minisymposia abstracts · index a absil, pierre-antoine, 44 acar, evrim, 35 agullo, emmanuel, 34,...

65
Minisymposia Abstracts

Upload: others

Post on 26-Mar-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Minisymposia Abstracts

IndexA

Absil, Pierre-Antoine, 44Acar, Evrim, 35Agullo, Emmanuel, 34, 46Ahuja, Kapil, 36, 40Aishima, Kensuke, 45Akian, Marianne, 48Alastair Spence, Alastair, 51Ali, Murtaza, 32Allamigeon, Xavier, 54Amat, Sergio, 62Amestoy, Patrick, 60Amodei, Luca, 64Amsallem, David, 10Antoine, Xavier, 19Antoulas, A.C., 38Arioli, Mario, 31, 40, 53Ashcraft, Cleve, 60Asif, M. Salman, 48Aurentz, Jared L., 59Avron, Haim, 16Axelsson, Owe, 12

BBaboulin, Marc, 12, 16Bai, Zhong-Zhi, 20Bailey, David H., 60Ballard, Grey, 22, 29Barioli, Francesco, 28Barlow, Jesse, 32Barreras, Alvaro, 9Barrett, Wayne, 28Barrio, Roberto, 60, 61Beattie, Christopher, 13, 38, 41Bebiano, Natalia, 14Becker, Dulceneia, 12, 16Becker, Stephen, 48Beckermann, Bernhard, 13Ben-Ari, Iddo, 19Bendito, Enrique, 14, 61Benner, Peter, 17, 36, 37Benning, Martin, 54Bergamaschi, Luca, 53Berman, Abraham, 28Beyer, Lucas, 32Bientinesi, Paolo, 32Bini, Dario A., 8Bioucas-Dias, Jose, 50Bıyıkoglu, Turker, 61Blank, Luise, 27Bognet, Brice, 21Boiteau, Olivier, 60

Boito, Paola, 58Bolten, Matthias, 9Boman, Erik G., 34Borm, Steffen, 51Borwein, Jonathan M., 60Bouhamidi, Abderrahman, 24Boutsidis, Christos, 13Boyanova, Petia, 12Bozovic, Nemanja, 39Brannick, James, 26Brauchart, Johann S., 64Breiten, Tobias, 17Bro, Rasmus, 35Bru, Rafael, 47Brualdi, Richard A., 21Buchot, Jean-Marie, 64Budko, Neil, 19Bueno, Maria I., 59Buluc, Aydın, 29Busquier, S., 62Butkovic, Peter, 55Buttari, Alfredo, 60Byckling, Mikko, 19

CCai, Xiao-Chuan, 40Calvin, Christophe, 45Campos, Carmen, 52Candes, Emmanuel, 48, 54Candela, Vicente F., 28, 62, 63Canto, Rafael, 9Carden, Russell L., 41, 55Carmona, Angeles, 14, 61Carson, Erin, 29, 30Castro Gonzalez, Nieves, 36Catral, Minerva, 9Cerdan, Juana, 53Chan, Raymond H., 56Chang, Xiao-Wen, 32Chaturantabut, Saifon, 13Chen, Jie, 24Chiang, Chun-Yueh, 42Chinesta, Francisco, 21Chitour, Yacine, 15Chow, Edmond, 31Chu, Delin, 42Chu, Eric King-wah, 39, 39Chu, Moody T., 39Cornelio, Anastasia, 30Cvetkovic-Ilic, Dragana S., 36

DDachsel, Holger, 43

2

3 2012 SIAM Conference on Applied Linear Algebra

Dahl, Geir, 21Damm, Tobias, 64Dandouna, Makarem, 45, 46de Hoop, Maarten V., 58, 58, 60de la Puente, Marıa Jesus, 54De Lathauwer, Lieven, 35De Sterck, Hans, 35de Sturler, Eric, 11, 13, 36, 40De Teran, Fernando, 59, 59De Vlieger, Jeroen, 56Deadman, Edvin, 8Demmel, James W., 22, 23, 29, 30Derevyagin, Maxim, 17Dhillon, Inderjit S., 24Di Benedetto, Maria Domenica, 14Diao, Huaian, 33, 43Djordjevi, Dragan S., 36Dolean, Victorita, 34Donatelli, Marco, 9, 30Dongarra, Jack, 12, 16, 46Dopico, Froilan M., 37, 37, 49, 59Drineas, Petros, 13Drmac, Zlatko, 41Druinsky, Alex, 22, 60Drummond, L. A. (Tony), 46, 52Druskin, Vladimir, 13, 17Du, Xiuhong, 26Dubois, Jerome, 45Duff, Iain S., 31, 31, 34Dunlavy, Daniel M., 35

EEffenberger, Cedric, 33Eidelman, Yuli, 52, 58Elman, Howard, 51Emad, Nahid, 45, 46Embree, Mark, 18, 55Encinas, Andres M., 14, 61Erhel, Jocelyne, 53Erway, Jennifer B., 50, 56Ezzatti, Pablo, 32

FFabregat-Traver, Diego, 32Fadili, Jalal, 48Falco, Antonio, 22, 29Falgout, Robert D., 11Fallat, Shaun M., 28, 28Farber, Miriam, 21Farhat, Charbel, 10Faverge, Mathieu, 46Feng, Lihong, 37Ferreira, Carla, 49Ferronato, Massimiliano, 47Fezzani, Riadh, 26Figueiredo, Mario, 50

Filar, Jerzy, 19Fornasini, Ettore, 15Freitag, Melina A., 51Fuenmayor, F. Javier, 21Furtado, Susana, 14Futamura, Yasunori, 45

GGalantai, Aurel, 46Gaubert, Stephane, 48, 54Gavalec, Martin, 49, 55Gemignani, Luca, 58Geuzaine, Christophe, 19Ghandchi Tehrani, Maryam, 57Ghnatios, Chady, 21Ghysels, Pieter, 29Gillis, Nicolas, 50Giner, Eugenio, 21Giraldi, Loic, 29Giraud, Luc, 34, 46Gohberg, Israel, 58Goldberg, Felix, 28Goldfarb, Donald, 54Goubault, Eric, 54Gracia, Juan-Miguel, 49Graf, Manuel, 65Grant, Michael, 48Grasedyck, Lars, 64Gratton, Serge, 27Greengard, Leslie, 52Greif, Chen, 15, 16Grigori, Laura, 23, 26, 29Grimes, Roger, 60Gu, Ming, 23Guermouche, Abdou, 34, 46Guerrero, Johana, 56Gugercin, Serkan, 11, 13, 36, 38, 41Guglielmi, Nicola, 10, 55, 57Guivarch, Roman, 34Guo, Chun-Hua, 63Gurbuzbalaban, Mert, 55Gutierrez, Jose M., 62Gutierrez-Gutierrez, Jesus, 18Guttel, Stefan, 13

HHached, Mustapha, 24Hackbusch, Wolfgang, 29Hall, H. Tracy, 28Han, Lixing, 9Hanke, Martin, 30Hartwig, R., 43Hauret, Patrice, 34Have, Pascal, 25Hayami, Ken, 33Hein, Matthias, 24

2012 SIAM Conference on Applied Linear Algebra 4

Heinecke, Alexander, 44Heins, Pia, 54Henson, Van Emden, 11Herault, Thomas, 46Hernandez, Araceli, 42Heroux, Michael A., 34Herzog, Roland, 27Heuveline, Vincent, 31Higham, Nicholas J., 8, 38Hilario, Lucıa, 22Ho, Kenneth L., 52Hogben, Leslie, 28Holodnak, John T., 40Holtz, Olga, 22Hsieh, Cho-Jui, 24Hu, Xiaofei, 50Huang, Tsung-Ming, 42Huckle, Thomas, 9, 47Huhs, Georg, 45Huhtanen, Marko, 19Humet, Matthias, 52Hunter, Jeffrey J., 19

IIannazzo, Bruno, 8Igual, Francisco D., 32Ipsen, Ilse C. F., 12, 40Ishteva, Mariya, 44

JJanna, Carlo, 47Jarlebring, Elias, 33, 64Jbilou, Khalide, 23, 24Jeuris, Ben, 44Jiang, Hao, 61Jimenez, Adrian, 54Jiranek, Pavel, 27Johnson, Charles R., 14Jokar, Sadegh, 38Jonsthovel, Tom, 47Joshi, Shantanu H., 56

KKabadshow, Ivo, 43Kahl, Karsten, 26Kambites, Mark, 49Karimi, Sahar, 54Karow, Michael, 49Katz, Ricardo D., 54Kaya, Oguz, 31Khabou, Amal, 23Khan, Mohammad Kazim, 30Kılıc, Mustafa, 55Kilmer, Misha E., 13Kirkland, Steve, 20Knight, Nicholas, 29, 30

Kolda, Tamara G., 35Kollmann, Markus, 20Kopal, Jirı, 37Koutis, Ioannis, 16Kraus, Johannes, 11Krendl, Wolfgang, 20Kressner, Daniel, 39, 56Kronbichler, Martin, 12Kunis, Susanne, 43Kuo, Chun-Hung, 39Kuo, Yueh-Cheng, 42

LL’Excellent, Jean-Yves, 60Landi, Germana, 50Langguth, Johannes, 31Langou, Julien, 46Latouche, Guy, 63Lattanzi, Marina, 42Laub, Alan J., 13Lecomte, Christophe, 57Legrain, Gregory, 29Leydold, Josef, 61Leygue, Adrien, 21Li, Ren-Cang, 63Li, Tiexiang, 39, 39Li, Xiaoye S., 58, 60Liesen, Jorg, 11Lim, Lek-Heng, 40Lim, Yongdo, 44Lin, Matthew M., 39, 42Lin, Wen-Wei, 39, 42Lipshitz, Benjamin, 22Litvinenko, Alexander, 22Lorenz, Dirk, 48Lubich, Christian, 10Lucas, Bob, 60Luce, Robert, 31Lukarski, Dimitar, 31

MMa, Shiqian, 54Mach, Susann, 27Mackey, D. Steven, 59, 59Maeda, Yasuyuki, 45Magdon-Ismail, Malik, 13Magrenan, A., 62Mahoney, Michael W., 16, 24Manguoglu, Murat, 25Marcia, Roummel F., 56Marek, Ivo, 20Marın, Jose, 47, 53Marques, Osni, 52Marquina, Antonio, 56Martin, David, 17Martin-Dorel, Erik, 61

5 2012 SIAM Conference on Applied Linear Algebra

Martınez, Angeles, 53Mas, Jose, 47, 53Mason, Paolo, 15Masson, Roland, 25Matthies, Hermann G., 22Mayo, Jackson R., 35McNally, Arielle Grim, 40Meerbergen, Karl, 16, 33, 51, 56Mehrmann, Volker, 38Melquiond, Guillaume, 61Meng, Xiangrui, 24Mengi, Emre, 55Michiels, Wim, 16, 33, 57, 64Miedlar, Agnieszka, 40Miller, Gary, 16Miller, Killian, 35Milzarek, Andre, 48Ming, Li, 40Mitjana, Margarida, 14, 61Moller, Michael, 54Montes, Nicolas, 22Mora, Marta C., 22Moreira Cardoso, Domingos, 21Morigi, Serena, 30Morikuni, Keiichi, 33Moro, Julio, 49Mosic, D., 36Moulding, Erin, 16Muhic, Andrej, 51Muller, Jean-Michel, 61

NNagy, James G., 30Nakata, Maho, 46Nakatsukasa, Yuji, 8, 34, 38, 45Napov, Artem, 58Nataf, Frederic, 25, 26, 34Neumann, Michael, 9Newton, Chris, 12Neytcheva, Maya, 12Ng, Esmond, G., 31Ng, Michael K., 50Nguyen, Giang T., 63Niknejad, Amir, 18Noschese, Silvia, 23Noutsos, Dimitrios, 9Nouy, Anthony, 21, 29Nuentsa Wakam, Desire, 53

OOgita, Takeshi, 46Oishi, Shin’ichi, 45Okoudjou, Kasso, 65Olshevsky, Vadim, 37Orban, Dominique, 16Osher, Stanley J., 56

Overton, Michael L., 55

PPaige, Chris, 40Pajonk, Oliver, 22Pan, Victor, 57Papathanasiou, Nikolaos, 9Parlett, Beresford, 49Patrıcio, Pedro, 43Pauca, Paul, 50Pearson, John, 20, 27Pechstein, Clemens, 34Pedregal, P., 62Pelaez, Marıa J., 9, 49Peled, Inon, 22Peng, Richard, 16Pena, Juan Manuel, 9Peris, Rosa M., 63Pestana, Jennifer, 15, 15Petiton, Serge G., 45PETSc Team, 46Pfetsch, Marc, 48Pfluger, Dirk, 44Piccolomini, Elena Loli, 30, 50Pinheiro, Eduardo, 18Pini, Giorgio, 47Pippig, Michael, 43Plavka, Jan, 55Plemmons, Robert J., 9, 50Plestenjak, Bor, 51Polizzi, Eric, 25Poloni, Federico, 41Popescu, Radu, 34Poulson, Jack, 53Prasad, Sudhakar, 50Providencia, J. da, 14Psarrakos, Panayiotis, 9

QQuraischi, Sarosh, 38

RRajamanickam, Siva, 34Ralha, Rui, 8Ramage, Alison, 12Ramaswami, V., 63Ramlau, Ronny, 30Rantzer, Anders, 38Raydan, Marcos, 56Rees, Tyrone, 15Reichel, Lothar, 17, 23, 30Reis, Timo, 41Remis, Rob, 13Robert, Yves, 46Rodenas, Juan Jose, 21Rohwedder, Thorsten, 10

2012 SIAM Conference on Applied Linear Algebra 6

Rojas, Marielba, 56Roman, Jean, 34, 46Roman, Jose E., 52Romberg, Justin, 48Romero Sanchez, Pantaleon David, 28, 62Rosic, Bojana, 22Rozloznık, Miroslav, 37Ruiz, Daniel, 34Rump, Siegfried M. , 61

SSaad, Yousef, 26Saak, Jens, 64Sachs, Ekkehard, 27Sadok, Hassane, 23Saied, Faisal, 25Sakurai, Tetsuya, 45Sameh, Ahmed H., 25Sanders, Geoffrey D., 11Sarbu, Lavinia, 27Sarkis, Marcus, 26Saunders, M. A., 24Scarpa, Giannicola, 28Schaerer, Christian E., 26Scheichl, Robert, 34Scheinberg, Katya, 54Schenk, Olaf, 25Schroder, Christian, 51Schroder, Jacob B., 11Schwartz, Oded, 22, 22, 29Sciriha, Irene, 21Scott, Jennifer, 53Sedlacek, Matous, 47Sergeev, Sergeı, 49Serra Capizzano, Stefano, 8Serrano, Sergio, 60Severini, Simone, 28Sgallari, Fiorella, 30Shader, Bryan, 28Sharify, Meisam, 48Shieh, Shih-Feng, 42Sifuentes, Josef, 18Sigalotti, Mario, 15Simic, Slobodan K., 62Simoncini, Valeria, 17, 20Smith, Barry, 46Smoktunowicz, Alicja, 37Solomonik, Edgar, 23Sonneveld, Peter, 26Soodhalter, Kirk M., 37Sootla, Aivar, 38Sorber, Laurent, 35Sorensen, D.C., 41Sou, Kin Cheong, 38Spence, Alastair, 51

Spillane, Nicole, 34Stevanovic, Dragan, 62Stoll, Martin, 27, 27Stovıcek, Pavel, 17Strakos, Zdenek, 11Stykel, Tatjana, 41Sun, Hai-Wei, 8Szydlarski, Mikolaj, 25Szyld, Daniel B., 15, 26, 36, 37

TTang, Ping Tak Peter, 25Taslaman, Leo, 34Thome, Nestor, 42Thornquist, Heidi, 53Tillmann, Andreas, 48Tisseur, Francoise, 34, 59Titley-Peloquin, David, 27Tobler, Christine, 39Toledo, Sivan, 16, 22, 60Tomov, Stanimire, 12Tran, John, 60Treister, Eran, 10Trenn, Stephan, 15Tuma, Miroslav, 37, 47Tyaglov, Mikhail, 14

UUcar, Bora, 31Ulbrich, Michael, 48Urbano, Ana M., 9Urquiza, Fabian, 42Uschmajew, Andre, 10, 44

VValcher, Maria Elena, 15Van Barel, Marc, 35, 52Van Beeumen, Roel, 16van de Geijn, Robert A., 32van den Berg, Ewout, 54van den Driessche, P., 28van der Holst, Hein, 28Van Dooren, Paul, 44van Gijzen, Martin B., 36, 47Van Houdt, Benny, 63Vandebril, Raf, 18, 44, 59Vandereycken, Bart, 44, 44Vanroose, Wim, 29Varona, J. L., 62Vassalos, Paris, 9Vassilevski, Panayot S., 11Vavasis, Stephen, 54Voss, Heinrich, 33Vuik, Kees, 47

7 2012 SIAM Conference on Applied Linear Algebra

WWagenbreth, Gene, 60Wang, Shen, 58, 60Wang, Wei-chao, 63Wang, Wei-guo, 63Wathen, Andrew J., 15, 15, 27Watkins, David S., 59Weber, Marcus, 18Wei, Yimin, 33, 43Weisbecker, Clement, 60Weiss, Jan-Philipp, 31Weng, Peter Chang-Yi, 39Willbold, Carina, 41Wirth, Fabian, 15Wittum, Gabriel, 26Womersley, Robert S., 65Wu, Chin-Tien, 42Wu, Minghao, 51Wu, Xunxun, 12Wulling, Wolfgang, 40Wyatt, Sarah, 11

XXia, Jianlin, 58, 58, 60Xu, Qingxiang, 43Xu, Shufang, 63Xue, Fei, 37Xue, Jungong, 63

YYamamoto, Kazuma, 45Yamanaka, Naoya, 45Yamazaki, Ichitaro, 45Yavneh, Irad, 10Ye, Qiang, 39Ye, Xuancan, 27Yıldırim, E. Alper, 55Yildiztekin, Kemal, 33Yuan, Xiaoming, 56

ZZangre, Ibrahim, 19Zaslavsky, Mikhail, 17Zemach, Ran, 10Zenadi, Mohamed, 34Zhang, Hong, 46Zhang, Jiani, 50Zhang, Qiang, 50Zhang, Wenxing, 56Zhao, Tao, 25Zhlobich, Pavel, 37Zietak, Krystyna, 8Zikatanov, Ludmil T., 11Zimmermann, Karel, 49Zinemanas, Alejandro, 32Zounon, Mawussi, 46

Zouros, Grigorios, 19Zuk, Or, 18Zulehner, Walter, 20, 20

2012 SIAM Conference on Applied Linear Algebra 8

MS 1. Recent advances in matrix functionsTalk 1. Computational issues related to the geometric meanof structured matricesIn several applications it is required to compute the geometricmean of a set of positive definite matrices. In certain cases, likein the design of radar systems, the matrices are structured, say,they are Toeplitz, and one expects that the mean still keeps thesame structure.Unfortunately, the available definitions of geometric mean, likethe Karcher mean, do not generally maintain the structure of theinput matrices.In this talk we introduce a definition of mean which preservesthe structure, analyze its properties and present algorithms for itscomputation.Dario A. BiniDipartimento di MatematicaUniversita di Pisa, [email protected]

Bruno IannazzoDipartimento di Matematica e InformaticaUniversita di Perugia, [email protected]

Talk 2. Efficient, communication-minimizing algorithms forthe symmetric eigenvalue decomposition and the singularvalue decompositionWe propose algorithms for computing the symmetric eigenvaluedecomposition and the singular value decomposition (SVD) thatminimize communication in the asymptotic sense whilesimultaneously having arithmetic operation costs within a factor3 of that for the most efficient existing algorithms. The essentialcost for the algorithms is in performing QR factorizations, ofwhich we require no more than six for the symmetriceigenproblem and twelve for the SVD. We establish backwardstability under mild assumptions, and numerical experimentsindicate that our algorithms tend to yield decompositions withconsiderably smaller backward error and eigen/singular vectorscloser to orthogonal than existing algorithms.Yuji NakatsukasaSchool of MathematicsThe University of [email protected]

Nicholas J. HighamSchool of MathematicsThe University of [email protected]

Talk 3. The Pade approximation and the matrix sign functionIn the talk we focus on the properties of the Pade approximantsof a certain hypergeometric function on which the Pade familiesof iterations for computing the matrix sign and sector functionsare based. We have determined location of poles of the Padeapproximants and we have proved that all coefficients of thepower series expansions of the reciprocals of the denominatorsof the Pade approximants and of the power series expansions ofthe Pade approximants are positive. These properties are crucialin the the proof of the conjecture – stated by Laszkiewicz andZietak, and extending the result of Kenney and Laub – on aregion of convergence of these Pade families of iterations.The talk is based on the paper by O. Gomilko, F. Greco, K.Zietak and the paper by O. Gomilko, D.B. Karp, M. Lin, K.Zietak. In the talk we also consider a certain different kind ofrational approximation to the sign function.

Krystyna ZietakInstitute of Mathematics and Computer ScienceWrocław University of [email protected]

Talk 4. A recursive blocked schur algorithm for computingthe matrix square rootThe Schur algorithm for computing a matrix square root reducesthe matrix to the Schur triangular form and then computes asquare root of the triangular matrix. In this talk I will describe arecursive blocking technique in which the computation of thesquare root of the triangular matrix can be made rich in matrixmultiplication. Numerical experiments making appropriate useof level 3 BLAS show significant speedups over the pointalgorithm, both in the square root phase and in the algorithm as awhole. The excellent numerical stability of the point algorithm isshown to be preserved by recursive blocking. Recursive blockingis also shown to be effective for multiplying triangular matrices.Edvin DeadmanUniversity of [email protected]

Nicholas J. HighamSchool of MathematicsUniversity of [email protected]

Rui RalhaCenter of MathematicsUniversity of Minho, Portugalr [email protected]

MS 2. Methods for Toeplitz matrices and theirapplicationTalk 1. Toeplitz operators with matrix-valued symbols andsome (unexpected) applicationsWe discuss the eigenvalue distribution in the Weyl sense ofgeneral matrix-sequences associated to a symbol. As a specificcase we consider Toeplitz sequences generated by matrix-valued(non Hermitian) bounded functions. We show that the canonicaldistribution can be proved under mild assumptions on thespectral range of the given symbol. Finally some applicationsare introduced and discussed.Stefano Serra CapizzanoDipartimento di Fisica e MatematicaUniversita dell’Insubria - sede di Como, [email protected]

Talk 2. Fast approximation to the Toeplitz matrix exponentialThe shift-invert Lanczos (or Arnoldi) method is employed togenerate an orthonormal basis from the Krylov subspacecorresponding to the real Toeplitz matrix and an initial vector.The vectors and recurrence coefficients produced by this methodare exploited to approximate the Toeplitz matrix exponential.Toeplitz matrix inversion formula and rapid Toeplitzmatrix-vector multiplications are utilized to lower thecomputational costs. For convergence analysis, a sufficientcondition is established to guarantee that the error bound isindependent of the norm of the matrix. Numerical results andapplications to the computational finance are given todemonstrate the efficiency of the method.Hai-Wei SunFaculty of Science and TechnologyUniversity of Macau, [email protected]

9 2012 SIAM Conference on Applied Linear Algebra

Talk 3. Matrix algebras sequences can be spectrallyequivalent with ill-conditioned Toeplitz onesThe construction of fast and efficient iterative procedure oreffective multigrid schemes, requires the study of the spectrumof the matrix sequences Ann, whereAnn = P−1

n (f)Tn(f)n. In this talk, we focus on the casewhere Tn(f) is a Toeplitz matrix generated by a nonnegativereal function f , and Pn(f) denotes matrices belonging to tau orcirculant algebras. Assuming that the symbol f has discreteroots of non integer order, we will show that under suitableassumptions the spectrum of the matrix sequence Ann isbounded by constants far away from zero and infinity. Using thedeveloped theory, we propose effective preconditioners and wegive hints for optimal multigrid schemes.Paris VassalosDept. of InformaticsAthens University of Economics and [email protected]

Dimitrios NoutsosDept. of MathematicsUniversity of [email protected]

Talk 4. Aggregation-based multigrid methods for ToeplitzmatricesAggregation and smoothed aggregation based multigrid methodscan be analyzed in the context of the convergence theory forToeplitz and circulant matrices. As in the aggregation-basedmultigrid methods, for generating symbols with a single isolatedzero at the origin aggregates are formed. The interpolation canthen be improved by applying an additional operator. Thisimprovement can be interpreted as the smoothing step in theoriginal smoothed aggregation method. Depending on theoriginal generating symbol, several smoothing steps can benecessary. Numerical examples show the efficiency of theaggregation based approach in the case of Toeplitz and circulantmatrices.Matthias BoltenDepartment of Mathematics and ScienceUniversity of Wuppertal, [email protected]

Marco DonatelliDipartimento di Fisica e MatematicaUniversita dell’Insubria - sede di Como, [email protected]

Thomas HuckleDept. of InformaticsTechnische Universitat Munchen, [email protected]

MS 3. Matrix factorizations and applicationsTalk 1. Classes of matrices with bidiagonal factorizationMatrices with a bidiagonal decomposition satisfying some signrestrictions are analyzed. They include all nonsingular totallypositive matrices, their matrices opposite in sign and theirinverses, as well as tridiagonal nonsingular H-matrices.Properties of these matrices are presented and the bidiagonalfactorization can be used to perform computations with highrelative accuracy.Alvaro BarrerasDept. of Applied Mathematics / IUMAUniversidad de [email protected]

Juan Manuel PenaDept. of Applied Mathematics / IUMAUniversidad de [email protected]

Talk 2. Cholesky factorization for singular matricesDirect and iterative solution methods for linear least-squaresproblems have been studied in Numerical Linear Algebra. Partof the difficulty for solving least-squares problems is the factthat many methods solve the system by implicitly solving thenormal equations ATAx = AT b with A ∈ Rn×m and b ∈ Rn.In this talk we consider this problem when A is a rank deficientmatrix. Then, ATA is a positive semidefinite matrix and itsCholesky factorization is not unique. So, we introduce a fullrank Cholesky decomposition LLT of the normal equationsmatrix without the need to form the normal matrix itself. Wepresent two algorithms to compute the entries of L by rows (orby columns) with the corresponding error analysis. Numericalexperiments illustrating the proposed algorithms are given.Rafael CantoDept. of Applied MathematicsUniversitat Politecnica de [email protected]

Ana M. UrbanoDept. of Applied MathematicsUniversitat Politecnica de [email protected]

Marıa J. PelaezDept. of MathematicsUniversidad Catolica del [email protected]

Talk 3. Applications of the singular value decomposition toperturbation theory of eigenvalues of matrix polynomialsIn this talk, motivated by a problem posed by Wilkinson, westudy the coefficient perturbations of a n× n matrix polynomialto n× n matrix polynomials which have a prescribed eigenvalueof specified algebraic multiplicity and index of annihilation. Foran n× n matrix polynomial P (λ) and a given scalar µ ∈ C, weintroduce two weighted spectral norm distances, Er(µ) andEr,k(µ), from P (λ) to the n× n matrix polynomials that have µas an eigenvalue of algebraic multiplicity at least r and to thosethat have µ as an eigenvalue of algebraic multiplicity at least rand maximum Jordan chain length exactly k, respectively. Thenwe obtain a singular value characterization of Er,1(µ), andderive a lower bound for Er,k(µ) and an upper bound for Er(µ),constructing associated perturbations of P (λ).Panayiotis PsarrakosDept. of MathematicsNational Technical University of [email protected]

Nikolaos PapathanasiouDept. of MathematicsNational Technical University of [email protected]

Talk 4. On reduced rank nonnegative matrix factorizationfor symmetric nonnegative matricesFor a nonnegative matrix V ∈ Rm,n, the nonnegative matrixfactorization (NNMF) problem consists of finding nonnegativematrix factors W ∈ Rm,r and H ∈ Rr,n such that V ≈WH .We consider the algorithm provided by Lee and Seung whichfinds nonnegative W and H such that ‖V −WH‖F isminimized. For the case m = n and in which V is symmetric,we present results concerning when the best approximate

2012 SIAM Conference on Applied Linear Algebra 10

factorization results in the product WH being symmetric and oncases in which the best approximation cannot be a symmetricmatrix. Results regarding other special cases as well asapplications are also discussed.Minerva CatralDept. of MathematicsXavier [email protected]

Lixing HanDept. of MathematicsUniversity of [email protected]

Michael NeumannDept. of MathematicsUniversity of Connecticut

Robert J. PlemmonsDept. of Mathematics and Computer ScienceWake Forest [email protected]

MS 4. Algorithms on manifolds of low-rank matricesand tensorsTalk 1. Low rank dynamics for computing extremal points ofreal and complex pseudospectraWe consider the real ε-pseudospectrum of a square matrix,which is the set of eigenvalues of all real matrices that areε-close to the given matrix, where closeness is measured ineither the 2-norm or the Frobenius norm.We characterize extremal points and compare the situation withthat for the unstructured ε-pseudospectrum. We presentdifferential equations for low rank (1 and 2) matrices for thecomputation of the extremal points of the pseudospectrum.Discretizations of the differential equations yield algorithms thatare fast and well suited for sparse large matrices. Based on theselow-rank differential equations, we further obtain an algorithmfor drawing boundary sections of the structured pseudospectrumwith respect to both considered norms.Nicola GuglielmiDipartimento di MatematicaUniversita di L’Aquila, [email protected]

Christian LubichMathematisches InstitutUniversitat [email protected]

Talk 2. Parametric model order reduction using stabilizedconsistent interpolation on matrix manifoldsA robust method for interpolating reduced-order linear operatorson their matrix manifolds is presented. Robustness is achievedby enforcing consistency between the different sets ofgeneralized coordinates underlying different instances ofparametric reduced-order models (ROMs), and explicitlystabilizing the final outcome of the interpolation process.Consistency is achieved by transforming the ROMs beforeinterpolation using a rotation operator obtained from the solutionof a generalized orthogonal Procrustes problem. Stabilization isachieved using a novel real-time algorithm based on semidefiniteprogramming. The overall method is illustrated with its on-lineapplication to the parametric fluid-structure analysis of awing-store configuration.David Amsallem

Department of Aeronautics and AstronauticsStanford [email protected]

Charbel FarhatDepartment of Aeronautics and AstronauticsDepartment of Mechanical EngineeringInstitute for Computational Methods in EngineeringStanford [email protected]

Talk 3. Treatment of high-dimensional problems by low-rankmanifolds of tensorsMany problems in e.g. natural sciences, data mining andstatistics are naturally posed as high-dimensionalproblems.Tensor product representations parametrize theseproblem on approximation manifolds, often fixed by a givenmaximal rank for the tensor approximation. A relatively newformat is the HT/TT format as developed by Hackbusch(Leipzig) and Oseledets/Tyrtyshnikov (Moscow), whichovercomes the shortcomings of older formats and gives a stableand often sparse opportunity to represent high-dimensionalquantities, so that the treatment of e.g. high-dimensional partialdifferential equations is on the verge of realization. In this talk, ageneral framework to realize such tasks in data-sparse tensorformats is presented. For the HT/TT format, we present sometheoretical as well as algorithmic results for the treatment ofhigh-dimensional optimization tasks and high-dimensionalevolution equations.Thorsten RohwedderInstitut fur MathematikTechnische Universitat [email protected]

Talk 4. Local convergence of alternating optimization ofmultivariate functions in the presence of scalingindeterminaciesAn easy and widely used approach to minimize functions withrespect to certain tensor formats (such as CP, Tucker, TT or HT)is the alternating optimization algorithm, also called nonlinearGauss–Seidel method. A prominent example is thePARAFAC-ALS algorithm. Due to the usual non-uniqueness oftensor representations, standard convergence results fornonlinear Gauss-Seidel are usually not directly applicable. Inthis talk we present a quite general approach to prove localconvergence, based on a geometric viewpoint, which regardstensors of fixed rank as orbits of a Lie group generatingequivalent representations.Andre UschmajewInstitut fur MathematikTechnische Universitat (TU) [email protected]

MS 5. Advances in algebraic multigrid - Newapproaches and applicationsTalk 1. Algebraic collocation coarse approximation multigridMost algebraic multigrid (AMG) methods define the coarseoperators by applying the (Petrov-)Galerkin coarseapproximation where the sparsity pattern and operatorcomplexity of the multigrid hierarchy is dictated by themultigrid prolongation and restriction. Therefore, AMGalgorithms usually must settle on some compromise between thequality of these operators and the aggressiveness of thecoarsening, which affects their rate of convergence and operatorcomplexity. In this paper we propose an algebraic generalization

11 2012 SIAM Conference on Applied Linear Algebra

of the collocation coarse approximation (CCA) approach ofWienands and Yavneh, where the choice of the sparsity patternof the coarse operators is independent of the choice of thehigh-quality transfer operators. The new algorithm is based onthe aggregation framework (smoothed and non-smoothed).Using a small set of low-energy eigenvectors, it computes thecoarse grid operator by a weighted least squares process.Numerical experiments for two dimensional diffusion problemswith sharply varying coefficients demonstrate the efficacy andpotential of this multigrid algorithm.Eran TreisterDepartment of Computer ScienceTechnion—Israel Institute of Technology, [email protected]

Ran ZemachDepartment of Computer ScienceTechnion—Israel Institute of Technology, [email protected]

Irad YavnehDepartment of Computer ScienceTechnion—Israel Institute of Technology, [email protected]

Talk 2. Energy-minimization interpolation for adaptivealgebraic multigridAdaptive algebraic multigrid (AMG) methods automaticallydetect the algebraically smooth error for a linear system and assuch, are robust black box approaches for solving difficultproblems. However, the two main families of methods,Bootstrap AMG (BAMG) and adaptive smoothed aggregation(αSA), suffer from drawbacks, such as noisy candidate vectors(BAMG) and potentially high operator complexity, especiallyfor scalar problems, (αSA). This work is intended to addresssome of these drawbacks by combining elements of BAMG andαSA through an energy-minimization interpolation framework.While the approach is general, the challenging, motivatingproblem is a biquadratic discretization of anisotropic diffusion.Jacob B. SchroderCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

Robert D. FalgoutCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

Talk 3. Algebraic multigrid (AMG) for complex networkcalculationsClustering, ranking, or measuring distance for vertices onscale-free graphs are all computational task of interest. Suchcalculations may be obtained by solving eigensystems or linearsystems involving matrices whose sparsity structure is related toan underlying scale-free graph. For many large systems ofinterest, classical iterative solvers (such as conjugate gradientand Lanczos) converge with prohibitively slow rates, due toill-conditioned matrices and small spectral gap ratios. Efficientlypreconditioning these systems with multilevel methods isdifficult due to the scale-free topology. For some large modelproblems and real-world networks, we investigate theperformance of a few AMG-related coarsening approaches.Geoffrey D. SandersCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

Van Emden HensonCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

Panayot S. VassilevskiCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

Talk 4. The polynomial of best uniform approximation to 1/xas smoother in two grid methodsWe discuss a simple convergence analysis of two level methodswhere the relaxation on the fine grid uses the polynomial of bestapproximation in the uniform norm to 1/x on a finite intervalwith positive endpoints. The construction of the latterpolynomial is of interest by itself, and we have included aderivation of a three-term recurrence relation for computing thispolynomial. We have also derived several inequalities related tothe error of best approximation, monotonicity of the polynomialsequence, and positivity which we applied in the analysis oftwo-level methods.Ludmil T. ZikatanovDepartment of MathematicsThe Pennsylvania State University, [email protected]

Johannes KrausJohann Radon Institute for Computational and Applied Mathematics(RICAM), [email protected]

Panayot S. VassilevskiCenter for Applied Scientific ComputingLawrence Livermore National Laboratory (LLNL), [email protected]

MS 6. Recent advances in fast iterative solvers - Part Iof IITalk 1. Challenges in analysis of Krylov subspace methodsThe current state-of-the art of iterative solvers is the outcome ofthe tremendous algorithmic development over the last fewdecades and of investigations of how to solve given problems. Inthis contribution we focus on Krylov subspace methods andmore on the dual question why things do or do not work. Inparticular, we will pose and discuss open questions such as whatthe spectral information tells us about the behaviour of Krylovsubspace methods, to which extent we can relax the accuracy oflocal operations in inexact Krylov subspace methods withoutcausing an unwanted delay, how important is considering ofrounding errors in various algorithmic techniques, whether it isuseful to view Krylov subspace methods as matching momentmodel reduction, and how the algebraic error can be includedinto locally efficient and fully computable a-posteriori errorbounds for adaptive PDE solvers.Zdenek StrakosCharles University in Prague and Academy of Science of the [email protected]

Jorg LiesenInstitute of MathematicsTechnical University of [email protected]

Talk 2. Updating preconditioners for parameterized systemsParameterized linear systems arise in a range of applications,such as model reduction, acoustics, and inverse problems basedon parametric level sets. Computing new preconditioners for

2012 SIAM Conference on Applied Linear Algebra 12

many small changes in parameter(s) would be very expensive.However, using one or only a few preconditioners for all systemsleads to an excessive total number of iterations. We showstrategies to efficiently compute effective multiplicative updatesto preconditioners. This has the advantage that, for a slightincrease in overhead, the update is more or less independent ofthe original preconditioner used.Eric de SturlerDepartment of MathematicsVirginia [email protected]

Sarah WyattDepartment of MathematicsVirginia [email protected]

Serkan GugercinDepartment of MathematicsVirginia [email protected]

Talk 3. Efficient preconditioning techniques for two-phaseflow simulationsIn this talk we present efficient preconditioning strategies,suitable for the iterative solution of algebraic systems, arising innumerical simulations of two-phase flow problems, modelled bythe Cahn-Hilliard equation.In this work, we decompose the original Cahn-Hilliard equation,which is nonlinear and of fourth-order, into a system of twosecond-order equations for the so-called ’concentration’ and’chemical potential’. The so-obtained nonlinear system isdiscretized using the finite element method and solved by twovariants of the inexact Newton method. The major focus of thework is to construct a preconditioner for the correcpondingJacobian, which is of block two-by-two form. The proposedpreconditioning techniques are based on approximatefactorization of the discrete Jacobian and utilise to a full extentthe properties of the underlying matrices.We propose a preconditioning method that reduces the problemof solving the non-symmetric discrete Cahn-Hilliard system tothe problem of solving systems with symmetric positive definitematrices, where off-the-shelf multilevel and multigrid algorithmsare directly applicable. The resulting iterative method exhibitsoptimal convergence and computational complexity properties.The efficiency of the resulting preconditioning method isillustrated via various numerical experiments, including largescale examples of both 2D and 3D problems.The preconditioning techniques are more generally applicablefor any algebraic system of the same structure as, e.g., whensolving complex symmetric linear systems.M. NeytchevaInstitute for Information TechnologyUppsala [email protected]

Owe AxelssonInstitute of GeonicsAcademy of Sciences of the Czech Republic, [email protected]

Petia BoyanovaInstitute for Information TechnologyUppsala [email protected]

Martin KronbichlerInstitute for Information TechnologyUppsala University

[email protected]

Xunxun WuInstitute for Information TechnologyUppsala [email protected]

Talk 4. Preconditioners in liquid crystal modellingLiquid crystal displays are ubiquitous in modern life, being usedextensively in monitors, televisions, gaming devices, watches,telephones etc. Appropriate models feature characteristic lengthand time scales varying by many orders of magnitude, whichprovides difficult numerical challenges to those trying tosimulate real-life dynamic industrial situations. The efficientsolution of the resulting linear algebra sub-problems is of crucialimportance for the overall effectiveness of the algorithms used.In this talk we will present some examples of saddle-pointsystems which arise in liquid crystal modelling and discuss theirefficient solution using appropriate preconditioned iterativemethods.Alison RamageDept of Mathematics and StatisticsUniversity of StrathclydeGlasgow, [email protected]

Chris NewtonNext Generation Displays GroupHewlett-Packard LaboratoriesBristol, [email protected]

MS 7. Application of statistics to numerical linearalgebra algorithms - Part I of IITalk 1. Fast linear system solvers based on randomizationtechniquesWe illustrate how dense linear algebra calculations can beenhanced by randomizing general or symmetric indefinitesystems. This approach, based on a multiplicativepreconditioning of the initial matrix, revisits the work from[Parker, 1995]. It enables us to avoid pivoting and then to reducesignificantly the amount of communication. This can beperformed at a very affordable computational price whileproviding a satisfying accuracy. We describe solvers based onrandomization that take advantage of the latest generation ofmulticore or hybrid multicore/GPU machines and we comparetheir Gflop/s performance with solvers from standard parallellibraries.Marc BaboulinInria Saclay Ile-de-FranceUniversity Paris-Sud, [email protected]

Dulceneia BeckerUniversity of Tennessee, [email protected]

Jack DongarraUniversity of Tennessee, [email protected]

Stanimire TomovUniversity of Tennessee, [email protected]

Talk 2. Numerical issues in randomized algorithmsRandomized algorithms are starting to find their way into a widevariety of applications that give rise to massive data sets. Thesealgorithms downsize the enormous matrices by picking and

13 2012 SIAM Conference on Applied Linear Algebra

choosing only particular columns or rows, thereby producingpotentially huge savings in storage and computing speed.Although randomized algorithms can be fast and efficient, notmuch is known about their numerical properties. We will discussthe numerical sensitivity and stability of randomized algorithms,as well as the error due to randomization, and the effect of thecoherence of the matrix. Algorithms under consideration includematrix multiplication, least squares solvers, and low-rankapproximations.Ilse IpsenDepartment of MathematicsNorth Carolina State University, [email protected]

Talk 3. Near-optimal column based matrix reconstructionWe consider low-rank reconstruction of a matrix using a subsetof its columns and we present asymptotically optimal algorithmsfor both spectral norm and Frobenius norm reconstruction. Themain tools we introduce to obtain our results are: (i) the use offast approximate SVD-like decompositions for column-basedmatrix reconstruction, and (ii) two deterministic algorithms forselecting rows from matrices with orthonormal columns,building upon the sparse representation theorem fordecompositions of the identity that appeared in [Batson,Spielman, Srivastava, 2011].Christos BoutsidisMathematical Sciences DepartmentIBM T. J. Watson Research Center, [email protected]

Petros DrineasComputer Science DepartmentRensselaer Polytechnic Institute, [email protected]

Malik Magdon-IsmailComputer Science DepartmentRensselaer Polytechnic Institute, [email protected]

Talk 4. Numerical experiments with statistical conditionestimationIn this talk we present the results of some numerical experimentsillustrating the use of statistical condition estimation (SCE).After a brief review of SCE, the use of the technique isdemonstrated on a few topics of general interest including theestimation of the condition of large sparse linear systems.Alan J. LaubDepartment of MathematicsUniversity of California Los Angeles, [email protected]

MS 8. Rational Krylov methods: analysis andapplications - Part I of IITalk 1. Solving Sylvester equations through rational GalerkinprojectionsRecently (B. Beckermann, An error analysis for rationalGalerkin projection applied to the Sylvester equation, SIAM J.Num. Anal. 49 (2012), 2430-2450), we suggested a new erroranalysis for the residual of Galerkin projection onto rationalKrylov spaces applied to a Sylvester equation with a rank 1right-hand side. In this talk we will consider more generallysmall rank right-hand sides, where block Krylov methods andtangential interpolation problems play an important role.Bernhard BeckermannDept. of Mathematics

University of [email protected]

Talk 2. Stability-corrected spectral Lanczos decompositionalgorithm for wave propagation in unbounded domainsApplying Krylov subspace methods to exterior wave fieldproblems has become an active topic of recent research. In thispaper, we introduce a new Lanczos-based solution method viastability-corrected operator exponents, allowing us to constructstructure-preserving reduced-order models (ROMs) respectingthe delicate spectral properties of the original scatteringproblem. The ROMs are unconditionally stable and are based ona renormalized Lanczos algorithm, which enables us toefficiently compute the solution in the frequency and timedomain. We illustrate the performance of our method through anumber of numerical examples in which we simulate 2Delectromagnetic wave propagation in unbounded domains.Rob RemisFaculty of Electrical Engineering, Mathematics and Computer ScienceDelft University of [email protected]

Vladimir DruskinSchlumberger-Doll [email protected]

Talk 3. Generalized rational Krylov decompositionsThe notion of an orthogonal rational Arnoldi decomposition, asintroduced by Axel Ruhe, is a natural generalization of thewell-known (polynomial) Arnoldi decomposition. Generalizedrational Krylov decompositions are obtained by removing theorthogonality assumption. Such decompositions are interestinglinear algebra objects by themselves and we will study some oftheir algebraic properties, as well as the rational Krylov methodsthat can be associated with them.Stefan GuttelMathematical InstituteUniversity of [email protected]

Talk 4. Interpolatory model reduction strategies for nonlinearparametric inversionWe will show how reduced order models can significantlyreduce the cost of general inverse problems approached throughparametric level set methods. Our method drastically reduces thesolution of forward problems in diffuse optimal tomography(DOT) by using interpolatory, i.e. rational Krylov based,parametric model reduction. In the DOT setting, these surrogatemodels can approximate both the cost functional and associatedJacobian with little loss of accuracy and significantly reducedcost.Serkan GugercinDept. of MathematicsVirginia [email protected]

Christopher A. BeattieDept. of MathematicsVirginia [email protected]

Saifon ChaturantabutDept. of MathematicsVirginia [email protected]

Eric de SturlerDept. of MathematicsVirginia [email protected]

2012 SIAM Conference on Applied Linear Algebra 14

Misha E. KilmerDept. of MathematicsTufts [email protected]

MS 9. New trends in tridiagonal matrices - Part I of IITalk 1. Direct and inverse problems on pseudo-JacobimatricesA theorem of Friedland and Melkman states the unique recoveryof a non-negative Jacobi matrix from the spectra of its upper andlower principal submatrices obtained by deleting the kth row andcolumn. Here this result is revisited and generalized. Otherrelated problems are also investigated, including existence anduniqueness theorems.Natalia BebianoDepartment of MathematicsUniversity of Coimbra3001-454 [email protected]

Susana FurtadoFaculty of EconomicsUniversity of Porto4200-464 [email protected]

J. da ProvidenciaDepartment of PhysicsUniversity of Coimbra3004-516 [email protected]

Talk 2. Schwartz’s matrices and generalized HurwitzpolynomialsWe present solutions of direct and inverse spectral problems fora kind of Schwarz’s matrices (tridiagonal matrices with only onenonzero main-diagonal entry). We study dependence of spectraof such matrices on the signs of their off-main-diagonal entries.We show that for certain distributions of those signs, thecharacteristic polynomial of the correspondent Schwarz’s matrixis a generalized Hurwitz polynomial. Recall that a realpolynomial p(z) = p0(z2) + zp1(z2), where p0 and p1 are theeven and odd parts of p, respectively, is generalized Hurwitz ifand only if the zeroes of p0 and p1 are real, simple andinterlacing.Mikhail TyaglovInstitut fur MathematikTechnische Universitat BerlinSekretariat MA 4-5Straße des 17. Juni 13610623 [email protected]

Talk 3. On the Moore-Penrose inverse of singular, symmetricand periodic Jacobi M-matricesWe aim here at determining the Moore–Penrose inverse, J† ofany singular, symmetric and periodic Jacobi M–matrix J, boththroughout direct computations or considering it as aperturbation of a singular, symmetric and nonperiodic JacobiM–matrix. We tackle the problem by applying methods fromthe operator theory on finite networks, since the off–diagonalentries of J can be identified with the conductance function of aweighted n–cycle. Then, J appears as a positive–semidefinite

Schrodinger operator on the cycle and hence, J† is nothing elsethat the corresponding Green operator.We also consider the problem of characterizing when J† is itselfan M -matrix.Andres M. EncinasDepartament de Matematica Aplicada IIIUniversitat Politecnica de Catalunya08034 [email protected]

Enrique BenditoDepartament de Matematica Aplicada IIIUniversitat Politecnica de Catalunya08034 [email protected]

Angeles CarmonaDepartament de Matematica Aplicada IIIUniversitat Politecnica de Catalunya08034 [email protected]

Margarida MitjanaDepartament de Matematica Aplicada IUniversitat Politecnica de Catalunya08034 [email protected]

Talk 4. The commutant of the tridiagonal patternWe consider patterns that allow real commutativity with anirreducible tridiagonal n-by-n pattern. All are combinatoriallysymmetric. All also allow a complex symmetric commuting pair,but only some allow a real symmetric commuting pair. We alsoshow that any matrix that commutes with an irreducibletridiagonal matrix satisfies certain ratio equations.Generalizations to matrices with tree patterns are alsoconsidered.Charles R. JohnsonDepartment of MathematicsCollege of William and MaryPO Box 8795Williamsburg, VA [email protected]

MS 10. Numerical algorithms for switching systems:from theory to applicationsTalk 1. Observer design for hybrid systemsThe state estimation problem has been the subject of intensivestudy for many years by both the computer science communityin the discrete domain and the control community in thecontinuous domain, but only scantly investigated in the hybridsystem domain.In this talk, we present a design methodology for dynamicalobservers of hybrid systems with linear continuous-timedynamics, which reconstruct the complete state from theknowledge of the inputs and outputs of a hybrid plant.We demonstrate the methodology by building a hybrid observerfor an industrial automotive control problem: on-lineidentification of the actual engaged gear.Maria Domenica Di BenedettoUniversity of L’[email protected]

15 2012 SIAM Conference on Applied Linear Algebra

Talk 2. About polynomial instability for linear switchedsystemsIn this talk we present recent results on the characterization ofmarginal instability for linear switched systems. Our maincontribution consists in pointing out a resonance phenomenonassociated with marginal instability. In particular we derive anupper bound of the norm of the state at time t, which ispolynomial in t and whose degree is computed from theresonance structure of the system. We also derive analogousresults for discrete-time linear switched systems.Paolo MasonLaboratoire des Signaux et [email protected]

Yacine ChitourLaboratoire des Signaux et SystemesUniversite Paris [email protected]

Mario SigalottiCMAP, INRIA [email protected]

Talk 3. Stability and stabilization of positive switchedsystems: state of the art and open problemsA positive switched system (PSS) consists of a family of positivestate-space models and a switching law, specifying when andhow the switching among the various models takes place. Thisclass of systems has some interesting practical applications:PSS’s have been adopted for describing networks employingTCP and other congestion control applications, for modelingconsensus and synchronization problems, and, quite recently, fordescribing the viral mutation dynamics under drug treatment.In the talk we will provide a comprehensive picture of theconditions for stability and for stabilizability, and we will pointout some open problems.Maria Elena ValcherDept. of Information EngineeringUniversity of [email protected]

Ettore FornasiniDept. of Information EngineeringUniversity of [email protected]

Talk 4. The joint spectral radius for semigroups generated byswitched differential algebraic equationsWe introduce the joint spectral radius for matrix semigroups thatare generated by switched linear ordinary differential equationswith jumps. The jumps are modeled by projectors which areassumed to commute with the generators of the flow. This setupcovers switched differential algebraic equations. It is shown thatan exponential growth rate can only be defined if the discretesemigroup generated by the projections is product bounded.Assuming this is true we show that Barabanov norms may bedefined in the irreducible case and that a converse Lyapunovtheorem holds. Some further properties of the exponentialgrowth rate are discussed.Fabian WirthDept. of MathematicsUniversity of Wurzburg, [email protected]

Stephan TrennDept. of MathematicsTechnical University of Kaiserslautern, [email protected]

MS 11. Recent advances in fast iterative solvers -Part II of IITalk 1. Combination preconditioning of saddle-point systemsfor positive definitenessThere are by now many preconditioning techniques forsaddle-point systems but also a growing number of applicationswhere such are required. In this talks we will discusspreconditioners which preserve self-adjointness in non-standardinner products and the combination of such preconditioners asintroduced by Martin Stoll and the first author in 2008.Here we will show how two preconditioners which ensureself-adjointness in different inner products, but which are bothindefinite may be combined to yield a positive definiteself-adjoint preconditioned system in a third inner product whichthus allows robust application of the Hestenes Steifel ConjugateGradient method in that inner product.Andrew J. WathenMathematical InstituteOxford University, [email protected]

Jennifer PestanaMathematical InstituteOxford University, [email protected]

Talk 2. Preconditioned iterative methods for nonsymmetricmatrices and nonstandard inner productsThe convergence of a minimum residual method applied to alinear system with a nonnormal coefficient matrix is not wellunderstood in general. This can make choosing an effectivepreconditioner difficult. In this talk we present a new GMRESconvergence bound. We also show that the convergence of anonstandard minimum residual method, applied to apreconditioned system, is bounded by a term that dependsprimarily on the eigenvalues of the preconditioned coefficientmatrix, provided the preconditioner and coefficient matrix areself-adjoint with respect to nearby Hermitian sesquilinear forms.We relate this result to the convergence of standard methods.Jennifer PestanaMathematical InstituteUniversity of [email protected]

Andrew J. WathenMathematical InstituteUniversity of [email protected]

Talk 3. Multi-preconditioned GMRESStandard Krylov subspace methods only allow the user to choosea single preconditioner. In many situations, however, there is no‘best’ choice, but a number of possible candidates. In this talkwe describe an extension of GMRES, multi-preconditionedGMRES, which allows the use of more than one preconditioner,and combines their properties in an optimal way. As well asdescribing some theoretical properties of the new algorithm wewill present numerical results which highlight the utility of theapproach.Tyrone ReesNumerical Analysis GroupRutherford Appleton [email protected]

Chen GreifDept. of Computer ScienceUniversity of British Columbia

2012 SIAM Conference on Applied Linear Algebra 16

[email protected]

Daniel B. SzyldDept. of MathematicsTemple [email protected]

Talk 4. Bounds on the eigenvalues of indefinite matricesarising from interior-point methodsInterior-point methods feature prominently in the solution ofconstrained optimization problems, and involve the need to solvea sequence of 3× 3 block indefinite matrices that becomeincreasingly ill-conditioned throughout the iteration. Mostsolution approaches are based on reducing the system size usinga block Gaussian elimination procedure. In this talk we useenergy estimates to obtain bounds on the eigenvalues of theoriginal, augmented matrix, which indicate that the spectralstructure of this matrix may be favorable compared to matricesobtained by performing a partial elimination of variables beforesolving the system.Chen GreifDepartment of Computer ScienceThe University of British [email protected]

Erin MouldingDepartment of MathematicsThe University of British [email protected]

Dominique OrbanMathematics and Industrial Engineering DepartmentEcole Polytechnique de [email protected]

MS 12. Application of statistics to numerical linearalgebra algorithms - Part II of IITalk 1. Spectral graph theory, sampling matrix sums, andnear-optimal SDD solversWe present a near-optimal solver for Symmetric DiagonallyDominant (SDD) linear systems. The solver is a great exampleof the power of statistical methods in linear algebra and inparticular of sampling sums of positive semi-definite matrices.Crucial to the speed of the solver is the ability to perform fastsampling. In turn, key to fast sampling are low-stretch spanningtrees, a central notion in spectral graph theory and combinatorialpreconditioning. We will see that a low-stretch tree is essentiallya ‘spectral spine’ of a graph Laplacian, allowing us to create ahierarchy of spectrally similar graphs which lends itself to thevery fast multilevel solver.Ioannis KoutisComputer Science DepartmentUniversity of Puerto [email protected]

Gary MillerComputer Science DepartmentCarnegie Mellon University, [email protected]

Richard PengComputer Science DepartmentCarnegie Mellon University, [email protected]

Talk 2. Implementation of a randomization algorithm fordense linear algebra librariesRandomization algorithms are an alternative to pivoting forfactoring dense matrices to solve systems of linear equations.These techniques are more suitable to parallel architectures

when compared to pivoting, requiring a reduced amount ofcommunication and no synchronization during the factorization.The core kernel requires the multiplication of sparse matriceswith a specific structure. The main issue in implementing suchalgorithms resides in the data dependency patterns, especiallyfor the symmetric case. An implementation for the PLASMAlibrary is described and traces are presented, together withperformance data.Dulceneia BeckerUniversity of Tennessee, [email protected]

Marc BaboulinInria Saclay Ile-de-FranceUniversity Paris-Sud, [email protected]

Jack DongarraUniversity of Tennessee, [email protected]

Talk 3. Implementing randomized matrix algorithms inlarge-scale parallel environmentsRecent work from theoretical computer science on developingrandomized matrix algorithms has recently led to thedevelopment of high-quality numerical implementations. Here,we describe our parallel iterative least-squares solver LSRN. Theparallelizable normal random projection in the preconditioningphase leads to a very well-conditioned system. Hence, thenumber of iterations is fully predictable if we apply LSQR or theChebyshev semi-iterative method to the preconditioned system.The latter method is particularly efficient for solving large-scaleproblems on clusters with high communication cost, e.g., onAmazon Elastic Compute Cloud clusters, that are increasinglycommon in large-scale data applications.Michael W. MahoneyDepartment of MathematicsStanford University, [email protected]

Talk 4. Random sampling preconditionersIn the talk we study the use of preconditioners based on randomsampling of rows for accelerating the solution of linear systems.We argue that the fusion of randomization with preconditioningis what enables fast and reliable algorithms.We will discuss both dense and sparse matrices. For densematrices, we will describe Blendenpik, a least-square solver fordense highly overdetermined systems that outperforms LAPACKby large factors, and scales significantly better than anyQR-based solver. For sparse matrices, we relate randomsampling preconditioners to fast SDD solvers, and generalizesome of the techniques to finite element matrices.Haim AvronMathematical Sciences DepartmentIBM T. J. Watson Research Center; [email protected]

Sivan ToledoDepartment of Computer ScienceTel-Aviv University, [email protected]

MS 13. Rational Krylov methods: analysis andapplications - Part II of IITalk 1. Rational Krylov methods for nonlinear matrixproblems

17 2012 SIAM Conference on Applied Linear Algebra

This talk is about the solution of non-linear eigenvalue problemsand linear systems with a nonlinear parameter. Krylov andRational Krylov methods are known to be efficient and reliablefor the solution of such matrix problems with a linear parameter.The nonlinear function can be approximated by a polynomial. Inearlier work, we suggested the use of Taylor expansions of thenonlinear function with an a priori undetermined degree. Thisled to the Taylor Arnoldi method that is an Arnoldi methodapplied to an infinite dimensional Companion linearization.When an interpolating polynomial is used instead of Taylorseries, there is a similar connection with the rational Krylovsubspace method applied on a linearization. The Krylovsubspaces enjoy similar properties as the linear case such asmoment matching and the convergence looks similar toconvergence for a linear problem. We present several choices ofpolynomials and also discuss ideas for future work.Karl MeerbergenDepartment of Computer [email protected]

Roel Van BeeumenDepartment of Computer [email protected]

Wim MichielsDepartment of Computer [email protected]

Talk 2. Block Gauss and anti-Gauss quadrature rules withapplication to networksThe symmetric and nonsymmetric block Lanczos processes canbe applied to compute Gauss quadrature rules for functionalswith matrix-valued measures. We show that estimates of upperand lower bounds for these functionals can be computedinexpensively by evaluating pairs of block Gauss and anti-Gaussrules. An application to network analysis is described.Lothar ReichelDept. of MathematicsKent State [email protected]

David MartinDept. of MathematicsKent State [email protected]

Talk 3. On optimality of rational Krylov based low-rankapproximations of large-scale matrix equationsIn this talk, we will discuss projection-based approximations forthe solution of Lyapunov equations of the form

AXET + EXAT +BBT = 0,

with A = AT , E = ET ∈ Rn×n and B ∈ Rn×m. Recently, arelation between minimizing the objective function

f :M→ R, X 7→ tr(XAXE +BBT

)on the manifoldM of symmetric positive semi-definite matricesof rank k and the L−norm defined by the operator

L := −E ⊗A−A⊗ E

together with the inner product 〈u, v〉L = 〈u,Lv〉 has beenshown. While so far this minimization problem was solved by

means of a Riemannian optimization approach, here we willdiscuss an interpolation-based method which leads to the sameresults but relies on projecting the original Lyapunov equationonto a rational Krylov subspace. It will turn out that this can beachieved by the iterative rational Krylov algorithm (IRKA).Besides a generalization for the case of A 6= AT , we will alsodiscuss an extension for more general equations of the form

AXE + FXB + CD = 0,

with A,F ∈ Rn×n, B,E ∈ Rm×m, C ∈ Rn×p andD ∈ Rp×m.Tobias BreitenComputational Methods in Systems and Control TheoryMax Planck Institute for Dynamics of Complex Technical [email protected]

Peter BennerComputational Methods in Systems and Control TheoryMax Planck Institute for Dynamics of Complex Technical [email protected]

Talk 4. Inverse problems for large-scale dynamical systems inthe H2-optimal model reduction frameworkIn this work we investigate the Rational Krylov subspace (RKS)projection method with application to the inverse problems. Wederive a representation for the reduced Jacobian as the productof a time-dependent and a stationary part. Then we show that theRKS satisfying the Meier-Luenberger necessary H2 optimalitycondition not only minimizes the approximation error butcompletely annuls its influence on the inversion result (even ifthe subspace is not optimal globally). More precisely, theapproximation error belongs to the (left) null-space of thereduced Jacobian. We compare inversion on such subspacesusing other nearly optimal RKS’s based on Zolotarev problemand adaptive pole selection algorithm.Mikhail ZaslavskySchlumberger-Doll [email protected]

Vladimir DruskinSchlumberger-Doll [email protected]

Valeria SimonciniDepartment of MathematicsUniversity of Bologna, [email protected]

MS 14. New trends in tridiagonal matrices - Part IIof IITalk 1. On generalized Jacobi matrices which are symmetricin Krein spacesI will give some motivations to introduce generalized Jacobimatrices of a special type. Then, some direct and inverseproblems for these matrices will be presented.Maxim DerevyaginInstitut fur MathematikTechnische Universitat BerlinSekretariat MA 4-5Straße des 17. Juni 13610623 [email protected]

Talk 2. On the characteristic function for Jacobi matricesFor a certain class of infinite Jacobi matrices with a discretespectrum it is introduced a characteristic function as an analytic

2012 SIAM Conference on Applied Linear Algebra 18

function on a suitable complex domain. It is shown that the zeroset of the characteristic function actually coincides with thepoint spectrum. As an illustration several explicitly solvableexamples are discussed.Pavel StovıcekDepartment of MathematicsFaculty of Nuclear Sciences and Physical EngineeringCzech Technical UniversityTrojanova 13120 00 PragueCzech [email protected]

Talk 3. Tridiagonal matrices in comb filtersIn the present talk we use some known and new results ontridiagonal matrices to compute the minimum mean square errorfor a decision feedback equalizer when the filter is a comb filter.We also apply those mathematical results on tridiagonal matricesto a problem in linear estimation with comb filters.Jesus Gutierrez-GutierrezCEIT and TecnunUniversidad de NavarraPaseo de Manuel Lardizabal, no. 1320018 Donostia - San [email protected]

Talk 4. The nullity theorem: forecasting structures in theinverses of sparse matricesThe Nullity Theorem for matrices was formulated by Markhamand Fiedler in 1986. At first sight, this theorem and its five-lineproof appear trivial. Examples show, however, that it is powerfultool in computations with sparse and structured rank matrices. Itis, e.g., a straightforward consequence of this theorem that theinverse of a tridiagonal is semiseparable, the inverse of a bandmatrix is higher order semiseparable, without even posingconstraints of irreducibility.We will reconsider the nullity theorem and extend it topredicting structures in LU and QR-factorizations.Raf VandebrilK.U. LeuvenDepartement Computerwetenschappen (Postbus 02402)Celestijnenlaan 200A3001 [email protected]

MS 15. Application of compressed sensing inBio-MedicineTalk 1. Evaluation of compressed sensing impact in cardiacsignals processing and transmissionSensor networks are a field which benefits significantly fromcompressed sensing (CS), as, for instance, quality of service, orbattery lifetime are directly improved by the subsequentreduction of traffic. Networks transmitting biomedical data, andwithin those, cardiac data, have particular advantages. ECG andPPG are the most studied 1D signals in the biomedical field, andwhere CS implementation has been more studied. One studyscenario is using the sparsity of wavelet decomposition of thesesignals, and Iterative Shrinkage/Thresholding algorithms. Focusis given in accuracy and computation overhead, the most criticalconstraints of cardiac data processing. Networking implicationsare quantified, for different types of real network models, as wellas the impact in signals quality, and on the extracted information

(heart rate, oxygen saturation, pulse wave velocity).Eduardo PinheiroInstituto de TelecomunicaesInstituto Superior Tcnico, Lisboa, [email protected]

Talk 2. Compressive sensing in drug discoveryThe design of new drugs is often based on certain interactionsbetween a small molecule (the drug) and a protein. Often a highaffinity between the small molecule and the protein is desirable.The affinity can be estimated by using statisticalthermodynamics. In order to provide statistical data,high-performance computing machines are used to explore theconformational space of the molecular system. Thecorresponding algorithms are based on simplified physicalmodels (force fields) for the interaction between the drugmolecule and the protein. These models mostly do not takequantum effects into account. The talk will present new ideas tobetter include quantum chemical information, it will providesome important application areas of molecular simulation indrug discovery, and it gives some hints about using compressivesensing in this field.Marcus WeberZuse Institute Berlin (ZIB), [email protected]

Talk 3. Reconstruction of bacterial communities using sparserepresentationDetermining the identities and frequencies of species present ina sample is a central problem in metagenomics, with scientific,environmental and clinical implications. A popular approach tothe problem is sequencing the Ribosomal 16s RNA gene in thesample using universal primers, and using variation in the gene’ssequence between different species to identify the speciespresent in the sample. We present a novel framework forcommunity reconstruction, based on sparse representation; whilemillions of microorganisms are present on earth, with known 16ssequences stored in a database, only a small minority (typically afew hundreds) are likely to be present in any given sample. Wediscuss the statistical framework, algorithms used and results interms of accuracy and species resolution.Or ZukBroad Institute, Cambridge, [email protected]

Talk 4. Sensing genome via factorizationSince the last decade the matrix factorization and dimensionreduction methods have been used for mining in genomics andproteomics. Also most recently the norm one optimization isemerging as a useful approach for this purpose. We propose ahybrid algorithm which uses the utility of SVD along withefficiency and robustness of norm one minimization forovercoming the shortcomings of each of these mentionedapproaches.Amir NiknejadCollege of Mount Saint Vincent, New York, [email protected]

MS 16. Preconditioning of non-normal linear systemsarising in scattering problemsTalk 1. Approximate deflation preconditioning methods forpenetrable scattering problemsIn this talk I consider the Lippmann-Schwinger (LS) integralequation for inhomogeneous acoustic scattering. I demonstrate

19 2012 SIAM Conference on Applied Linear Algebra

that spectral properties of the LS equations suggest that deflationbased preconditioning might be effective in accelerating theconvergence of a restarted GMRES method. Much of theconvergence results on deflation based preconditioning are basedon using exact invariant subspaces. I will present an analyticalframework for convergence theory of general approximatedeflation that is widely applicable. Furthermore, numericalillustrations of the spectral properties also reveal that asignificant portion of the spectrum is approximated well oncoarse grids. To exploit this, I develop a novel restarted GMRESmethod with adaptive preconditioning based on spectralapproximations on multiple grids.Josef SifuentesCourant Institute of Mathematical SciencesNew York [email protected]

Mark EmbreeDepartment of Computational and Applied MathematicsRice [email protected]

Talk 2. Direct approximate factoring of the inverseTo precondition a large and sparse linear system, two directmethods for approximate factoring of the inverse are devised.The algorithms are fully parallelizable and appear to be morerobust than the iterative methods suggested for the task. Amethod to compute one of the matrix subspaces optimally isderived. These approaches generalize the approximate inversepreconditioning techniques in several natural ways.Marko HuhtanenDepartment of Electrical and Information EngineeringUniversity of [email protected]

Mikko [email protected]

Talk 3. Regularization of singular integral operators as apreconditioning strategyIn this talk we consider the singular integral equation arising inelectromagnetic scattering on inhomogeneous objects in freespace. To understand the apparent poor convergence of iterativemethods, such as GMRES, we analyze the spectrum of both theintegral operator and the system matrix. It turns out that theconvergence is slowed down by dense spectral clusters, whoseorigin and location may be traced to the nontrivial essentialspectrum of the integral operator. We propose a multiplicativeregularizer which reduces the extent of these clusters and inconjunction with the deflation of largest-magnitude eigenvaluesmakes for a good preconditioner.Neil BudkoNumerical Analysis, Delft Institute of Applied MathematicsDelft University of [email protected]

Grigorios ZourosSchool of Electrical and Computer EngineeringNational Technical University of [email protected]

Talk 4. High-order shifted Laplace preconditioners for waveequationsShifted Laplace preconditioners techniques have beenintroduced by Erlangga, Vuik and Oosterlee (Applied NumericalMathematics, 2004) for solving PDEs problems related towave-like equations. The aim of this talk is to propose a

generalization of shifted Laplace preconditioning methods byusing operator representation combined with complex Padeapproximants. We will show that the resulting high-order shiftedLaplace preconditioners are highly efficient and robust for two-and three-dimensional scattering problems that exhibit complexgeometrical features (e.g. resonant structures). Furthermore, theconvergence is proved to be weakly frequency dependent.Xavier AntoineInstitut Elie Cartan de NancyUniversity of [email protected]

Christophe GeuzaineInstitut MontefioreUniversity of [email protected]

Ibrahim ZangreInstitut Elie Cartan de NancyUniversity of [email protected]

MS 17. Markov ChainsTalk 1. Markov Chain properties in terms of column sums ofthe transition matrixQuestions are posed regarding the influence that the columnsums of the transition probabilities of a stochastic matrix (withrow sums all one) have on the stationary distribution, the meanfirst passage times and the Kemeny constant of the associatedirreducible discrete time Markov chain. Some new relationships,including some inequalities, and partial answers to the questions,are given using a special generalized matrix inverse that has notpreviously been considered in the literature on Markov chains.Jeffrey J. HunterSchool of Computing and Mathematical SciencesAuckland University of Technology, Auckland, New [email protected]

Talk 2. Hamiltonian cycle problem and Markov chainsWe consider the famous Hamiltonian cycle problem (HCP)embedded in a Markov decision process (MDP). Morespecifically, we consider the HCP as an optimization problemover the space of either stationary policies, or of occupationalmeasures induced by these stationary policies. This approachhas led to a number of alternative formulations and algorithmicapproaches involving researchers from a number of countries.These formulations exploit properties of a range of matrices thatusually accompany investigations of Markov chains. It will beshown that when these matrices are induced by the given graphsome of the graph’s underlying structures can be detected intheir matrix analytic properties.Jerzy FilarSchool of Computer Science, Engineering and MathematicsFlinders University, Bedford Park, SA, [email protected]

Talk 3. Inequalities for functions of transition matricesConsider an irreducible finite-state Markov chain on N states. Itis known that the N by N matrix vanishing on the diagonal andwhich is equal to the mean first passage matrix elsewhere, isinvertible. We prove that if N is greater than two, the diagonalentries of the inverse are negative, and obtain some relatedinequalities. Analogous results for a closely related matrix arederived as well. Joint with Michael Neumann and OlgaPryporova.

2012 SIAM Conference on Applied Linear Algebra 20

Iddo Ben-AriDepartment of MathematicsUniversity of Connecticut, Storrs, CT, United States of [email protected]

Talk 4. Compartmental systems and computation of theirstationary probability vectorsTo compute stationary probability vectors of Markov chainswhose transition matrices are cyclic of index p may be a difficulttask if p becomes large. A class of iterativeaggregation/disaggregation methods (IAD) is proposed toovercome the difficulty. It is shown that the rate of convergenceof the proposed IAD processes is governed by the maximalmodulus of the eigenvalues laying out of the peripheral spectrumof the smoothing matrix. The examined generators of Markovchains come from compartmental systems and cause that thetransition matrices under consideration may depend upon theappropriate stationary probability vectors. The nonlinearityrepresents further difficulties in computation.Ivo MarekSchool of Civil EngineeringCzech Institute of Technology, Praha, Czech [email protected]

MS 18. Preconditioning for PDE-constrainedoptimization - Part I of IITalk 1. Structural spectral properties of symmetric saddlepoint problemsSymmetric and indefinite block structured matrices often ariseafter the discretization of a large variety of application problems,where the block form stems from the presence of more than onepartial differential equation (PDE) in the problem, or from theimposition of some constraints, usually associated with PDEs.Structure-aware preconditioning strategies have emerged aswinning devises for efficiently and optimally solving theassociated large linear systems. In some relevant applications,the coefficient matrix shows a particular spectral symmetryaround the origin, which should be taken into account whenselecting both the iterative system solver and the mostappropriate preconditioner. In this talk we analyze in detail thissymmetry, and discuss its consequences in the solution of threemodel problems in optimal control, distributed time-harmonicparabolic control and distributed time-harmonic Stokes control.Valeria SimonciniDipartimento di MatematicaUniversita di Bologna, [email protected]

Wolfgang KrendlDoctoral Program Computational MathematicsJohannes Kepler University, Linz, [email protected]

Walter ZulehnerInstitute of Computational MathematicsJohannes Kepler University, 4040 Linz, [email protected]

Talk 2. Preconditioned iterative methods for Stokes andNavier-Stokes control problemsThe development of iterative solvers for PDE-constrainedoptimization problems is a subject of considerable recent interestin numerical analysis. In this talk, we consider the numericalsolution of two such problems, namely those of Stokes controland Navier-Stokes control. We describe the role of saddle pointtheory, mass matrix and Schur complement approximation,

multigrid routines, and commutator arguments, in theconstruction of solvers for these problems. We also discussissues involved with regularization terms, and, in the case ofNavier-Stokes control, the outer iteration employed. We displaynumerical results that demonstrate the effectiveness of oursolvers.John PearsonMathematical InstituteUniversity of [email protected]

Talk 3. Preconditioners for elliptic optimal control problemswith inequality constraintsIn this talk we consider optimal control problems with an ellipticboundary value problem as state equation. Based on recentresults for distributed elliptic optimal control problems withoutinequality constraints we present and discuss preconditioners forthe optimality system in the presence of additional controlconstraints or state constraints. The focus is on the performanceof these preconditioners within a Krylov subspace method withrespect to model and discretization parameters.A possible extension of the presented approach to optimalcontrol problems with other state equations, like the steady-stateStokes equations, is discussed.Walter ZulehnerInstitute of Computational MathematicsJohannes Kepler University, Linz, [email protected]

Markus KollmannDoctoral Program Computational MathematicsJohannes Kepler University, Linz, [email protected]

Talk 4. Nearly optimal block preconditioners for blocktwo-by-two linear systemsFor a class of block two-by-two systems of linear equations, weconstruct block preconditioning matrices and discuss theeigen-properties of the corresponding preconditioned matrices.The block preconditioners can be employed to accelerate theconvergence rates of Krylov subspace iteration methods such asMINRES and GMRES. Numerical experiments show that theblock preconditioners are feasible and effective when used insolving the block two-by-two linear systems arising from thefinite-element discretizations of a class of PDE-constrainedoptimization problems.Zhong-Zhi BaiInstitute of Computational MathematicsAcademy of Mathematics and Systems ScienceChinese Academy of SciencesP. O. Box 2719, Beijing 100190, [email protected]

MS 19. Matrices and graphs - Part I of IITalk 1. (0,1) matrices and the analysis of social networksIn the sociology literature, an m× n (0, 1) matrix is called anactor–event matrix. In many examples, the matrix A is uniquelydetermined by AAT , ATA, and the fact that A is (0, 1). In thistalk, we present a result on pairs of actor–event matrices that aredistinct but ‘close’, and yield the same AAT and ATA.Specifically, using techniques from combinatorial matrix theory,we characterise all pairs of (0, 1) matrices A1, A2 such thatA1A

T1 = A2A

T2 , A

T1 A1 = AT2 A2, and the (0, 1,−1) matrix

A1−A2 has at most one 1 and one−1 in each row and column.

21 2012 SIAM Conference on Applied Linear Algebra

Steve KirklandHamilton InstituteNational University of Ireland [email protected]

Talk 2. Necessary and sufficient conditions for a HamiltoniangraphA graph is singular if the zero eigenvalue is in the spectrum of its0-1 adjacency matrix A. If an eigenvector belonging to the zeroeigenspace of A has no zero entries, then the singular graph issaid to be a core graph. A (κ, τ)-regular set is a subset of thevertices inducing a κ-regular subgraph such that every vertex notin the subset has τ neighbours in it. We consider the case whenκ = τ which relates to the eigenvalue zero under certainconditions. We show that if a regular graph has a (κ, κ)-regularset, then it is a core graph. By considering the walk matrix, wedevelop an algorithm to extract (κ, κ)-regular sets and formulatea necessary and sufficient condition for a graph to beHamiltonian.Irene ScirihaDept of Mathematics, Faculty of ScienceUniv. of [email protected]

Domingos Moreira CardosoDepartamento de MatemticaUniv. de [email protected]

Talk 3. On the eigenvalues of symmetric matrices associatedwith graphsLet G be a weighted graph of order n. It is known that λ1 ≥ d1and λ2 ≥ d3 where λi and di are the ith largest Laplacianeigenvalue and degree of G, respectively. For λ3 and smallereigenvalues we show that it is possible to construct a weightedgraph G of any order n such that λj < dn.To obtain these results we show that for every k ≥ 2 andn ≥ k + 1 there exists a weighted graph of order n whoseadjacency matrix has exactly k negative eigenvalues.This is done by obtaining some beautiful properties of the Schurcomplements of the adjacency matrix.Miriam FarberDepartment of [email protected]

Talk 4. An extension of the polytope of doubly stochasticmatricesWe consider a class of matrices whose row and column sumvectors are majorized by given vectors b and c, and whoseentries lie in the interval [0, 1]. This class generalizes the class ofdoubly stochastic matrices. We investigate the correspondingpolytope Ω(b|c) of such matrices. Main results include ageneralization of the Birkhoff - von Neumann theorem and acharacterization of the faces, including edges, of Ω(b|c).Richard A. BrualdiDepartment of MathematicsUniversity of [email protected]

Geir DahlCenter of Mathematics for ApplicationsDepartments of Mathematics and InformaticsUniversity of [email protected]

MS 20. Tensor based methods for high dimensionalproblems in scientific computing - Part I of II

Talk 1. Optimal a priori Tensor Decomposition for theSolution of High Dimensional ProblemsTensor-based methods are receiving a growing attention for theiruse in high dimensional applications in scientific computingwhere functions of multiple parameters have to beapproximated. These methods are based on the construction oftensor decompositions that provide approximate representationsof functions on adapted low dimensional reduced bases. Here,we propose algorithms that are able to directly construct anapproximation of optimal tensor decompositions with respect toa desired metric, without a priori information on the solution.Connections with optimal model reduction will be discussed.Anthony NouyLUNAM Universite, GeM, UMR CNRS 6183, Ecole Centrale deNantes, Universit de Nantes, [email protected]

Talk 2. Application of the Proper Generalized Decomposition(PGD) to 3D cracked plates and estimation of thediscretization errorWe present a new numerical approach to solve cracked plates inan efficient way by obtaining a single solution in which thePoisson’s ratio ν and the plate thickness B are non-fixedparameters. With a single analysis, it is possible to obtain anumerical solution for a cracked plate problem that can beparticularized for any ν and B by simple post-processing, thusproviding an efficient tool for the numerical analysis of theseproblems. The method makes use of the Proper GeneralizedDecomposition (PGD). In this method, the sought displacementfield u(x, y, z, ν, B) is written as a series expansion of functionproducts. The formulation proposed here uses a 2Ddiscretization for the plane XY and a 1D discretization in theZ-direction. In addition, independent functions of ν and B areincluded in the tentative solution to be converged. An iterativeprocess is then carried out by which the necessary terms of theseries expansion are added in order to attain the requiredaccuracy. We have also implemented a recovery type errorestimator to evaluate the discretization error in energy norm bothin the XY plane (2D problem) and along the Z-direction (1Dproblem). The combination of both errors provides the localerror distribution that can be used to drive the h-adaptiverefinement.Eugenio GinerCentro Investigacion Tecnologıa de Vehıculos - CITV, U. PolitecnicaValencia, [email protected]

Chady GhnatiosEADS Chair at Ecole Centrale de Nantes, [email protected]

Brice BognetEADS Chair at Ecole Centrale de Nantes, [email protected]

Adrien LeygueEADS Chair at Ecole Centrale de Nantes, [email protected]

Juan Jose RodenasCentro Investigacion Tecnologıa de Vehıculos - CITV, U. PolitecnicaValencia, [email protected]

F. Javier FuenmayorCentro Investigacion Tecnologıa de Vehıculos - CITV, U. PolitecnicaValencia, [email protected]

Francisco Chinesta

2012 SIAM Conference on Applied Linear Algebra 22

EADS Chair at Ecole Centrale de Nantes, [email protected]

Talk 3. A tensor calculus approach for Bezier shapedeformationThe goal of our research is to obtain a computationally efficientdeformation of a parametric curve. This research is applied toengineering applications as, for example, mobile robotics andliquid composite moulding. A new technique was developed tomodify a Bezier curve by minimizing the changes of its shape.This shape modification is computed through a field of vectors,applying them on particular points of the Bezier curve (BezierShape Deformation, BSD). In these kinds of applications thecomputational cost is a critical issue because the parametriccurve must be modified in real time. Therefore, the original BSDalgorithm has been reformulated using algebra of tensors toreduce the computational time (Tensor-Bezier ShapeDeformation, T-BSD). While BSD grows exponentially, T-BSDgrows linearly. As a consequence, the Bezier curve is modifiedwith the lowest computational cost.Lucıa HilarioDepartamento de Ciencias Fısicas, Matematicas y de la Computacion,Universidad CEU Cardenal Herrera, [email protected]

Nicolas MontesDepartamento de Ingenierıa de la Edificacion y Produccion Industrial,Universidad CEU Cardenal Herrera ,[email protected]

Marta C. MoraDepartamento de Ingenierıa Mecanica y Construccion, UniversitatJaume I , [email protected]

Antonio FalcoDepartamento de Economıa Financiera y Contabilidad, Universidad deAlicante, [email protected]

Talk 4. Tensor approximation methods for parameteridentificationIn this talk we present methods for parameteridentification—here the coefficient fields of partial differentialequations—based on Bayesian procedures. As is well known,Bayes’s theorem for random variables (RVs) of finite variance isan orthogonal projection onto a space of functions generated bythe observations. Both the quantity to be identified as well as thespace to project onto are described as functions of known RVs.All this takes place in high-dimensional tensor product spaces.To limit the numerical effort, we use low-rank tensorapproximations for the forward as well as the inverse problem.Hermann G. MatthiesInstitute of Scientific ComputingTU [email protected]

Alexander LitvinenkoInstitute of Scientific ComputingTU [email protected]

Bojana RosicInstitute of Scientific ComputingTU [email protected]

Oliver PajonkInstitute of Scientific ComputingTU [email protected]

MS 21. Reducing communication in linear algebra -Part I of IITalk 1. Communication-optimal parallel algorithm forStrassen’s matrix multiplicationParallel matrix multiplication is one of the most studiedfundamental problems in distributed and high performancecomputing. We obtain a new parallel algorithm that is based onStrassen’s fast matrix multiplication and minimizescommunication. The algorithm outperforms all other parallelmatrix multiplication algorithms, classical and Strassen-based,both asymptotically and in practice.A critical bottleneck in parallelization of Strassen’s algorithm isthe communication between the processors. Ballard, Demmel,Holtz, and Schwartz (SPAA’11) provide lower bounds on thesecommunication costs, using expansion properties of theunderlying computation graph. Our algorithm matches theselower bounds, and so is communication-optimal. It has perfectstrong scaling within the maximum possible range. Ourparallelization approach generalizes to other fast matrixmultiplication algorithms.Oded SchwartzComputer Science DivisionUC [email protected]

Grey BallardComputer Science DivisionUC [email protected]

James DemmelComputer Science DivisionUC [email protected]

Olga HoltzMath DepartmentUC Berkeley and TU [email protected]

Benjamin LipshitzComputer Science DivisionUC [email protected]

Talk 2. A communication-avoiding symmetric-indefinitefactorizationWe describe a communication-avoiding symmetric indefinitefactorization. The factorization works in two phases. In the firstphase, we factor A = LTLT where L is unit lower triangularand T is banded with bandwidth Θ(

√M) where M is the size

of the cache. This phase uses a block version of Aasen’salgorithm. In the second phase, we factor T ; this can be doneefficiently using Kaufman’s retraction algorithm, or LU withpartial pivoting, or successive band reduction. We show that thealgorithm is asymptotically optimal in terms of communication,is work efficient (n3/3 + o(n3) flops), and is backward stable.Sivan ToledoSchool of Computer ScienceTel-Aviv [email protected]

Grey BallardComputer Science DivisionUC [email protected]

James DemmelComputer Science DivisionUC [email protected]

23 2012 SIAM Conference on Applied Linear Algebra

Alex DruinskySchool of Computer ScienceTel-Aviv [email protected]

Inon PeledSchool of Computer ScienceTel-Aviv [email protected]

Oded SchwartzComputer Science DivisionUC [email protected]

Talk 3. LU factorisation with panel rank revealing pivotingand its communication avoiding versionWe present the LU decomposition with panel rank revealingpivoting (LU PRRP), an LU factorization algorithm based onstrong rank revealing QR panel factorization. LU PRRP is morestable than Gaussian elimination with partial pivoting (GEPP).Our extensive numerical experiments show that the newfactorization scheme is as numerically stable as GEPP inpractice, but it is more resistant to pathological cases and easilysolves the Wilkinson matrix and the Foster matrix.We also present CALU PRRP, a communication avoidingversion of LU PRRP that minimizes communication.CALU PRRP is more stable than CALU, the communicationavoiding version of GEPP.Amal KhabouLaboratoire de Recherche en Informatique Universite Paris-Sud 11,INRIA Saclay- Ile de [email protected]

James W. DemmelComputer Science Division and Mathematics Department, UC Berkeley,CA 94720-1776, [email protected]

Laura GrigoriINRIA Saclay - Ile de France, Laboratoire e Recherche en InformatiqueUniversite Paris-Sud 11, [email protected]

Ming GuMathematics Department, UC Berkeley, CA 94720-1776, [email protected]

Talk 4. 2.5D algorithms for parallel dense linear algebraWe give a ‘2.5D’ parallelization scheme for dense linear algebracomputations. Our scheme exploits any extra available memoryadaptively. The scheme generalizes previous work on 2D and 3Dalgorithms in matrix multiplication. Further, we extend 2.5Dalgorithms to LU, Cholesky, and Cholesky-QR factorizations,triangular solve, and the all-pairs-shortest-path problem. Thesenew algorithms perform asymptotically less communication thanany previous counterparts. 2.5D algorithms are also practicaland map well to torus networks. We benchmark 2.5D matrixmultiplication and LU factorization on 65,536 cores of a BlueGene/P supercomputer. Our results demonstrate that memoryreplication improves strong scalability and overall efficiencysignificantly.Edgar SolomonikDept. of Compiter ScienceUniversity of California, [email protected]

James DemmelDept. of Compiter Science and MathematicsUniversity of California, [email protected]

MS 22. Linear algebra for inverse problems -Part I of IITalk 1. Block-extrapolation methods for linear matrixill-posed problemsIn this talk, we consider large-scale linear discrete ill-posedproblems where the right-hand side contains noise. In manyapplications such as in image restoration the coefficient matrix isgiven as a sum of Kronecker products of matrices and thenTikhonov regularization problem leads to a generalized linearmatrix equation. We define some block extrapolation methodsand show how they could be used in combination with thesingular value decomposition (SVD) to solve these problems. Inparticular, the proposed methods will be applied in imagerestoration. Some theoretical results and numerical tests inimage restoration are also given.Khalide JbilouLMPA, Universite du Littoral,50, rue F. Buisson, BP 699,62228 CALAIS-Cedex, [email protected]

Talk 2. Convergence properties of GMRES and RRGMESmethod for ill posed problemsRange restricted iterative methods (RRGMRES) based on theArnoldi process are attractive for the solution of largenonsymmetric linear discrete ill-posed problems witherror-contaminated data (right-hand side).GMRES is one of the most widely used iterative methods for thesolution of linear system of equations, with a large real orcomplex nonsingular matrix. Convergence properties ofGMRES are discussed by many authors.We present some theoretical results for GMRES and RRGMRESand we give some numerical tests.Hassane SadokLMPA, Universite du Littoral,50, rue F. Buisson, BP 699,62228 CALAIS-Cedex, [email protected]

Lothar ReichelDepartment of Mathematical SciencesKent State [email protected]

Talk 3. Inverse problems for regularization matricesDiscrete ill-posed problems are difficult to solve, because theirsolution is very sensitive to errors in the data and to round-offerrors introduced during the solution process. Tikhonovregularization replaces the given discrete ill-posed problem by anearby penalized least-squares problem whose solution is lesssensitive to perturbations. The penalization term is defined by aregularization matrix, whose choice may affect the quality of thecomputed solution significantly. We describe several inversematrix problems whose solution yields regularization matricesadapted to the desired solution. Numerical examples illustratethe performance of the regularization matrices determined.Silvia NoscheseDipartimento di Matematica “Guido Castelnuovo”SAPIENZA Universita di [email protected]

Lothar ReichelDepartment of Mathematical SciencesKent State [email protected]

Talk 4. Meshless regularization for the numerical

2012 SIAM Conference on Applied Linear Algebra 24

computation of the solution of steady Burgers-type equations

In this talk, we introduce a meshless method approximation forthe numerical solution of the Burgers-type steady equations. Thenumerical approximation to the exact solution is obtained fromthe exact values of the analytical solution on a finite set ofscattered data points in the interior of the domain. The exactvalues are in fact unkown and contamained by a noise.Regularization techniques are needed to obtain the numericalapproximation as the smoothing thin plate splines. This strategyleads to a nonlinear system which may be solved by using theTikhonov regularization method. The estimation of theregularization parameter is obtained by using the classical GCVor the L-curve criteria. We present some theoretical results andwe give some numerical tests.Abderrahman BouhamidiLMPA, Universite du Littoral,50, rue F. Buisson, BP 699,62228 CALAIS-Cedex, [email protected]

Mustapha HachedLMPA, Universite du Littoral,50, rue F. Buisson, BP 699,62228 CALAIS-Cedex, [email protected]

Khalide JbilouLMPA, Universite du Littoral,50, rue F. Buisson, BP 699,62228 CALAIS-Cedex, [email protected]

MS 23. Modern matrix methods for large scale dataand networksTalk 1. Nonlinear eigenproblems in data analysis and graphpartitioningIt turns out that many problems in data analysis and beyond havenatural formulations as nonlinear eigenproblems. In this talk Ipresent our recent line of research in this area. This incluces theefficient computation of nonlinear eigenvectors via ageneralization of the inverse power method and a general resultshowing that a certain class of NP-hard combinatorial problemssuch as balanced graph cuts or the (constrained) maximumdensity subgraph have tight relaxations as nonlineareigenproblems.Matthias HeinFaculty of Mathematics and Computer ScienceSaarland [email protected]

Talk 2. LSRN: a parallel iterative solver for strongly over- orunder-determined systemsWe develop a parallel iterative least squares solver named LSRNthat is based on random normal projection. It computes theunique minimum-length solution to minx∈Rn ‖Ax− b‖2, whereA ∈ Rm×n can be rank-deficient with m n or m n. Acan be a dense matrix, a sparse matrix, or a linear operator, andLSRN speeds up automatically on sparse matrices and fastoperators. The preconditioning phase consists of a randomnormal projection, which is embarrassingly parallel, and asingular value decomposition of sizedγmin(m,n)e ×min(m,n), where γ is moderately larger than1, e.g., γ = 2. We show that the preconditioned system iswell-conditioned with a strong concentration result on the

extreme singular values. Hence the number of iterations is fullypredictable if we apply LSQR or the Chebyshev semi-iterativemethod to the preconditioned system. The latter method isparticularly efficient for solving large-scale problems on clusterswith high communication cost. LSRN is also capable of handlingcertain types of Tikhonov regularization. Numerical results showthat LSRN outperforms LAPACK’s DGELSD on large-scaledense problems and MATLAB’s backslash (SuiteSparseQR) onsparse problems on a shared memory machine, and LSRN scaleswell on Amazon Elastic Compute Cloud (EC2) clusters.Xiangrui MengInstitute for Computational and Mathematical EngineeringStanford [email protected]

M. A. SaundersInstitute for Computational and Mathematical EngineeringStanford [email protected]

M. W. MahoneyStanford [email protected]

Talk 3. Solving large dense linear systems with covariancematricesGaussian processes are a fundamental tool in spatial/temporalstatistics, and they have broad applications to fields such asnuclear engineering and climate science. The maximumlikelihood approach for fitting a Gaussian model requires themanipulation of the covariance matrix through inversion,logarithmic action and trace computation, all of which posesignificant challenges for large dense matrices. We consider areformulation of the maximum likelihood through a stochasticapproximation framework, which narrows down the severallinear algebra challenges to solving a linear system with thecovariance matrix for multiple right-hand sides. In this talk I willpresent an iterative technique integrating the designs of thesolver, the preconditioner and the matrix-vector multiplications,which lead to a scalable method for solving the maximumlikelihood problem in data analysis.Jie ChenMathematics and Computer Science DivisionArgonne National [email protected]

Talk 4. Fast coordinate descent methods with variableselection for non-negative matrix factorizationNon-negative Matrix Factorization (NMF) is a dimensionreduction method for non-negative matrices, and has proven tobe useful in many areas, such as text mining, bioinformatics andimage processing. In this talk, I will present a new coordinatedescent algorithm for solving the least squares NMF problem.The algorithm uses a variable selection scheme that employs thegradient of the objective function. This new method isconsiderably faster in practice, especially when the solution issparse, as is often the case in real applications. In these cases,our method benefits by selecting important variables to updatemore often, thus resulting in higher speed. As an example, on atext dataset RCV1, our method is 7 times faster than FastHals,which is a cyclic coordinate descent algorithm, and more than 15times faster when the sparsity is increased by adding an L1penalty. We also develop new coordinate descent methods whenerror in NMF is measured by KL- divergence by applying theNewton method to solve the one-variable sub-problems.

25 2012 SIAM Conference on Applied Linear Algebra

Experiments indicate that our algorithm for minimizing theKL-divergence is faster than the Lee & Seung multiplicative ruleby a factor of 10 on the CBCL image dataset.Inderjit S. DhillonDepartment of Computer ScienceThe University of Texas at [email protected]

Cho-Jui HsiehDepartment of Computer ScienceThe University of Texas at [email protected]

MS 24. Novel and synergetic algorithms for multicoreand multinode architectureTalk 1. Novel and synergetic linear algebra algorithms onmulticore and multinode architectureThe advent of parallel computers from multiple CPUs to, morerecently, multiple processor cores within a single CPU hascontinued to spur innovative linear algebra algorithms. Thisminisymposium aims to present a number of works that not onlyexhibit new algorithmic ideas suitable for these moderncomputer architectures, but also present interesting synergiesamong themselves in several levels. This talk gives an overviewon how a set of such algorithms work together and sets the stagefor the details to follow in each of the subsequent presentations.Ping Tak Peter TangIntel [email protected]

Talk 2. Towards hybrid factorization methods for solvinglarge sparse systemsThe availability of large-scale computing platforms comprised ofthousands of multicore processors motivates the need for highlyscalable sparse linear system solvers. In this talk I will reviewsome recent work by researchers at Purdue University,University of Lugano, and Intel Corporation on the combineduse of direct and iterative methods for solving very large linearsystems of equations. We will present recent progress in the areaof parallel direct solvers and present a new parallel solver thatcombines the desirable characteristics of direct methods(robustness) and iterative solvers (computational efficiency),while alleviating their drawbacks (high memory requirements,lack of robustness).Olaf SchenkInstitute of Computational ScienceUniversity of [email protected]

Ahmed SamehDept. of Computer SciencePurdue [email protected]

Talk 3. Eigensolver-based reordering and parallel traceMINThe eigenvector corresponding to the second smallest eigenvalueof the Laplacian of a graph, known as the Fiedler vector, has anumber of applications in areas that include matrix reordering,graph partitioning, protein analysis, data mining, machinelearning, and web search. The computation of the Fiedler vectorhas been regarded as an expensive process as it involves solvinga large eigenvalue problem. We present an efficient parallelalgorithm for computing the Fiedler vector of large graphs basedon the Trace Minimization algorithm. We compare the parallelperformance of our method against one of the bestimplementations of a sequential multilevel scheme, designed

specifically for computing the Fiedler vector – routine MC73 ofthe Harwell Subroutine Library (HSL). In addition, we presentresults showing how such reordering: (i) enhances data localityfor sparse matrix-vector multiplication on parallel architectures,as well as (ii) helps in extracting effective and scalablepreconditioners.Murat ManguogluDepartment of Computer EngineeringMiddle East Technical University, 06800 Ankara, [email protected]

Faisal SaiedDepartment of Computer SciencePurdue University, West Lafayette, 47906 IN, [email protected], [email protected]

Ahmed H. SamehDepartment of Computer SciencePurdue University, West Lafayette, 47906 IN, [email protected], [email protected]

Talk 4. FEAST a density matrix based eigensolverThe FEAST algorithm takes its inspiration from the densitymatrix representation and contour integration technique inquantum mechanics. It combines simplicity and efficiency andoffers many important capabilities for achieving highperformance, robustness, accuracy, and scalability for solvinglarge sparse eigenvalue problems on parallel architectures.Starting from a random set of vectors, FEAST’s maincomputational tasks consist of solving a few complex-valuedindependent linear systems with multiple right-hand sides, andone reduced eigenvalue problem orders of magnitude smallerthan the original one. Accuracy can be systematically improvedthrough an iterative process that usually converges in two tothree iterations. In this talk, FEAST will first be presented as anouter-layer interface to any linear system solvers eithershared-memory or distributed such as PSPIKE. In addition, itwill be shown that a single FEAST iteration provides the bestsuitable eigen-subspace which can be used for initiating theTraceMIN algorithm.Eric PolizziDepartment of Electrical and Computer EngineeringUniversity of Massachusetts, Amherst, [email protected]

MS 25. Direction preserving and filtering methods forsolving sparse linear systemsTalk 1. Algebraic two-level domain decomposition methodsWe consider the solving of linear systems arising from porousmedia flow simulations with high heterogeneities. The parallelsolver is a Schwarz domain decomposition method. Theunknowns are partitioned with a criterion based on the entries ofthe input matrix. This leads to substantial gains compared to apartition based only on the adjacency graph of the matrix. Fromthe information generated during the solving of the first linearsystem, it is possible to build a coarse space for a two-leveldomain decomposition algorithm. We compare two coarsespaces: a classical approach and a new one adapted to parallelimplementation.Frederic NatafLaboratory J.L. LionsUniversity of Paris [email protected]

Pascal HaveIFP

2012 SIAM Conference on Applied Linear Algebra 26

Rueil [email protected]

Roland MassonLaboratoire J.A. DieudonneUniversity of [email protected]

Mikolaj [email protected]

Tao ZhaoLaboratory J.L. LionsUniversity of Paris [email protected]

Talk 2. Filtering solversIn this talk, we give an overview on filtering solvers. Thiscomprises frequency filtering and the smoothing correctionscheme, the tangential frequency filtering (TFFD) by Wagner,the two frequency decomposition by Buzdin et al. We thendiscuss the adaptive filtering method. The adaptive test vectoriterative method allows the combination of the tangentialfrequency decomposition and other iterative methods such asmulti-grid. We further discuss the Filtering Algebraic Multigrid(FAMG) method. Interface problems as well as problems withstochastically distributed properties are considered. Realisticnumerical experiments confirm the efficiency of the presentedalgorithms.Gabriel WittumG-CSC, [email protected]

Talk 3. Bootstrap algebraic multigridBy the time of its development Algebraic Multigrid (AMG) wasthought of as a black box solver for systems of linear equations.Yet, the classical formulation of AMG turned out to lack therobustness to overcome challenges encountered in many oftoday’s computational simulations which lead to thedevelopment of adaptive techniques in AMG methods.We present in this talk a Bootstrap approach to adaptive AMGintroducing the so-called “Least Squares Interpolation” (LSI)which allows to define interpolation operators solely based onprototypes of algebraically smooth error. Furthermore, weintroduce a “Bootstrap Setup” which enables us to computeaccurate LSI operators using a multigrid approach in the setup toefficiently compute the required prototypes of algebraicallysmooth error.We demonstrate the potential of the Bootstrap AMG approach inthe application to a variety of problems, each illustrating acertain aspect of the method.Karsten KahlDept. of Mathematics and Natural SciencesUniversity of [email protected]

James BrannickDept. of MathematicsPennsylvania State [email protected]

Talk 4. Block filtering decompositionWe describe a preconditioning technique that is suitable formatrices arising from the discretization of a system of PDEs onunstructured grids. The preconditioner satisfies a filteringproperty, which ensures that the input matrix is identical with thepreconditioner on a given filtering vector. This vector is chosento alleviate the effect of low frequency modes on convergence.

We present a general approach that allows to ensure that thefiltering condition is satisfied in a matrix decomposition. Theinput matrix can have an arbitrary sparse structure, and can bereordered using nested dissection to allow a parallelimplementation.Laura GrigoriINRIA SaclayUniversity of Paris 11, [email protected]

Frederic NatafDept. of Applied MathematicsUniversity of Paris 6, [email protected]

Riadh FezzanniINRIA SaclayUniversity of Paris 11, [email protected]

MS 26. Advances in Krylov subspace methodsTalk 1. The new challenges to Krylov subspace methodsSolution methods based on preconditioned Krylov subspacemethods have reached a certain level of maturity and research inthis area is now mostly concerned with issues related topreconditioning. This state of equilibrium is currently beingchallenged by the emerging computational architectures. Forexample, general purpose GPUs put a high cost on innerproducts on which Krylov methods rely heavily. At the sametime the many-core environment will make standard Krylovmethod unreliable due to the potentially high number of possiblehardware failures when tens or hundreds of thousands of coresare deployed. In this talk we will explore these challenges andsee what options are available to face them.Yousef SaadDept. of Computer Science and EngineeringUniversity of Minnesota, Twin Cities, [email protected]

Talk 2. Random shadow vectors in IDR(s): an explanation ofits GMRES-like convergenceThe IDR(s) method [Sonneveld and van Gijzen, SISC 31,1035-1062 (2008)], is a family of short-term recurrent Krylovsubspace solvers for large sparse, not necessarily symmetriclinear systems. For increasing s, the convergence behaviourshows an increasing similarity with full GMRES. The residualsof IDR(s) are, roughly speaking, orthogonalized with respect tos shadow vectors. Usually these are randomly chosen. Eachiteration step can be related to a Galerkin approximation in acorresponding Krylov subspace. In this talk we describe why thequality of this Galerkin approximation comes close to theoptimal (GMRES), if s is large enough.Peter SonneveldDelft Institute of Applied MathematicsDelft University of Technology, [email protected]

Talk 3. Truncated and inexact Krylov subspace methods forparabolic control problemsWe study the use of inexact and truncated Krylov subspacemethods for the solution of linear systems arising in thediscretized solution of optimal control of a parabolic partialdifferential equation. An all-at-once temporal discretization anda reduced Hessian approach are used. The solutions of the twolinear systems involved in this reduced Hessian can beapproximated, and in fact they can be less and less exact as the

27 2012 SIAM Conference on Applied Linear Algebra

iterations progress. The option we propose is the use of theparareal-in-time algorithm for approximating the solution ofthese two linear systems. Truncated methods can be usedwithout much delay in convergence, but with important savingsin storage. Spectral bounds are provided and numericalexperiments are presented, illustrating the potential of theproposed methods.Daniel B. SzyldDept. of MathematicsTemple University, [email protected]

Xiuhong DuDept. of MathematicsAlfred University, [email protected]

Marcus SarkisDept. of MathematicsWorcester Polytechnic Institute, [email protected]

Christian E. SchaererDept. of Computer ScienceNational University of Asuncion, [email protected]

Talk 4. Convergence of iterative solution algorithms forleast-squares problemsWe consider the iterative solution of linear least-squaresproblems. We ask the following questions: which quantitiesshould be used to monitor convergence, and how can thesequantities be estimated at every iteration of commonly usedalgorithms? We argue that the backward error, appropriatelyscaled, is an excellent measure of convergence. We show howcertain projections of the residual vector can be used to bound orestimate it. We present theoretical results and numericalexperiments on the convergence of the backward error and itsestimates in the algorithms LSQR of Paige and Saunders andLSMR of Fong and Saunders.David Titley-PeloquinMathematical InstituteUniversity of Oxford, [email protected]

Serge GrattonENSEEIHT, [email protected]

Pavel JiranekCERFACS, [email protected]

MS 27. Preconditioning for PDE-constrainedoptimization - Part II of IITalk 1. On linear systems arising in Trust-region methodsTrust-region methods are widely used for the numerical solutionof nonlinear constrained and unconstrained optimizationproblems. Similar as line-search methods, they take turns inminimizing local models of the objective. In contrast toline-search methods, however, trust-region approaches do notrequire nor promote the positive definiteness of the local Hessianapproximations.In this presentation, we address issues arising in the context ofthe preconditioned approximate solution of these local models.The talk will highlight connections between algorithmicoptimization and linear algebra.Susann MachDept. of Mathematics

University of Technology [email protected]

Roland HerzogDept. of MathematicsUniversity of Technology [email protected]

Talk 2. Preconditioning for PDE-constrained optimizationusing proper orthogonal decompositionThe main effort of solving a PDE constrained optimizationproblem is devoted to solving the corresponding large scalelinear system, which is usually sparse and ill conditioned. As aresult, a suitable Krylov subspace solver is favourable, if aproper preconditioner is embedded.Other than the commonlyused block preconditioners, we exploit knowledge of properorthogonal decomposition (POD) for preconditioning andachieve some interesting features. Numerical results onnonlinear test problems are proposed.Ekkehard SachsDepartment of Mathematics University of Trier 54286 Trier, [email protected]

Xuancan YeDepartment of Mathematics University of Trier 54286 Trier, [email protected]

Talk 3. Preconditioning for Allen-Cahn variationalinequalities with non-local constraintsThe solution of Allen-Cahn variational inequalities with massconstraints is of interest in many applications. This problem canbe solved in its scalar and vector-valued form as aPDE-constrained optimization problem. At the heart of thismethod lies the solution of linear systems in saddle point form.In this talk we propose the use of Krylov-subspace solvers andsuitable preconditioners for the saddle point systems. Numericalresults illustrate the competitiveness of this approach.Luise BlankDept. of MathematicsUniversity of [email protected]

Martin StollMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

Lavinia Sarbu

Talk 4. A one-shot approach to time-dependent PDE controlIn this talk, we motivate, derive and test effectivepreconditioners to be used with the MINRES algorithm forsolving a number of saddle point systems, which arise in PDEconstrained optimization problems. We consider the distributedcontrol and boundary control problems subject to partialdifferential equations such as the heat equation or Stokesequations. Crucial to the effectiveness of our preconditioners ineach case is an effective approximation of the Schur complementof the matrix system. In each case, we state the problem beingsolved, propose the preconditioning approach, and providenumerical results which demonstrate that our solvers areeffective for a wide range of regularization parameter values, aswell as mesh sizes and time-steps.Martin StollComputational Methods in Systems and Control TheoryMax Planck Institute for Dynamics of Complex Technical SystemsSandtorstr. 1, 39106 Magdeburg Germany

2012 SIAM Conference on Applied Linear Algebra 28

[email protected]

Andy WathenNumerical Analysis GroupMathematical Institute, 24–29 St Giles’, Oxford, OX1 3LB, [email protected]

John PearsonNumerical Analysis GroupMathematical Institute, 24–29 St Giles’, Oxford, OX1 3LB, [email protected]

MS 28. Matrices and graphs - Part II of IITalk 1. Parameters related to maximum nullity, zero forcingnumber, and tree-width of a graphTree-width, and variants that restrict the allowable treedecompositions, play an important role in the study of graphalgorithms and have application to computer science. The zeroforcing number is used to study the maximum nullity/minimumrank of the family of symmetric matrices described by a graph.Relationships between these parameters and several Colin deVerdiere type parameters, and numerous variations (includingthe minor monotone floors and ceilings of some of theseparameters) are discussed.Leslie HogbenDepartment of MathematicsIowa State [email protected]

Francesco BarioliDepartment of MathematicsUniversity of Tennessee at [email protected]

Wayne BarrettDepartment of MathematicsBrigham Young [email protected]

Shaun M. FallatDepartment of Mathematics and StatisticsUniversity of [email protected]

H. Tracy HallDepartment of MathematicsBrigham Young [email protected]

Bryan ShaderDepartment of MathematicsUniversity of [email protected]

P. van den DriesscheDepartment of Mathematics and StatisticsUniversity of [email protected]

Hein van der HolstDepartment of Mathematics and StatisticsGeorgia State [email protected]

Talk 2. Colin de Verdiere numbers of chordal and split graphs

I will discuss the Colin de Verdiere number of a chordal graph,showing that it always takes one of two possible values. For theimportant subcase of split graphs, I will present a completedistinction between the two cases. Finally, I will show how todeduce from the result on chordal graphs an improvement offactor 2 to Colin de Verdiere’s well-known tree-width upper

bound for the eponymous number. This work was done while theauthor was at Technion.Felix GoldbergHamilton InstituteNational University of Ireland [email protected]

Abraham BermanDepartment of [email protected]

Talk 3. Kochen-Specker sets and the rank-1 quantumchromatic numberThe quantum chromatic number of a graph G is sandwichedbetween its chromatic number and its clique number, which arewell known NP-hard quantities. We restrict our attention to therank-1 quantum chromatic number χ(1)

q (G), which upperbounds the quantum chromatic number, but is defined understronger constraints. We study its relation with the chromaticnumber χ(G) and the minimum dimension of orthogonalrepresentations ξ(G). It is known thatξ(G) ≤ χ(1)

q (G) ≤ χ(G). We answer three open questionsabout these relations: we give a necessary and sufficientcondition to have ξ(G) = χ

(1)q (G), we exhibit a class of graphs

such that ξ(G) < χ(1)q (G), and we give a necessary and

sufficient condition to have χ(1)q (G) < χ(G). Our main tools

are Kochen-Specker sets, collections of vectors with atraditionally important role in the study of noncontextuality ofphysical theories, and more recently in the quantification ofquantum zero-error capacities. Finally, as a corollary of ourresults and a result by Avis, Hasegawa, Kikuchi, and Sasaki onthe quantum chromatic number, we give a family ofKochen-Specker sets of growing dimension.Simone SeveriniDepartment of Computer Science and Department of Physics &Astronomy, University College [email protected]

Giannicola ScarpaCentrum Wiskunde & Informatica (CWI)[email protected]

Talk 4. On the null vectors of graphsEigenvalues and eigenvectors associated with graphs have beenimportant areas of research for many years. My plan for this talkit to revisit a basic, yet important, result by Fiedler on the inertiaof acyclic matrices and discuss its impact on some previousrelated work and highlight some interesting new implications oncertain null vectors associated with a graph.Shaun M. FallatDepartment of Mathematics and StatisticsUniversity of [email protected]

MS 29. Tensor based methods for high dimensionalproblems in scientific computing - Part II of IITalk 1. A greedy algorithm for the convergence of a Laplacianoperators in the blind deconvolution problemIn this talk we present new results on the a greedy algorithm tostudy the convergence of the Fractional Blind Deconvolution ina Hilbert Space. We will show the theoretical results and theapplication to the artistic Restoration.Pantaleon David Romero SanchezDepartamento de Ciencias Fısicas, Matematicas y de la Computacion

29 2012 SIAM Conference on Applied Linear Algebra

Universidad CEU-Cardenal [email protected]

Vicente F. CandelaDepartamento de Matematica AplicadaUniversidad de [email protected]

Talk 2. Algorithms for approximate inverse of operators forpreconditioning systems of equations in tensor tormatWe here propose and analyze greedy algorithms for theapproximation of the inverse of an operator in tensor format.Algorithms are based on successive best approximations withrespect to non usual norms that makes possible thedecomposition of the inverse operator without any a prioriinformation. This approximate inverse is then used forpreconditioning iterative solvers and PGD algorithms for thesolution of high dimensional PDEs in tensor format. Theefficiency of the proposed preconditioner is illustrated onnumerical examples, where it is compared to otherpreconditioners.Loic GiraldiLUNAM Universite, GeM, UMR CNRS 6183, Ecole Centrale deNantes, Universite de Nantes, [email protected]

Anthony NouyLUNAM Universite, GeM, UMR CNRS 6183, Ecole Centrale deNantes, Universite de Nantes, [email protected]

Gregory LegrainLUNAM Universite, GeM, UMR CNRS 6183, Ecole Centrale deNantes, Universite de Nantes, [email protected]

Talk 3. Geometric structures in tensor representationsIn this talk we discuss about the geometric structures associatedwith tensor representations based in subspaces. We introduce ageneralization of the hierarchical Tucker format. Then we give aBanach manifold structure to the set of tensors in generalizedhierarchical Tucker format with fixed hierarchical rank. It allowsto extend the dynamical low-rank tensor approximationframework to tensor Banach spaces.Antonio FalcoDept. de Economıa Financiera y ContabilidadUniversidad de Alicante, [email protected]

Wolfgang HackbuschMax-Planck-Institut Mathematik in den NaturwissenschaftenLeipzig, [email protected]

Anthony NouyLUNAM Universite, GeM, UMR CNRS 6183, Ecole Centrale deNantes, Universite de Nantes, [email protected]

MS 30. Reducing communication in linear algebra -Part II of IITalk 1. Communication-avoiding sparse matrix-matrixmultiplicationSparse matrix-matrix multiplication is a key primitive for manyhigh performance graph algorithms as well as some linearsolvers, such as algebraic multigrid. There is a significant gapbetween communication costs of existing algorithms and knownlower bounds. We present new 1D and 2.5D algorithms thatperform less communication than existing algorithms. Thesealgorithms have different memory requirements, and they scale

differently with increasing matrix density and processor count.We also report on our performance results on large-scaleexperiments and our recent progress in obtaining tighter lowerbounds.Aydın BulucComputational Research DivisionLawrence Berkeley National [email protected]

Grey BallardComputer Science DivisionUC [email protected]

James DemmelComputer Science DivisionUC [email protected]

Laura GrigoriINRIA Saclay - Ile de France, Laboratoire e Recherche en InformatiqueUniversite Paris-Sud 11, [email protected]

Oded SchwartzComputer Science DivisionUC [email protected]

Talk 2. Improving the stability of communication-avoidingKrylov subspace methodsKrylov Subspace Methods (KSMs) are commonly used forsolving linear systems, eigenvalue problems, and singular valueproblems. Standard KSMs are communication-bound on moderncomputer architectures, due to required sparse matrix-vectormultiplication and projection operations in each iteration. Thismotivated s-step KSMs, which can use blocking strategies toincrease temporal locality, allowing an O(s) reduction incommunication cost. Despite attractive performance benefits,these variants are often considered impractical, as increasederror in finite precision can negatively affect stability. Wediscuss practical techniques for alleviating these problems ins-step methods while still achieving an asymptotic reduction incommunication cost.Erin CarsonDept. of Computer ScienceUniversity of California, [email protected]

Nicholas KnightDept. of Computer ScienceUniversity of California, [email protected]

James DemmelDept. of Computer Science, MathematicsUniversity of California, [email protected]

Talk 3. Hiding global synchronization latencies in Krylovmethods for systems of linear equationsIn Krylov methods, global synchronization due to reductionoperations for dot-products is becoming a bottleneck on parallelmachines. We adapt GMRES and CG such that this globalcommunication latency is completely overlapped by other localwork. This requires the global communication to be performedin a non-blocking or asynchronous way. To maintain stabilityeven at the strong scaling limit, different Krylov bases, likeNewton and Chebychev bases, can be used. Our performancemodel predicts large benefits for future exascale machines aswell as for current-scale applications such as solving the coarsestlevel of a multigrid hierarchy in parallel.

2012 SIAM Conference on Applied Linear Algebra 30

Pieter GhyselsDept. of Math and Computer ScienceUniversity of Antwerp, [email protected]

Wim VanrooseDept. of Math and Computer ScienceUniversity of Antwerp, [email protected]

Talk 4. Avoiding communication with hierarchical matricesWe show how to reorganize the construction of a Krylov basis[x,Ax, . . . , Asx] to asymptotically reduce data movement,when A is a hierarchical matrix. Our approach extends theblocking covers algorithm of Leiserson, Rao, and Toledo, andrequires that off-diagonal blocks of A are low-rank. Thisapproach enables communication-avoiding s-step Krylovsubspace methods with hierarchical preconditioners. We alsodiscuss extensions to multigrid and fast multipole methods.Nick KnightDept. of Electrical Engineering and Computer ScienceUniversity of California, [email protected]

Erin CarsonDept. of Electrical Engineering and Computer ScienceUniversity of California, [email protected]

James DemmelDept. of Electrical Engineering and Computer ScienceUniversity of California, [email protected]

MS 31. Linear algebra for inverse problems -Part II of IITalk 1. Implicit filtering methods for inverse problemsIn this talk we consider a nonlinear least squares framework tosolve separable nonlinear ill-posed inverse problems. It is shownthat with proper constraints and well chosen regularizationparameters, it is possible to obtain an objective function that isfairly well behaved. Although uncertainties in the data andinaccuracies of linear solvers make it unlikely to obtain a smoothand convex objective function, it is shown that implicit filteringoptimization methods can be used to avoid becoming trapped inlocal minima. An application to blind deconvolution is used forillustration.James G. NagyDept. of Mathematics and Computer ScienceEmory [email protected]

Anastasia CornelioDept. of MathematicsUniversity of Modena and Reggio [email protected]

Elena Loli PiccolominiDepartment of MathematicsUniversity of [email protected]

Talk 2. Iterative reconstruction methods for adaptive opticsThe image quality of large ground based astronomical telescopessuffers from turbulences in the atmosphere. Adaptive Optics is ahardware-based technique that corrects the influence of theturbulence. To this end, wavefront measurements from theincoming wavefront of several guide stars are used for thereconstruction of the turbulent layers in the atmosphere. Thereconstructed atmosphere is then used to compute the surface ofseveral deformable mirrors such that, in the reflected light, the

influence of the turbulence is removed. The reconstruction of theturbulence in the atmosphere is related to a limited angletomography problem and therefore severely ill posed. For thenew class of extremely large telescopes, the numerical task is toinvert a linear system with a dimension of approximately60.000× 10.000 every millisecond.We present iterative methods, based on Kaczmarz and cg, for thereconstruction of the layers. The methods will be evaluated indifferent function spaces. In particular, we will include themodeling of specific effects that are related to the use of laserguide stars. We will also demonstrate that our method are able tocompute the correcting shape of the deformable mirrors withinthe available time frame.Ronny RamlauIndustrial Mathematics InstituteKepler University Linz, [email protected]

Talk 3. Approximated nonstationary iterated Tikhonov withapplication to image deblurringIn this talk we present new iterative regularization methods,based on the approximation of nonstationary iterated Tikhonov.In particular we investigate the image deblurring problem, wherethe blurring matrix is not easily invertible with a lowcomputational cost, while an approximation with such propertyis available. This is for instance the case of block Toeplitzmatrices with Toeplitz blocks that can be well approximated byblock circulant matrices with circulant blocks matrices whichcan be diagonalized by two dimensional fast Fourier transforms.Matrices arising from the imposition of other boundaryconditions can be considered as well.A detailed analysis is proposed in the stationary case and wediscuss relations with preconditioned Landweber method andother known methods. A large numerical experimentation withdifferent boundary conditions, blurring phenomena and noiselevel is presented.Marco DonatelliDept. of Scienza e Alta TecnologiaUniversity of [email protected]

Martin HankeInstitute of MathematicsUniversity of [email protected]

Talk 4. On the Richardson-Lucy method for imagerestorationImage deconvolution problems with a symmetric point-spreadfunction arise in many areas of science and engineering. Theseproblems often are solved by the Richardson- Lucy method, anonlinear iterative method. In this talk we first show aconvergence result for the Richardson-Lucy method. The proofsheds light on why the method may converge slowly.Subsequently, we describe an iterative active set method thatimposes the same constraints on the computed solution as theRichardson- Lucy method. Computed examples show the lattermethod to yield better restorations than the Richardson- Lucymethod with less computational effort.Fiorella SgallariDept. of Mathematics- CIRAMUniversity of Bologna,Bologna, [email protected]

M. Kazim Khan

31 2012 SIAM Conference on Applied Linear Algebra

Dept. of Mathematical Sciences,Kent State University, Kent, OHIO, [email protected]

Serena MorigiDept. of Mathematics- CIRAMUniversity of Bologna,Bologna, [email protected]

Lothar ReichelDept. of Mathematical Sciences,Kent State University, Kent, OHIO, [email protected]

MS 32. Orderings in sparse matrix computationTalk 1. Orderings and solvers for “non-uniform sparsematrices”Sparse matrices from non-PDE problems come in rich varieties.We specifically address sparse matrices that have a non-uniformdistribution of nonzeros per row (power-law graphs are a popularexample that lead to these), which may be called “non-uniformsparse matrices.” We examine the appropriateness of traditionalorderings to solve various problems for these types of matrices.For example, instead of RCM reordering for ILU(k), it is naturaland indeed better to use an ordering by increasing degree. Bystudying why such orderings are effective, we are lead to“incomplete” versions of minimum degree ordering, i.e., theordering operates like minimum degree, but the maximumdegree is capped by a parameter, just as fill-in level in ILU(k) iscapped. Lee, Raghavan, and Ng proposed such a technique forpivoting in 2006. In our studies, tie-breaking and thecomplementary strategies of reducing the number of fill-ins andthe aggregate size of fill-ins are important issues. Sparsematrices from PDEs have well-defined model problems; tofurther study non-uniform sparse matrices, we propose models(structure and values) derived from realistic applications.Edmond ChowSchool of Computational Science and EngineeringCollege of ComputingGeorgia Institute of [email protected]

Oguz KayaSchool of Computational Science and EngineeringCollege of ComputingGeorgia Institute of [email protected]

Talk 2. On hypergraph partitioning based ordering methodsfor sparse matrix factorizationWe discuss the use of hypergraph-based methods for orderingsof sparse matrices in Cholesky, LU and QR factorizations. Forthe Cholesky factorization case, we investigate a recent result onpattern-wise decomposition of sparse matrices, generalize theresult and develop effective ordering algorithms. Thegeneralized results help us formulate the ordering problem in LUas we do in the Cholesky case, without ever symmetrizing thegiven matrix. For the QR factorization case, the use ofhypergraph models is fairly standard; the method does notsymmetrize the given matrix. We present comparisons with themost common alternatives in all three cases.Bora UcarCNRS and ENS [email protected]

Iain S. DuffRAL Oxon, OX11 0QX, England and

CERFACS, Toulouse, [email protected]

Johannes LangguthENS [email protected]

Talk 3. Orderings governed by numerical factorizationWe study the solution of sparse least-squares problems using anaugmented system approach:(

I AAT 0

)(rx

)=

(b0

)If the null space approach for constrained optimization is used, acrucial aspect is the selection of the basis rows from theoverdetermined matrix A. We discuss the effect of this showingthat the concept of condition number and conditioning needs tobe rethought in this case. We illustrate our discussion with runsusing a basis selection routine from HSL that involves a sparsefactorization with rook pivoting and a subsequent solution of theaugmented system using iterative methods.Iain S. DuffRutherford Appleton [email protected]

Mario ArioliRutherford Appleton [email protected]

Talk 4. Reordering sparse Cholesky factorization: minimumfill vs. minimum FLOP countGiven a sparse positive definite matrix A, we discuss theproblem of finding a permutation matrix P such that the numberof floating point operations for computing the Choleskyfactorization PAPT = LLT is minimum. Two theoreticalresults are presented: First, we outline a reduction fromMAXCUT in order to give an NP-hardness result. Second, wediscuss the relationship of minimizing FLOPs and the minimumfill problem. Using an explicit construction we show thatorderings which minimize the operation count may benon-optimal in terms of the fill and vice versa.Robert LuceDept. of MathematicsTechnical University of [email protected]

Esmond G. NgLawrence Berkeley National [email protected]

MS 33. Moving from multicore to manycore in appliedlinear algebraTalk 1. Parallel preconditioners and multigrid methods forsparse systems on GPUsLarge-scale numerical simulation relies on both efficient parallelsolution schemes and platform-optimized parallelimplementations. Due to the paradigm shift towards multicoreand manycore technologies, both aspects have become moreintricate. In this talk we address fine-grained parallelpreconditioning techniques and multigrid solvers that arecompliant with SIMD-like parallelism of graphics processingunits. By means of multi-coloring reordering combined with thepower(q)-pattern method for incomplete LU decompositionswith fill-ins we show how scalable parallelism can be introducedto the triangular solution phase of smoothers andpreconditioners. Performance results demonstrate efficiency and

2012 SIAM Conference on Applied Linear Algebra 32

scalability of our approach on recent GPUs and onmulticore-CPUs.Jan-Philipp WeissEngineering Mathematics and Computing Lab (EMCL)Karlsruhe Institute of Technology (KIT), [email protected]

Dimitar LukarskiEngineering Mathematics and Computing Lab (EMCL)Karlsruhe Institute of Technology (KIT), [email protected]

Vincent HeuvelineEngineering Mathematics and Computing Lab (EMCL)Karlsruhe Institute of Technology (KIT), [email protected]

Talk 2. Towards a GPU-accelerated direct sparse solverWe present a detailed study about the acceleration of a directmethod to solve sparse linear systems using GPU-basedcomputing. Particularly, we describe a variant for thefactorization stage of the multifrontal method which combinesCPU and GPU computations. Additionally, the developedroutine is included in the CSparse library to cover the wholesolver. We evaluate the proposal and compare it with an ad-hocmulticore implementation and the implementation included inMUMPS library (combined with Goto BLAS) The resultsobtained show that this is a promising research line to acceleratethis kind of sparse matrix solvers on cheap hardware platforms.Pablo EzzattiInstituto de ComputacionUniv. de la Republica, Montevideo, [email protected]

Alejandro ZinemanasInstituto de ComputacionUniv. de la Republica, Montevideo, [email protected]

Talk 3. Exploiting the flexibility of libflame for novelmulti-core and many-core architecturesThe libflame library is a modern, high-performance denselinear algebra library that is extensible, easy to use, and availableunder an open source license, and offers competitive (and inmany cases superior) real-world performance when compared totraditional libraries like LAPACK. It can optionally exploitmultiple cores by casting an algorithm as an algorithm-by-blockswhich can generate a directed acyclic graph (DAG) that is thenscheduled to cores via a runtime system called SuperMatrix.In this talk we demonstrate its flexibility for running efficientlydense linear algebra codes on general-purpose architectures,including single core, multi-core and GPU-based architectures.In addition, we adapt our runtime to support a novel, highlyefficient HPC architecture: the Texas Instrument’s C6678multi-core DSP.Francisco D. IgualDept. de Ingenierıa y Ciencia de los ComputadoresUniversity Jaume I de [email protected]

Murtaza AliTexas [email protected]

Robert A. van de GeijnThe University of Texas at [email protected]

Talk 4. High-performance genome studiesIn the context of the genome-wide association study (GWAS),one has to solve long sequences of generalized least-squares

problems; such a problem presents two limiting factors:execution time –often in the range of days or weeks– and datamanagement –in the order of Terabytes–. We present algorithmsthat obviate both issues. By taking advantage of domain-specificknowledge, exploiting parallelism provided by multicores andGPU, and handling data effciently, our algorithms attainunequalled high performance. When compared to GenABEL,one of the most widely used libraries for GWAS, we obtainspeedups up to a factor of 50.Lucas BeyerAachen Institute for Advanced Study in Computational EngineeringScience (AICES)RWTH [email protected]

Diego Fabregat-TraverAICESRWTH [email protected]

Paolo BientinesiAICESRWTH [email protected]

MS 34. Least squares methods and applicationsTalk 1. Block Gram–Schmidt algorithms withreorthogonalizationThe talk discusses reorthogonalized block classicalGram–Schmidt algorithms for factoring a matrix A intoA = QR where Q is left orthogonal (has orthonormal columns)and R is upper triangular. These algorithms are useful indeveloping BLAS–3 versions of orthogonal factorization in theimplementation of Krylov space methods and in modifying anexisting orthogonal decomposition.In previous work, we have made assumptions about the diagonalblocks of R to insure that a block classical Gram–Schmidtalgorithm with reorthogonalization will produce a backwardstable decomposition with a near left orthogonal Q. The contextof this talk is where these diagonal blocks violate thoseassumptions and are allowed to be singular or veryill-conditioned. A strategy for using rank–revealingdecompositions to deal with that contingency is given thatinsures a similar result to the case where the assumptions aboutthe blocks of R hold. The algorithm is considered in the contextof block downdating.Jesse BarlowDepartment of Computer Science and EngineeringPenn State UniversityUniversity Park, PA 16802, [email protected]

Talk 2. A numerical method for a mixed discrete bilinear leastsquares problemFor CPU utilization control in distributed real-time embeddedsystems, we recently proposed a new scheme to makesynchronous rate and frequency adjustment to enforce theutilization set point. In this scheme, we need to solve a mixeddiscrete bilinear least squares problem, in which one unknownvector is subject to a box constraint and the other unknownvector’s each entry has to be taken from a discrete set ofnumbers. In this talk we propose an alternating iterative methodto solve this problem.This is joint work with Xi Chen and Xue Liu.Xiao-Wen Chang

33 2012 SIAM Conference on Applied Linear Algebra

School of Computer ScienceMcGill UniversityMontreal, Quebec H3A 1Y1, [email protected]

Talk 3. On condition numbers for constrained linear leastsquares problemsCondition numbers are important in numerical linear algebra,which can tell us the posterior error bounds for the computedsolution. Classical condition numbers are normwise, but theyignore the input data sparsity and/or scaling. Componentwiseanalysis have been introduced, which gives a powerful tool tostudy the perturbations on input and output data regarding thesparsity and scaling. In this paper under componentwiseperturbation analysis we will study the condition numbers forconstrained linear least squares problems. The obtainedexpressions of the condition numbers avoid the explicit formingof Kronecker products, which can be estimated by powermethods efficiently. Numerical examples show that ourcondition numbers can give better error bounds.Huaian DiaoSchool of Mathematics and StatisticsNortheast Normal UniversityChang Chun 130024P.R. [email protected]

Yimin WeiSchool of Mathematical SciencesFudan UniversityShanghai, 200433P.R. [email protected]

Talk 4. SOR inner-iteration GMRES for underdeterminedleast squares problemsSuccessive over-relaxation (SOR) inner iterations are proposedfor preconditioning the generalized minimal residual method(GMRES) for underdetermined least squares problems. Theright-preconditioned GMRES (AB-GMRES) may fail toconverge for inconsistent problems. Instead, theleft-preconditioned GMRES method (BA-GMRES) works sincewe can transform an inconsistent system to a consistent one.Numerical experiments show that BA-GMRES with SOR for thenormal equation works efficiently. Moreover, we showconditions under which BA-GMRES determines aminimum-norm least squares solution without breakdown andpresent numerical results.Keiichi MorikuniThe Department of InformaticsThe School of Multidisciplinary ScienceThe Graduate University for Advanced Studies2-1-2 Hitotsubashi, Chiyoda-kuTokyo [email protected]

Ken HayamiThe Department of InformaticsThe School of Multidisciplinary ScienceThe Graduate University for Advanced Studies2-1-2 Hitotsubashi, Chiyoda-kuTokyo [email protected]

MS 35. Nonlinear eigenvalue problems

Talk 1. Computable error bounds for nonlinear eigenvalueproblems allowing for a minimax characterizationFor a nonlinear eigenvalue problem

T (λ)x = 0

allowing for a variational characterization of its eigenvalues wediscuss computable error bounds of Krylov–Bogoliubov and ofKato–Temple type.Heinrich VossInstitute of Numerical SimulationHamburg University of [email protected]

Kemal YildiztekinInstitute of Numerical SimulationHamburg University of [email protected]

Talk 2. A restarting technique for the infinite Arnoldi method

Different adaptions of the Arnoldi method are often used tocompute partial Schur factorizations. We propose here atechnique to compute a partial Schur factorization of a nonlineareigenvalue problem (NEP). The technique is inspired by thealgorithm in [E. Jarlebring, K. Meerbergen, W. Michiels, Alinear eigenvalue algorithm for the nonlinear eigenvalueproblem, 2012], now called the infinite Arnoldi method, forwhich we design an appropriate restarting technique. Thetechnique is based on a characterization of the invariant pairs ofthe NEP, which turn out to be equivalent to the invariant pairs ofan operator. We characterize the structure of the invariant pairsof the operator and show how one can carry out a modificationof the infinite Arnoldi method by respecting this structure. Thisalso allows us to naturally add the feature known as locking. Wenest this algorithm with an outer iteration, where the infiniteArnoldi method for a particular type of structured functions isappropriately restarted. The restarting exploits the structure andis inspired by the well-known implicitly restarted Arnoldimethod for standard eigenvalue problems.Elias JarlebringCSC / Numerical analysisKTH - Royal institute of [email protected]

Karl MeerbergenDept. of Comp. [email protected]

Wim MichielsDept. of Comp. [email protected]

Talk 3. Robust successive computation of eigenpairs fornonlinear eigenvalue problemsThe successive computation of several eigenpairs of a nonlineareigenvalue problem requires a means to prevent our algorithmfrom repeatedly converging to the same eigenpair. For lineareigenproblems, this is usually accomplished by restricting theproblem to the subspace orthogonal to all previously computedeigenvectors. In the nonlinear case, though, this strategy isharmful as it may cause eigenpairs to be missed due to possiblelinear dependencies among the eigenvectors. In this talk, wepresent a modified orthogonalization scheme based on theconcept of minimal invariant pairs, which does not bear thisdanger.

2012 SIAM Conference on Applied Linear Algebra 34

Cedric EffenbergerMathematics Institute of Computational Science and EngineeringSwiss Federal Institute of Technology [email protected]

Talk 4. Triangularization of matrix polynomialsUsing similarity transformations that preserve the structure ofthe companion matrix, we show that any matrix polynomialP (λ) over the complex numbers can be triangularized in thesense that there always exists a triangular matrix polynomialT (λ) having the same degree and elementary divisors (finite andinfinite) as P (λ) . Although the result is theoretical the approachis constructive and hints on possible numerical algorithms.Leo TaslamanSchool of MathematicsUniversity of [email protected]

Yuji NakatsukasaSchool of MathematicsUniversity of [email protected]

Francoise TisseurSchool of MathematicsUniversity of [email protected]

MS 36. Hybrid solvers for sparse linear equationsTalk 1. The augmented block-Cimmino distributed methodIn row projection methods, such as Block-Cimmino, near lineardependency between partitions implies small eigenvalues in thespectrum of the iteration matrix. In this work we try to break thisnear linear dependency by augmenting the matrix to enforceorthogonality between partitions. We formulate a linear systemin a super-space, where the extra equations that are introducedfor consistency are obtained by projections and can be handledimplicitly. In the ideal case the resulting iteration matrix has alleigenvalues almost equal to one. We investigate the numericalcomponents of this approach and ways to reduce the number ofsuper-space variables.Mohamed [email protected]

Iain DuffCERFACS and Rutherford Appleton [email protected]

Ronan [email protected]

Daniel [email protected]

Talk 2. On a parallel hierarchical algebraic domaindecomposition method for a large scale sparse linear solverThe emerging petaflop computers have processing nodes basedon multi- /many-core chips. To fully utilize such parallelarchitectures, a natural approach is to exploit medium-grainparallelism through multi-threading within a node andcoarse-grain parallelism using message-passing (MPI) betweennodes. For a hybrid direct/iterative solver, this can beimplemented by using a parallel sparse direct solver within anode on local subproblems. The available multi-threaded sparsedirect solvers can be used to take advantage of the multicorearchitecture of the nodes. The MPI paradigm is then used toimplement the iterative scheme for the interfaces between the

subproblems. In this talk, we will present such a parallelimplementation and illustrate its performance scalability on a setof academic and industrial test problems. In addition, we willpropose a model to compute the subsequent computational andmemory costs, depending on the ratio of computation performedin the direct solver. We apply this model to several classes ofproblems and illustrate the possible trade-offs in terms ofcomputational and memory complexity.Luc GiraudInria, joint Inria-CERFACS [email protected]

Emmanuel [email protected]

Abdou GuermoucheBordeaux [email protected]

Jean [email protected]

Talk 3. A two-level Schwarz method for systems with highcontrastsTwo-level Schwarz methods are popular domain decompositionmethods. They are based on direct solves in each subdomain aswell as in a specified coarse space. Used as a preconditioner forthe Conjugate Gradient algorithm they lead to efficient hybridsolvers.Heterogeneous coefficients in the PDEs are a known challengefor robustness in domain decomposition, specially if theheterogeneities are across interfaces. We introduce a new coarsespace for which the Schwarz method is robust even in this case.It relies on solving local generalized eigenvalue problems in theoverlaps between subdomains which isolate exactly the termsresponsible for slow convergence.Nicole SpillaneLaboratoire Jacques-Louis LionsUniversite Pierre et Marie [email protected]

Victorita DoleanLaboratoire J.-A. DieudonneUniversite de Nice-Sophia [email protected]

Patrice HauretCentre de technologieMFP [email protected]

Frederic NatafLaboratoire Jacques-Louis LionsUniversite Pierre et Marie [email protected]

Clemens PechsteinInstitute of Computational MathematicsJohannes Kepler [email protected]

Robert ScheichlDepartment of Mathematical SciencesUniversity of [email protected]

Talk 4. A 3-level parallel hybrid preconditioner for sparselinear systemsWe describe ShyLU, a hybrid direct-iterative preconditioner andsolver for general sparse linear systems. We use a Schurcomplement approach where subproblems are solved using adirect solver while the Schur complement is solved iteratively.

35 2012 SIAM Conference on Applied Linear Algebra

We employ a block decomposition, which is formed in a fullyautomatic way using state-of-the-art partitioning and orderingalgorithms in the Zoltan and Scotch software packages. Themethod can be viewed as a two-level scheme. We show resultsthat demonstrate the robustness and effectiveness of ShyLU forproblems from a variety of applications. We have found it isparticularly effective on circuit simulation problems where manyiterative methods fail. We also show results from a hybrid MPI +threads version of ShyLU. The Schur complement strategy inShyLU works well for moderate amount of parallelism. For verylarge problems on very large parallel systems, we introduce athird level based on domain decomposition to improvescalability. We study how ShyLU behaves as a subdomainsolver. This strategy looks promising for peta- and exa-scalesystems, which usually have a hierarchical structure.Erik G. BomanSandia National [email protected]

Siva RajamanickamSandia National [email protected]

Michael A. HerouxSandia National [email protected]

Radu PopescuEPFL, [email protected]

MS 37. Optimization methods for tensordecompositionTalk 1. Efficient algorithms for tensor decompositionsThe canonical polyadic and rank-(Lr, Lr, 1) block termdecomposition (CPD and BTD, respectively) are two closelyrelated tensor decompositions. The CPD is an important tool inpsychometrics, chemometrics, neuroscience and data mining,while the rank-(Lr, Lr, 1) BTD is an emerging decompositionin signal processing and, recently, blind source separation. Wepresent a decomposition that generalizes these two and developalgorithms for its computation. Among these algorithms arealternating least squares schemes, several general unconstrainedoptimization techniques, as well as matrix-free nonlinear leastsquares methods. In the latter we exploit the structure of theJacobian’s Gramian by means of efficient expressions for itsmatrix-vector product. Combined with an effectivepreconditioner, numerical experiments confirm that thesemethods are among the most efficient and robust currentlyavailable for computing the CPD, rank-(Lr, Lr, 1) BTD andtheir generalized decomposition.Laurent SorberDept. of Computer Science KU [email protected]

Marc Van BarelDept. of Computer Science KU [email protected]

Lieven De LathauwerGroup Science, Engineering and Technology, KU Leuven CampusKortrijk& Dept. of Electrical Engineering (ESAT), KU [email protected]

Talk 2. Symmetric tensor decomposition via a power methodfor the generalized tensor eigenproblemWe present an approach for approximate (least-squares)decomposition of a symmetric positive-definite tensor of

dimension n into a given number p of symmetric rank-1 terms.This is accomplished in two parts. The first part is to transformthe decomposition problem into a ”generalized tensoreigenproblem” (GTEP) for tensors of dimension np. The GTEPis of independent interest and, for example, subsumes thepreviously defined Z-eigenpairs and H-eigenpairs. The secondpart is to extend a previously developed power method forZ-eigenpairs to solve the GTEP. We discuss the conceptsunderlying these techniques and provide numerical examples.Jackson R. MayoSandia National [email protected]

Tamara G. KoldaSandia National [email protected]

Talk 3. All-at-once optimization for coupled matrix andtensor factorizationsJoint analysis of data from multiple sources can enhanceknowledge discovery. The task of fusing data, however, ischallenging since data may be incomplete and heterogeneous,i.e., data consists of matrices and higher-order tensors. Weformulate this problem as a coupled matrix and tensorfactorization problem where heterogeneous data are modeled byfitting outer-product models to higher-order tensors and matricesin a coupled manner. Unlike traditional approaches usingalternating algorithms, we use a gradient-based all-at-onceoptimization approach. Using numerical experiments, wedemonstrate that the all-at-once approach is often more accuratethan the alternating approach and discuss the advantages ofcoupled analysis in terms of missing data recovery.Evrim AcarUniversity of [email protected]

Tamara G. KoldaSandia National [email protected]

Daniel M. DunlavySandia National [email protected]

Rasmus BroUniversity of [email protected]

Talk 4. An algebraic multigrid optimization method forlow-rank canonical tensor decompositionA new algorithm based on algebraic multigrid is presented forcomputing the rank-R canonical decomposition of a tensor forsmall R. Standard alternating least squares (ALS) is used as therelaxation method. Transfer operators and coarse-level tensorsare constructed in an adaptive setup phase that combinesmultiplicative correction and bootstrap algebraic multigrid. Anaccurate solution is computed by an additive solve phase basedon the full approximation scheme. Numerical tests show that forcertain test problems, our multilevel method significantlyoutperforms standalone ALS when a high level of accuracy isrequired.Killian MillerUniversity of [email protected]

Hans De SterckUniversity of [email protected]

MS 38. Generalized inverses and applications - Part I

2012 SIAM Conference on Applied Linear Algebra 36

of IITalk 1. The group inverse of additively modified matricesLet A ∈ Cn×n. We recall that the group inverse of A, is theunique matrix A# ∈ Cn×n, if it exists, which satisfies theequations: A#AA# = A#; AA#A = A; AA# = A#A. If Ais nonsingular then A# = A−1. It is well known that the groupinverse of A exists if rank(A) = rank(A2). Alternatively, A#

exists if A+ I −AA− is nonsingular, independently of thechoice of A−, where A− denotes an inner inverse of A, i.e.AA−A = A. Let B ∈ Cn×r , C ∈ Cr×n, and r < n. If A andI − CA−1B are nonsingular, then to adapt the inverse of themodified matrix A−BC we can apply theSherman-Morrison-Woodbury formula:(A−BC)−1 = A−1 +A−1B(I − CA−1B)−1CA−1; and,hence, we need to compute the inverse of I − CA−1B, whichhas size r < n. This talk will be focused on the group inverse ofadditively modified matrices, (A−BC)#, when it exists.Extensions of the previous formula will be considered. Anapplication to the perturbation of the group inverse ofA = I − T , where T is the transition matrix of a Markov chainwill be given. The research is partially supported by ProjectMTM2010-18057, “Ministerio de Ciencia e Innovacion” ofSpain.Nieves Castro GonzalezDepartamento de MatematicaUniversidad Politecnica de Madrid, [email protected]

Talk 2. The Moore-Penrose inverse of a linear combination ofcommuting generalized and hypergeneralized projectorsWe present new results concerning the representation of theMoore-Penrose inverse of a linear combination of generalizedand hypergeneralized projectors and give the form for theMoore-Penrose inverse, i.e., the group inverse of c1A+ c2B,where A, B are two commuting generalized or hypergeneralizedprojectors and c1; c2 ∈ C\0 and c31 + c32 6= 0. Furthermore,we show that under that conditions the invertibility ofc1A+ c2B is independent of the choice of the scalars c1; c2.Also, we present some results relating different matrix partialorderings and the invertibility of a linear combination of EPmatrices.Dragana S. Cvetkovic-IlicFaculty of Mathematics and SciencesUniversity of Nis, [email protected]

Talk 3. Generalized inverses of operators on HilbertC∗-modulesWe present new results on the theory and applications ofgeneralized inverses of operators between Hilbert C∗-modules.Dragan S. DjordjevicFaculty of Sciences and MathematicsUniversity of Nis, [email protected]

Talk 4. Some results on the reverse order lawWe presented some necessary and sufficient conditionsconcerning the reverse order laws for the group inverses ofelements in rings and rings with involution. In particular,assuming that a and b are Moore-Penrose invertible or groupinvertible elements, we study equivalent conditions which arerelated to the reverse order law for the group inverse of theproduct ab. Also, some equivalent condition which ensure that

the product ab is EP element are consider too. Severalconditions for (ab)# = b#(a#abb#)#a# to hold in rings areinvestigate. We extend the recent results to more generalsettings, giving some new conditions and providing simplerproofs to already existing conditions.D. MosicFaculty of Sciences and MathematicsUniversity of Nis, [email protected]

MS 39. Challenges for the solution andpreconditioning of multiple linear systems - Part I of IITalk 1. Preconditioners for sequences of shifted linear systems

Sequences of shifted linear systems with the same right-handside can be solved with multi-shift Krylov methods at almost thesame cost as for solving the unshifted system. This is possibledue to the fact that the Krylov subspaces for the shifted and theunshifted problems are the same. The main drawback ofmulti-shift Krylov methods is that they cannot be combined withan arbitrary preconditioner since in general the Krylov subspacesfor the preconditioned shifted problems will be different. In thetalk we will discuss this problem in detail and also explain howto construct preconditioners that can be combined withmulti-shift Krylov methods.Martin B. van GijzenDelft Institute of Applied MathematicsDelft University of Technology, the [email protected]

Daniel B. SzyldDepartment of MathematicsTemple University, Philadelphia, Pa., [email protected]

Talk 2. Solving sequences of linear systems with application tomodel reductionTo obtain efficient solutions for sequences of linear systemsarising in interpolatory model order reduction (both parametricand nonparametric), we recycle Krylov subspaces from onesystem (or pair of systems) in a sequence to the next. We firstintroduce a generalization of BiCG. Here we show that even fornon-Hermitian matrices one can build bi-orthogonal bases (forthe associated two Krylov subspaces) using a short-termrecurrence. We then adapt and apply Recycling BiCG andRecycling BiCGSTAB to model reduction algorithms. For amodel reduction problem we demonstrate that solving theproblem without recycling leads to (about) a 50% increase inruntime.Kapil AhujaMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

Eric de SturlerDept. of MathematicsVirginia Tech, Blacksburg, Va., [email protected]

Serkan GugercinDept. of MathematicsVirginia Tech, Blacksburg, Va., [email protected]

Peter BennerMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

37 2012 SIAM Conference on Applied Linear Algebra

Talk 3. Krylov subspace recycling for families of shifted linearsystemsWe address the solution of a sequence of families of linearsystems. For each family, there is a base coefficient matrix Ai,and the coefficient matrices for all systems in the family differfrom Ai by a multiple of the identity, e.g.,

Aixi = bi and (Ai + σ(`)i I)x

(`)i = bi for ` = 1 . . . Li,

where Li is the number of shifts at step i. We propose a newmethod which solves the base system using GMRES withsubspace recycling while constructing approximate correctionsto the solutions of the shifted systems. At convergence of thebase system solution, GMRES with subspace recycling isapplied to further improve the solutions of the shifted systems totolerance. We present analysis of this method and numericalresults involving systems arising in lattice quantumchromodynamics.Kirk M. SoodhalterDepartment of MathematicsTemple University, Philadelphia, Pa., [email protected]

Daniel B. SzyldDepartment of MathematicsTemple University, Philadelphia, Pa., [email protected]

Fei XueDepartment of MathematicsTemple University, Philadelphia, Pa., [email protected]

Talk 4. Krylov subspace recycling for faster model reductionalgorithmsThe numerical solution of shifted linear system with multipleright-hand sides is a recurring task in many computationalproblems related to model reduction. Shifts and right-hand sidesmay be known in advance or may became available only afterthe previous linear system has been solved. We will survey somemodel reduction algorithms where these tasks occur and willdemonstrate that Krylov subspace recycling may significantlyaccelerate the computation of reduced-order models in thesesituations.Peter BennerMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

Lihong FengMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

MS 40. Different perspectives on conditioning andnumerical stability - Part I of IITalk 1. Highly accurate numerical linear algebra via rankrevealing decompositionsHigh accuracy algorithms are those that produce relative forwarderrors of order the unit roundoff of the computer even formatrices that are very ill-conditioned in a traditional sense. Thistype of algorithms only exist for certain structured matrices. Awide class of these matrices are the matrices for which it ispossible to compute accurate rank revealing decompositions(RRDs), i.e., factorizations XDY where D is diagonal andnon-singular, and X and Y are well conditioned. This classcomprise many important structured matrices, like

Vandermonde, Cauchy, graded matrices and many others.Originally, high accuracy algorithms acting on the factors ofRRDs (instead of acting directly on the matrix) were designedfor computing Singular Value Decompositions (1999), then forcomputing eigenvalues/eigenvectors of symmetric matrices(2003, 2009), and very recently for computing solutions of linearsystems (2011) and least square problems (2012). The purposeof this talk is to present a unified description of all thesealgorithms and to pose some open problems in this area.Froilan M. DopicoICMAT and Department of MathematicsUniversidad Carlos III de Madrid (Spain)[email protected]

Talk 2. Stability of numerical algorithms with quasiseparablematricesQuasiseparable matrices are encountered in PDEs, computationswith polynomials (interpolation, root finding etc.), design ofdigital filters and other areas of applied mathematics. There havebeen major interest in fast algorithms with quasiseparablematrices in the last decade. Many linear complexity algorithmshave been developed including QR factorization, system solver,QR iterations and others. However, their numerical stability hasnot received equal attention of the scientific community. In ourtalk we will present first results of error analysis of QRdecomposition and system solvers. We will also present ageneralization of Parlett’s dqds algorithm and discussperspectives of using quasiseparable matrices for stableevaluation of polynomials’ roots.Pavel ZhlobichSchool of MathematicsUniversity of [email protected]

Froilan DopicoDepartamento de MatematicasUniversidad Carlos III de [email protected]

Vadim OlshevskyDepartment of MathematicsUniversity of [email protected]

Talk 3. Gram-Schmidt orthogonalization with standard andnon-standard inner product: rounding error analysisIn this contribution we consider the most important schemesused for orthogonalization with respect to the standard andnon-standard inner product and review the main results on theirbehavior in finite precision arithmetic. Although all the schemesare mathematically equivalent, their numerical behavior can besignificantly different. We treat separately the particular case ofthe standard inner product and show that similar results hold alsofor the case when the inner product is induced by a positivediagonal matrix. We will show that in the case of general innerproduct the orthogonality between computed vectors besides thelinear independence of initial vectors depends also on thecondition number of the matrix that induces the non-standardinner product. Finally we discuss the extension of this theory tosome bilinear forms used in the context of variousstructure-preserving transformations.Miroslav RozloznıkInstitute of Computer Science, Czech Academy of SciencesPrague, Czech [email protected]

Jirı Kopal

2012 SIAM Conference on Applied Linear Algebra 38

Institute of Novel Technologies and Applied InformaticsTechnical University of Liberec, Czech [email protected]

Alicja SmoktunowiczFaculty of Mathematics and Information ScienceWarsaw University of Technology, [email protected]

Miroslav TumaInstitute of Computer Science, Czech Academy of SciencesPrague, Czech [email protected]

Talk 4. Backward stability of iterations for computing thepolar decompositionAmong the many iterations available for computing the polardecomposition the most practically useful are the scaled Newtoniteration and the recently proposed dynamically weighted Halleyiteration. Effective ways to scale these and other iterations areknown, but their numerical stability is much less wellunderstood. In this work we show that a general iterationXk+1 = f(Xk) for computing the unitary polar factor isbackward stable under two conditions. The first conditionrequires that the iteration is implemented in a mixedbackward–forward stable manner and the second requires thatthe mapping f does not significantly decrease the size of anysingular value relative to the largest singular value. Using thisresult we show that the dynamically weighted Halley iteration isbackward stable when it is implemented using Householder QRfactorization with column pivoting and either row pivoting orrow sorting. We also prove the backward stability of the scaledNewton iteration under the assumption that matrix inverses arecomputed in a mixed backward–forward stable fashion; ourproof is much shorter than a previous one of Kiełbasinski andZietak. We also use our analysis to explain the instability of theinverse Newton iteration and to show that the Newton–Schulziteration is only conditionally stable. This work shows that bycarefully blending perturbation analysis with rounding erroranalysis it is possible to produce a general result that can provethe backward stability or predict or explain the instability (as thecase may be) of a wide range of practically interesting iterationsfor the polar decomposition.Nicholas J. Higham

School of MathematicsUniversity of [email protected]

Yuji NakatsukasaSchool of MathematicsUniversity of [email protected]

MS 41. Recent advances in model reduction - Part I ofIITalk 1. The Loewner framework in data-driven modelreductionIn this talk we will survey recent advances in model reductionconcentrating on the method of model reduction directly fromdata using the Loewner framework. In addition we will showhow the Loewner framework can be extended to handlereduction of systems depending on parameters.A.C. AntoulasRice [email protected]

Talk 2. Robust computational approaches toH2-optimal

model reductionWe present an approach to model reduction for MIMO lineardynamical systems that is numerically stable, nonintrusive, andcomputationally tractable even for very large order systems. Ourstrategy is a hybrid of the Iterative Rational Krylov Algorithm(IRKA) and a trust-region method with logarithmic barrier thatguarantees stability. The method produces a monotonicallyimproving (with respect toH2 error) sequence of stable reducedorder models that is rapidly and globally convergent. Theeffectiveness of the algorithm is illustrated through numericalexamples drawn from a variety of challenging dynamicalsystems that arise in practical settings.Christopher BeattieDept. of MathematicsVirginia Polytechnic Institute and State [email protected]

Serkan GugercinDept. of MathematicsVirginia Polytechnic Institute and State [email protected]

Talk 3. Reduced order modeling via framesIn this talk we present a new concept that arises in thediscretization of PDEs. Instead of using bases (like hierarchicalfinite elements) to represent the solution in a certain functionspace, we use frames, which consist of a set of standard basisfunctions enriched by specific functions that allow to addressspecific solution behavior like singularities, boundary layers orsmall amplitude, high frequency oscillations. We thendetermined sparse solutions represented in these frames, bysolving the under-determined problems to compute sparsesolutions. We show that by doing this in a multilevel approachwe can achieve much smaller models than with classicaladaptive FEM.Volker MehrmannInst. f. MathematikTU [email protected]

Sadegh JokarInst. f. Mathematik TU [email protected]

Sarosh QuraischiInst. f. Mathematik TU [email protected]

Talk 4. Semidefinite Hankel-type model reduction based onfrequency response matchingIn this talk a model reduction method for linear time-invariantsystems is discussed. It is based on frequency response matchingand semidefinite programming techniques. The method isrelated to Hankel model reduction. It can be claimed that thepresented method is a scalable approach to Hankel modelreduction. Numerical simulations show that the accuracy of themethod is comparable to the well-known techniques, such asbalanced truncation and Hankel model reduction. The developedmethod can be applied to various problems. In this talk we aregoing to briefly discuss two: parameterized model reduction andreduction in the nu-gap metric.Aivar SootlaDept. of Automatic ControlLund [email protected]

Anders RantzerDept. of Automatic Control

39 2012 SIAM Conference on Applied Linear Algebra

Lund [email protected]

Kin Cheong SouKTH School of Electrical Engineering, Automatic Control,[email protected]

MS 42. Structured solution of nonlinear matrixequations and applications - Part I of IITalk 1. Structured solution of large-scale algebraic Riccatiand nonlinear matrix equationsWe consider the solution of large-scale algebraic Riccati andnonlinear matrix equations (AREs, NMEs). For discrete-timeAREs, the structure-preserving doubling algorithm will beadapted for the corresponding sparsity and low-rankedstructures. For continuous-time AREs, the Cayley transform isapplied before doubling. Similar methodology is applied tolarge-scale nonsymmetric AREs and NMEs. With n being thedimension of the equations, the resulting algorithms are of O(n)complexity. Numerical results will be presented. As an example,a DARE with 3.19 billion unknowns was solved using MATLABon a MacBook Pro to machine accuracy within 1,100 seconds.Eric King-wah ChuSchool of Mathematical SciencesBuilding 28, Monash University, VIC 3800, [email protected]

Tiexiang LiDept. of MathematicsSoutheast University, Nanjing 211189, People’s Republic of [email protected]

Wen-Wei LinDept. of Applied MathematicsNational Chiao Tung University, Hsinchu 300, [email protected]

Peter Chang-Yi WengSchool of Mathematical SciencesBuilding 28, Monash University, VIC 3800, [email protected]

Talk 2. Accurate solutions of nonlinear matrix equations inqueueing modelsIn this talk, we discuss numerical issues arising in finiteprecision implementations of iterative methods for solvingnonlinear matrix equations arising in queueing models.Exploring the structure of the problem, we shall present somenumerically more stable implementations. A rounding erroranalysis together with numerical examples are given todemonstrate the higher accuracy achieved by the refinedimplementations.Qiang YeDept. of MathematicsUniversity of Kentucky, Lexington KY 40506-0027, [email protected]

Talk 3. A numerical approach for solving nonlinear matrixequations in economic dynamicsModern economic theory views the economy as a dynamicalsystem. The dynamics includes changes over time of marketbehavior such as consumption, investment, labor supply, andtechnology innovation. To make analytical analysis, the naiveapproach is to set up the Euler equation and subsequently solveit by finding the policy functions. Indeed, this is a process ofsolving the nonlinear matrix equation. In this work, we proposea Newton iterative scheme on approximating the unknown policyfunctions. Applications to the neoclassical growth model withleisure choice are used to demonstrate the working of the idea.

Matthew M. LinDept. of MathematicsNational Chung Cheng [email protected]

Moody T. ChuDept. of Applied MathematicsNorth Carolina State [email protected]

Chun-Hung KuoDept. of EconomicsNorth Carolina State [email protected]

Talk 4. A structure-preserving doubling algorithm forquadratic eigenvalue problems arising from time-delaysystemsWe propose a structure-preserving doubling algorithm for aquadratic eigenvalue problem arising from the stability analysisof time-delay systems. We are particularly interested in theeigenvalues on the unit circle, which are difficult to estimate.The convergence and backward error of the algorithm areanalyzed and three numerical examples are presented. Ourexperience shows that our algorithm is efficient in comparison tothe few existing approaches for small to medium size problem.Tiexiang LiDept. of MathematicsSoutheast University, Nanjing 211189, People’s Republic of [email protected]

Eric King-wah ChuSchool of Mathematical SciencesBuilding 28, Monash University 3800, [email protected]

Wen-Wei LinDept. of Applied MathematicsNational Chiao Tung University, Hsinchu 300, [email protected]

MS 43. Challenges for the solution andpreconditioning of multiple linear systems - Part II ofIITalk 1. Low-rank techniques for parameter-dependent linearsystems and eigenvalue problemsMotivated by the need for quantifying the impact ofuncertainties in engineering applications, recently a number ofnew approaches for solving parameter-dependent and stochasticPDEs have been developed. In particular, this includes sparsegrid Galerkin and collocation methods, as well as low-ranktechniques. Depending on the regularity of the parameterdependence, these methods are able to deal with potentiallymany parameters. The aim of this talk to provide a survey ofrecent work on low-rank techniques for solving linear systemsarising from the spatial discretization of parameter-dependentPDEs. The extension of these techniques to generalparameter-dependent eigenvalue problems is nontrivial and willalso be briefly discussed.Christine ToblerMATHICSE-ANCHPEPF Lausanne, [email protected]

Daniel KressnerMATHICSE-ANCHPEPF Lausanne, [email protected]

Talk 2. Recycling Krylov subspace information in sequencesof linear systems

2012 SIAM Conference on Applied Linear Algebra 40

Many problems in numerical simulations in physics require thesolution of long sequences of slowly changing linear systems.One problem that is of interest to us arises in Lattice QCDsimulations, e.g., while computing masses of elementaryparticles. In each time step, we have to solve a linear systemwith a Dirac operator which changes slightly from time step totime step. Based on the work of M. Parks, E. de Sturler etal.(Recycling Krylov subspaces for sequences of linear systems,SIAM J. on Scientific Computing, 28(2006), 1651− 1674) wewill show how the cost of solving subsequent systems is reducedby recycling selected subspaces generated for previous systems.Furthermore, we have investigated how the algorithm behaveswhen we use the solution of the previous system as an initialguess and also, when we use different kinds of extrapolations ofthe previous solutions as an initial guess. We will show theseresults and the effectiveness of the algorithm in comparison tothe algorithm that solves each system separately.Nemanja BozovicFachbereich C, Mathematik und NaturwissenschaftenBergische Universitat [email protected]

Talk 3. Efficiently updating preconditioners in quantumMonte Carlo simulationsIn the quantum Monte Carlo method for computing properties ofmaterials, we need to solve a very long sequence of linearsystems arising from importance sampling. Each system in thesequence is a so-called Slater matrix, and the matrices change byone row (or a few rows) at a time corresponding to small movesby particles. We will combine a method for incremental matrixreordering with multilevel preconditioning to develop effectivepreconditioners that can be recycled efficiently.Arielle Grim McNallyDepartment of MathematicsVirginia Tech, Blacksburg, Va., [email protected]

Eric de SturlerDepartment of MathematicsVirginia Tech, Blacksburg, Va., [email protected]

Kapil AhujaMax Planck Institute for Dynamics of Complex Technical SystemsMagdeburg, [email protected]

Li MingDepartment of MathematicsVirginia Tech, Blacksburg, Va., [email protected]

Talk 4. A domain decomposition preconditioned recyclingGMRES for stochastic parabolic PDEWe discuss an implicit space-time approach for solvingstochastic parabolic PDEs. We first decouple the space-timediscretized stochastic equation into some uncoupleddeterministic systems by using a Karhunen-Loeve expansion anddouble orthogonal polynomials. And then a domaindecomposition method is combined with recycling GMRES tosolve the large number of systems with similar structures. Wereport experiments obtained on a parallel computer with a largenumber of processors. This is a joint work with Cui Cong.Xiao-Chuan CaiDepartment of Computer ScienceUniversity of Colorado at BoulderBoulder, Co., [email protected]

MS 44. Different perspectives on conditioning andnumerical stability - Part II of IITalk 1. Accuracy and sensitivity of Monte Carlo matrixmultiplication algorithmsRandomized matrix multiplication algorithms were designed toapproximate very large matrix products for which adeterministic algorithm is prohibitively expensive. For analgorithm introduced by Drineas, Kannan, and Mahoney, weanalyze the error resulting from randomization and derive abound that improves existing results. In addition, we formulate ameasure for the sensitivity of the algorithm to perturbations inthe input and present a bound for this measure. We also comparethe sensitivity of the randomized algorithm to the sensitivity ofthe deterministic algorithm.John T. HolodnakDept. of MathematicsNorth Carolina State UniversityRaleigh, NC 27695, [email protected]

Ilse C. F. IpsenDept. of MathematicsNorth Carolina State UniversityRaleigh, NC 27695, [email protected]

Talk 2. Hyperdeterminant and the condition number of amultilinear systemThe condition number of a problem is the reciprocal of itsnormalized distance to the nearest ill-posed instance in that classof problems. We shall see that in solving a system of multilinearequations, the set of ill-posed problems is given by the set ofcoefficient tensors with vanishing hyperdeterminant. Theparallel with matrices goes further: the hyperdeterminant of atensor is zero iff it has a zero singular value; and in the case of asymmetric tensor, iff it has a zero eigenvalue. Moreover, thecriterion is invariant under Gaussian elimination orHouseholder/Givens transformations applied to ‘all k sides’ ofthe tensor viewed as a k-dimensional hypermatrix.Lek-Heng LimDepartment of StatisticsUniversity of [email protected]

Talk 3. Condition numbers and backward errors in functionalsettingIn this talk we present a functional perturbation results, i.e., afunctional normwise backward error, for the PDE eigenvalueproblems. Inspired by the work of Arioli et al. for linear systemsarising from the finite element discretization of boundary valueproblems we extend the ideas of the functional CompatibilityTheorem and condition numbers to eigenvalue problems in theirvariational formulation. Moreover, we discuss a new spacewisebackward error for PDEs using the componentwise erroranalysis, i.e., by performing the error analysis using theso-called hypernorms of order k.Agnieszka MiedlarInstitute of MathematicsTechnical University of [email protected]

Mario ArioliSTFC Rutherford Appleton [email protected]

Talk 4. Orthogonality and stability in large-sparse-matrixiterative algorithms

41 2012 SIAM Conference on Applied Linear Algebra

Many iterative algorithms for large sparse matrix problems arebased on orthogonality, but this can rapidly be lost using vectororthogonalization (subtracting multiples of earlier vectors fromthe latest vector to produce the next orthogonal vector). Yetthese are among the best algorithms we have, and include theLanczos process, CG, Golub and Kahan bidiagonalization, andMGS-GMRES.Here we describe a form of orthogonal matrix that arises fromany sequence of supposedly orthogonal vectors. We illustratesome of its properties, including a beautiful measure oforthogonality of the original set of vectors. We will show howthis leads to new concepts of conditioning and stability for thesealgorithms.Chris PaigeSchool of Computer ScienceMcGill [email protected]

Wolfgang WullingW2 Actuarial & Math Services [email protected]

MS 45. Recent advances in model reduction - Part IIof IITalk 1. Automating DEIM for nonlinear model reductionThe discrete empirical interpolation method (DEIM) can providespectacular dimension and complexity reduction for challengingsystems large scale nonlinear ordinary differential equations.The DEIM replaces orthogonal projection of POD with aninterpolatory projection of the nonlinear term that only requiresthe evaluation of a few selected components. However, theimplementation at present is intrusive in the sense that a usermust provide a scheme for evaluating the selected componentsof the nonlinear terms in order to integrate the reduced ordersystem. We have utilized some of the techniques of automaticdifferentiation to fully automate this process. In particular, thisapproach will automatically generate a new code derived directlyfrom the original C-code for the right hand side of a first orderODE.D.C. SorensenDept. of Computational and Applied Math.Rice [email protected]

Russell L. CardenDept. of Computational and Applied Math.Rice [email protected]

Talk 2. Model reduction for optimal control problems infield-flow fractionationWe discuss the application of model order reduction to optimalcontrol problems governed by coupled systems of theStokes-Brinkmann and advection diffusion equations. Suchproblems arise in field-flow fractionation processes for theefficient and fast separation of particles of different size inmicrofluidic flows. Our approach is based on a combination ofbalanced truncation and tangential interpolation for modelreduction of the semidiscretized optimality system. Numericalresults demonstrate the properties of this approach.Tatjana StykelInstitut fur MathematikUniversitat [email protected]

Carina Willbold

Institut fur MathematikUniversitat [email protected]

Talk 3. Numerical implementation of the iterative rationalKrylov algorithm for optimalH2 model order reductionThe Iterative Rational Krylov (IRKA) algorithm for model orderreduction (Gugercin, Antoulas, Beattie 2008.) has recentlyattracted attention because of its effectiveness in real worldapplications, as well as because of its mathematical elegance.Our current work is focused on the development of efficient andnumerically reliable mathematical software that implements theIRKA algorithm. The first step is, necessarily, a theoretical studyof the algorithm. We analyze the convergence of fixed pointiterations that run in the background of IRKA, in particular themorphology of the mapping σ(k+1) = φ(σ(k)) (fixed points,periodic points and their classification). Other theoretical issuesinclude perturbation theory to analyze stability of the shifts,revealing relevant condition numbers, Cauchy–like structure ofcertain key matrices, connection of the fixed point iterations andpole placement, proper stopping criterion that translates into abackward stability relation, etc.Besides rich theory, IRKA also offers many numericalchallenges. How to implement the algorithm efficiently usingdirect or iterative solvers forV(σ) = ((σ1I−A)−1b, . . . , (σrI−A)−1b),W(σ) = ((σ1I−AT )−1c, . . . , (σrI−AT )−1c), and how todeal with numerical rank deficiency? How to adapt iterativesolvers in an inner loop that communicates with the outer fixedpoint iterations loop? When to stop? This requires estimates intheH2 space, and we show to combine them together with theusual stopping scheme for fixed point iterations. All these andmany other questions (e.g. implementations on modern parallelarchitectures such as CPU/GPU) are analyzed during softwaredevelopment. We will give some answers and illustrate theperformances of the software package.Zlatko DrmacDept. of MathematicsUniversity of [email protected]

Christopher BeattieDept. of MathematicsVirginia Polytechnic Institute and State [email protected]

Serkan GugercinDept. of MathematicsVirginia Polytechnic Institute and State [email protected]

Talk 4. Low rank deflative/iterative solutions of Lur’eequationsThe bottleneck of model reduction by passivity-preservingbalanced truncation is the numerical solution of Lur’e equations.A typical approach is an a priori perturbation leading to analgebraic Riccati equation. This is however, from an analyticalpoint of view, insufficiently understood and, from a numericalpoint of view, leads to an ill-conditioned problem. Hence we arefollowing an alternative approach that is basically consisting oftwo steps:

a) ’Filter out’ the ’critical part’ of the Lur’e equations: Thiswill lead us to an eigenproblem of moderate complexity,the outcome is an algebraic Riccati equation on somesubspace.

2012 SIAM Conference on Applied Linear Algebra 42

b) Solve algebraic Riccati equation on the subspace.

We show that this method provides low rank approximativesolutions of the Lur’e equations. Especially this makes thepresented method feasible for large-scale model order reductionproblems.Timo ReisFachbereich MathematikUniversitat [email protected]

Federico PoloniInstitut fur MathematikTechnische Universitat [email protected]

MS 46. Structured solution of nonlinear matrixequations and applications - Part II of IITalk 1. Inertia and rank characterizations of the expressionsA−BXB∗ − CY C∗ and A−BXC∗ ± CX∗B∗In this talk we consider the admissible inertias and ranks of theexpressions A−BXB∗ − CY C∗ and A−BXC∗ ± CX∗B∗with unknowns X and Y in the four cases when theseexpressions are: (i) complex self-adjoint, (ii) com- plexskew-adjoint, (iii) real symmetric, (iv) real skew symmetric. Wealso pro- vide a construction for X and Y to achieve the desiredinertia/rank, that uses only unitary/orthogonal transformationthus leading to a numerically reliable construction.Consequently, necessary and sufficient solvability conditions formatrix equations

A−BXB∗ − CY C∗ = 0;

andA−BXC∗ ± CX∗B∗ = 0

are provided.Delin ChuDept. of MathematicsNational University of Singapore, [email protected]

Talk 2. Structure-preserving Arnoldi-type algorithm forsolving eigenvalue problems in leaky surface wavepropagationWe study the generalized eigenvalue problems (GEP) arisingfrom modeling leaky surface waves propagation in a acousticresonator with infinitely many periodically arranged interdigitaltransducers. The constitution equations are discretized by finiteelement method with mesh refinement along the electrodeinterface and corners. The associated GEP is then transformed toa T-palindromic quadratic eigenvalue problem so that theeigenpairs can be accurately and efficiently computed by usingstructure-preserving algorithm with a generalizedT-skew-Hamiltonian implicity-restarted Arnoldi method. Ournumerical results show that the eigenpairs produced by theproposed structure-preserving method not only preserve thereciprocal property but also possess high efficiency and accuracy.Tsung-Ming HuangDept. of MathematicsNational Taiwan Normal University, Taipei 116, [email protected]

Wen-Wei LinDept. of Applied MathematicsNational Chiao Tung University, Hsinchu 300, [email protected]

Chin-Tien WuDept. of Applied MathematicsNational Chiao Tung University, Hsinchu 300, [email protected]

Talk 3. Structure-preserving curve for symplectic pairsWe study the stabilizing solution of the equationX +A∗X−1A = Q, where Q is Hermitian positive definite.We construct a smooth curve, parameterized by t ≥ 1, ofsymplectic pairs with a special structure, in which the curvepasses through all iteration points generated by the knownnumerical methods, including the fixed-point iteration,structured preserving doubling algorithm and Newton’s method,under specified conditions. We give a necessary and sufficientcondition for the existence of this structured symplectic pairsand characterize the behavior of this curve. We also use thiscurve to measure the convergence rates of these numericalmethods. Some numerical results are presented.Yueh-Cheng KuoDept. of Applied MathematicsNational University of Kaohsiung, Kaohsiung, 811, [email protected]

Shih-Feng ShiehDept. of MathematicsNational Taiwan Normal University, Taipei, [email protected]

Talk 4. A doubling algorithm with shift for solving anonsymmetric algebraic Riccati equationIn this talk, we analyze a special instance of nonsymmetricalgebraic matrix Riccati equation (NARE) arising from transporttheory. Traditional approaches for finding the minimalnonnegative solution of NARE are based on the fixed pointiteration and the speed of the convergence is linear. Recently, astructure-preserving doubling algorithm (SDA) with quadraticconvergence is designed for improving the speed of convergence.Our contribution is to show that applied with a shifted technique,the SDA is guaranteed to converge quadratically with nobreakdown. Also, we modify the conventional simple iterationalgorithm to dramatically improve the speed of convergence.Chun-Yueh ChiangCenter for General EducationNational Formosa [email protected]

Matthew M. LinDepartment of MathematicsNational Chung Cheng [email protected]

MS 47. Generalized inverses and applications - Part IIof IITalk 1. On a partial order defined on certain matricesFor a given matrix A ∈ Cm×n, we recall that the weightedMoore-Penrose inverse with respect to two Hermitian positivedefinite matrices M ∈ Cm×m and N ∈ Cn×n is the uniquesolution X ∈ Cn×m satisfying the equations: AXA = A,XAX = X , (MAX)∗ = MAX , (NXA)∗ = NXA. Thismatrix will be used to define a partial order on certain class ofcomplex matrices. Some properties on predecessors andsuccessors of a given matrix in that class will be established.This paper has been partially supported by Universidad Nacionalde La Pampa (Facultad de Ingenierıa) of Argentina (grant Resol.N. 049/11) and by the Ministry of Education of Spain (GrantDGI MTM2010-18228).

43 2012 SIAM Conference on Applied Linear Algebra

Nestor ThomeInstituto Universitario de Matematica MultidisciplinarUniversitat Politecnica de Valencia, [email protected]

Araceli HernandezFacultad de IngenierıaUniversidad Nacional de La Pampa, General Pico, [email protected]

Marina LattanziFacultad de Ciencias Exactas y NaturalesUniversidad Nacional de La Pampa, Santa Rosa, [email protected]

Fabian UrquizaFacultad de IngenierıaUniversidad Nacional de La Pampa, General Pico, [email protected]

Talk 2. Generalized inverses and path productsLet M = [aij ] be a lower triangular matrix over an arbitrary ringwith unity 1, and set ai := aii. We may split M asM = D +N , where D = diag(a1, . . . , an) and N = M −D.The latter is strictly lower triangular and hence nilpotent.Associated with the matrix M , we will consider the weighteddigraph G = (V,E) where V = Si is a set of nodes (or sites)and E = (Si, Sj) ∈ V × V is a collection of arcs such that(Si, Sj) ∈ E if aij 6= 0. We add a loop at site Si if ai 6= 0. Wewill relate the existence of the generalized inverse, namely vonNeumann, group and the Moore-Penrose inverse, of a lowertriangular matrix again gives a matrix of the same type by meansof path products obtained in the graph associated to the matrix.Pedro PatrıcioDepartamento de Matematica e AplicacoesUniversidade do Minho, [email protected]

R. HartwigDept. of MathematicsN.C.S.U. Raleigh, [email protected]

Talk 3. On structured condition numbers for a linearfunctional of Tikhonov regularized solutionA structured componentwise and normwise perturbation analysisof Tikhonov regularization problems are presented. Thestructured matrices we consider include: Toeplitz, Hankel,Vandermonde, and Cauchy matrices. Structured normwise,mixed and componentwise condition numbers for theseTikhonov regularization problems are introduced and theirexpressions are derived. Such expressions for many other classesof matrices can be similarly derived. By means of the powermethod, the fast condition estimated algorithms are proposed.The condition numbers and perturbation bounds are examinedon some numerical examples and compared with unstructurednormwise, mixed and componentwise condition numbers. Thenumerical examples show that the structured mixed conditionnumbers give better perturbation bounds than others.Yimin WeiDepartment of MathematicsFudan University, P.R. [email protected]

Huaian DiaoSchool of Mathematics and StatisticsNortheast Normal University, Chang Chun 130024, P.R. [email protected]

Talk 4. Explicit characterization of the Drazin indexLet B (X) be the set of bounded linear operators on a Banachspace X , and A ∈ B (X) be Drazin invertible. An element

B ∈ B (X) is said to be a stable perturbation of A if B is Drazininvertible and I −Aπ −Bπ is invertible, where I is the identityoperator on X , Aπ and Bπ are the spectral projectors of A andB respectively. Under the condition that B is a stableperturbation of A, a formula for the Drazin inverse BD isderived. Based on this formula, a new approach is provided tothe computation of the explicit Drazin indices of certain 2× 2operator matricesQingxiang XuDepartment of MathematicsShanghai Normal University, Shanghai, 200234, PR [email protected]

MS 48. Parallelization of efficient algorithmsTalk 1. A highly scalable error-controlled fast multipolemethodWe present a linear scaling, error-controlled FMMimplementation for long-range interactions of particle systemswith open, 1D, 2D and 3D periodic boundary conditions.Similarly to other fast summation algorithms the FMM allows toreduce the total runtime significantly. This runtime advantagehowever comes with considerably increased memoryrequirements posing constraints to the overall particle systemsize. In this talk we focus on the reduced memory footprint aswell as the communication pattern for trillions of particles. Thecurrent code is designed to lower the memory consumption toonly 45Byte/particle already including a small 16.2% paralleloverhead for 300k MPI ranks.Ivo KabadshowJulich Supercomputing CentreResearch Centre [email protected]

Holger DachselJulich Supercomputing CentreResearch Centre [email protected]

Talk 2. A parallel fast Coulomb solver based onnonequispaced Fourier transformsThe fast calculation of long-range Coulomb interactions is acomputational demanding problem in particle simulation.Therefore, several fast approximate algorithms have beendeveloped, which reduce the quadratic arithmetic complexity ofthe plain summation to linear complexity (up to a logarithmicfactor). This talk focuses on Fourier based methods with specialattention to the application of the nonequispaced fast Fouriertransform and its parallelization.We present a massively parallel Coulomb solver software libraryfor distributed memory architectures. The underlying fastalgorithms for periodic and non-periodic boundary conditionswill be explained and extensive performance evaluations will bepresented.Michael PippigDept. of MathematicsChemnitz University of [email protected]

Talk 3. Generalized fast Fourier transforms via CUDAThe fast Fourier transform (FFT) belongs to the algorithms withlarge impact on science and engineering. By appropriateapproximations, this scheme has been generalized for arbitraryspatial sampling points. We discuss the computational costs indetail for this so called nonequispaced FFT and its variations.

2012 SIAM Conference on Applied Linear Algebra 44

Because of the evolution of programmable graphic processorunits into highly parallel, multithreaded, manycore processorswith enormous computational capacity and very high memorybandwith, we parallelized the nonequispaced FFT by means ofthe so called Compute Unified Device Architecture (CUDA)using the CUDA FFT library and a dedicated parallelization ofthe approximation scheme.Susanne KunisDept. of MathematicsUniversity of [email protected]

Talk 4. Efficient regularization and parallelization for sparsegrid regressionRegression, (high-dimensional) function reconstruction fromscattered data, is a common problem in data-driven tasks.Typically, meshfree methods are employed to circumvent thecurse of dimensionality. To deal with regression by discretizingthe feature space, sparse grids can be employed. Due to theirprimarily data-independent approach, sparse grids enable one todeal with massive amounts of data. Adaptive refinement thenallows to adapt to the pecularities of the problem at hand.To be able to deal with large, noisy datasets, efficient algorithmsand parallelizations have to be employed, and the full potentialof modern hardware architectures has to be exploited. This canbe complicated having to deal with hierarchical basis functionson multiple levels. We present the special challenges posed tohierarchical and recursive sparse grid algorithms and discussefficient solutions for regularization and parallelization.Dirk PflugerDept. of InformaticsTechnische Universitat [email protected]

Alexander HeineckeDept. of InformaticsTechnische Universitat [email protected]

MS 49. Analysis and computation on matrix manifoldTalk 1. Best low multilinear rank approximation of symmetrictensors by Jacobi rotationsWe consider third-order symmetric tensors and seek their bestlow multilinear rank approximations. The proposed algorithm isbased on Jacobi rotations and symmetry is preserved at eachiteration. Examples are provided that illustrate the need of suchalgorithms. Our algorithm converges to stationary points of theobjective function. The proof of convergence can be consideredas an advantage of the algorithm over existingsymmetry-preserving algorithms in the literature.Pierre-Antoine AbsilDepartment of Mathematical Engineering, ICTEAM InstituteUniversite catholique de [email protected]

Mariya IshtevaSchool of Computational Science and EngineeringGeorgia Institute of [email protected]

Paul Van DoorenDepartment of Mathematical Engineering, ICTEAM InstituteUniversite catholique de [email protected]

Talk 2. Differential geometry for tensors with fixedhierarchical Tucker rank

A number of authors have recently proposed several geometriesfor rank-structured matrix and tensor spaces, namely formatrices or tensors with fixed matrix, Tucker, and TT rank. Inthis talk we present a unifying approach for establishing asmooth, differential structure on the set of tensors with fixedhierarchical Tucker rank. Our approach describes this set as asmooth submanifold, globally embedded in the space of realtensors. The previous spaces are shown to be specific instancesof a particular hierarchical Tucker tree (possibly with additionalconstraints on the frames). As numerical example we show howthis geometry can be used to dynamically update a time-varyingtensor. This approach compares favorable to point-wiseSVD-based computations.Bart VandereyckenMathematics Section, ANCHPEcole Polytechnique Federale de [email protected]

Andre UschmajewDepartment of MathematicsTechnische Universitat (TU) [email protected]

Talk 3. Deterministic approaches to the Karcher mean ofpositive definite matricesWe propose a deterministic approach to the Karcher mean ofpositive definite matrices via geometric power means. For amatrix geometric mean G, we construct one-parameter group Gtof geometric means varying continuously over t ∈ [−1, 1] \ 0and approaching the Karcher mean as t→ 0. Each of thesemeans arises as unique positive definite solution of a non-linearmatrix equation and has the distance less ≤

√t/2 to the

Karcher mean. This provides not only a structured anddeterministic sequence of matrix means converging to theKarcher mean, but also a simple proof of the monotonicity of theKarcher mean, conjectured by Bhatia and Holbrook, and othernew properties, which have recently been established by Lawsonand Lim and also Bhatia and Karandikar using probabilisticmethods on the metric structure of positive definite matricesequipped with the trace metric.Yongdo LimDept. of MathematicsKyungpook National [email protected]

Talk 4. The Karcher mean: first and second orderoptimization techniques on matrix manifoldsIn this talk, we present a collection of implementations for theKarcher mean, which is a specific instance of the matrixgeometric mean. The Karcher mean is defined as the solution toan optimization problem on the manifold of positive definitematrices, where it exhibits an appealing analogy with thearithmetic mean. Generalization of classical optimizationschemes results in Riemannian optimization, where the intrinsicproperties of the manifold are maintained and exploitedthroughout the algorithm. We examine several optimizationtechniques, such as SD, CG, Trust Region, and BFGS, andcompare the results with the ALM, BMP and CHEAP meanalgorithms.Ben JeurisDept. of Computer ScienceKatholieke Universiteit [email protected]

Raf Vandebril

45 2012 SIAM Conference on Applied Linear Algebra

Dept. of Computer ScienceKatholieke Universiteit [email protected]

Bart VandereyckenMathematics Institute of Computational Science and EngineeringEcole Polytechnique Federale de [email protected]

MS 50. Advanced methods for large eigenvalueproblems and their applicationsTalk 1. DQDS with aggressive early deflation for computingsingular valuesThe DQDS algorithm is the standard method for computing allthe singular values of a bidiagonal matrix with high relativeaccuracy. Its efficient implementation is now available as aLAPACK subroutine DLASQ. In order to reduce the DQDSruntime, we incorporate into DQDS a technique calledaggressive early deflation, which has been applied successfullyto the Hessenberg QR algorithm. In addition, a shift-free versionof our algorithm has a potential to be parallelized in a pipelinedfashion. Our mixed forward-backward stability analysis provesthat with our proposed deflation strategy, DQDS computes allthe singular values to high relative accuracy.Kensuke AishimaDept. of Mathematics of InformaticsUniversity of TokyoKensuke [email protected]

Yuji NakatsukasaSchool of MathematicsUniversity of [email protected]

Ichitaro YamazakiDept. of Computer ScienceUniversity of [email protected]

Talk 2. A scalable parallel method for large-scale nonlineareigenvalue problemsIn this presentation, we present parallel software for nonlineareigenvalue problems that has interfaces for PETSc. The softwareimplements an eigensolver based on contour integral. It finds apartial set of eigenpairs of large sparse nonlinear eigenvalueproblems and has high scalability. We demonstrate parallelperformance of the implemented method through numericalexperiments arising from several practical applications.Kazuma YamamotoDept. of Computer ScienceUniversity of [email protected]

Yasuyuki MaedaDept. of Computer ScienceUniversity of [email protected]

Yasunori FutamuraDept. of Computer ScienceUniversity of [email protected]

Tetsuya SakuraiDept. of Computer ScienceUniversity of [email protected]

Talk 3. Application of the Sakurai-Sugiura method in the fieldof density functional theory on highly parallel systemsDensity Functional Theory (DFT) is one of the most importantmethods in computational material science. Despite the steadily

increasing computational power, computation time is still thelimiting factor for many systems of interest. In many cases themost time consuming part is solving a series of generalizedeigenvalue problems that emerge inside an iterative loop. In thecontext of Siesta, a widely spread DFT software, theSakurai-Sugiura method is a promising approach. This talk willshow how, due to its three obvious levels of parallelization, thisalgorithm scales much better than other widely used libraries(e.g. ScaLAPACK), how it offers possibilities to use many GPUsin parallel, and how it also supports dealing with sparse matrices.Georg HuhsBarcelona Supercomputing Center,Centro Nacional de Supercomputacion, Barcelona, [email protected]

Talk 4. MERAM for neutron physics applications using YMLenvironment on post petascale heterogeneous architectureEigenvalue problem is one of the key elements in neutronphysics applications. Studying and designing efficient andsacalable eigenvalue solvers is thus necessary for post petascaleneutrons physics applications. In this talk, after recalling theprinciple of multiple explicitly restarted Arnoldi method(MERAM), we will present the design model used for itsimplementation on distributed heterogeneous architectures. Theperformance results on CURIE and GRID5000 platformsmaking use of the software environments KASH (Krylov bAsedSolvers for Hybrid architectures) YML and PETSc/SLEPc fortypical neutron physics problems will be presented.Christophe CalvinCEA DEN DANS DM2S SERMA CEA [email protected]

Nahid EmadUniversity of [email protected]

Serge G. PetitonINIRIA University of [email protected]

Jerome DuboisCEA DEN DANS DM2S SERMA CEA [email protected]

Makarem DandounaUniversity of [email protected]

MS 51. Accurate and verified numerical computationsfor numerical linear algebraTalk 1. Product decomposition and its applicationsA product decomposition algorithm of a real number on floatingpoint system is proposed. The product decomposition of a realnumber x is a floating point decomposition defined by

x ' x1 (1 + x2) (1 + x3) · · · (1 + xn) ,

where xi denote floating point numbers (1 ≤ i ≤ n). x1 impliesan approximation of x and xi (2 ≤ i ≤ n) involveapproximations of relative errors of an approximate value by thedecomposition using x1, x2, · · · , xi−1. This decomposition isused in numerical analysis to calculate the accurate logarithmvalue and so on. In this talk we present an efficient algorithm toconstruct the product decomposition and its applications.Naoya YamanakaResearch Institute for Science and Engineering, Waseda Universitynaoya [email protected]

Shin’ichi Oishi

2012 SIAM Conference on Applied Linear Algebra 46

Faculty of Science and Engineering, Waseda [email protected]

Talk 2. The MPACK: multiple precision version of BLAS andLAPACKWe have been developing a multiple precision version of linearalgebra package based on BLAS and LAPACK. We translatedand reimplimented FORTRAN by C++, and the MPACKsupports GMP, MPFR, and QD (DD) multiple precisionlibraries; users can choose the library on their needs. CurrentlyBLAS part is completed and 100 LAPACK routines areimplemented and well tested. Moreover DD version ofmatrix-matrix multiplication routine has been accelerated usingNVIDIA C2050 GPU. Development is ongoing athttp://mplapack.sourceforge.net/, and available under opensource (2-clause BSD) license.Maho NakataAdvanced Center for Computing and Communication, [email protected]

Talk 3. On eigenvalue computations of nonderogatorymatricesIn this talk I present some problems and results concerning theeigenvalue problem of nonderogatory matrices. The first groupof problems is related to the detection of multiple eigenvalues ofunreduced upper Hessenberg matrices and their refinements inmultiple floating point arithmetic. The second group of problemsis the perturbation of invariant subspaces and itscharacterizations in terms of the matrix perturbation.Aurel GalantaiInstitute of Intelligent Engineering SystemsObuda University, 1034 Budapest, Becsi ut 96/b, [email protected]

Talk 4. Verified solutions of sparse linear systemsAlgorithms for calculating verified solutions of sparse linearsystems are proposed. The proposed algorithms are based onstandard numerical algorithms for a block LDLT factorizationand error estimates for specified eigenvalues by Lehmann’stheorem. Numerical results are presented for illustrating theperformance of the proposed algorithms.Takeshi OgitaDivision of Mathematical SciencesTokyo Woman’s Christian [email protected]

MS 52. Numerical linear algebra libraries for high endcomputing - Part I of IITalk 1. Large-scale eigenvalue computation with PETSc andYMLIn the context of parallel and distributed computation, thecurrently existing numerical libraries do not allow sequential andparallel code reuse. Besides, they are not able to exploit themulti-level parallelism offered by modern emerging numericalmethods.In this talk, we present a design model for numerical librariesbased on a component approach allowing code reuse andproblem solving scalability. We present then, an implementationof this design using YML scientific workflow environmentjointly with the object oriented library PETSc. Some numericalexperiments on GRID5000 platform and NERSC computersvalidate our approach and show its efficiency.Makarem DandounaPRISM laboratory, University of Versailles, France

[email protected]

Nahid EmadPRISM laboratory University of Versailles, [email protected]

L. A. (Tony) DrummondComputational Research Division, Lawrence Berkeley NationalLaboratory, [email protected]

Talk 2. Sparse matrix-matrix operations in PETScSparse matrix-matrix operations, A ∗B, AT ∗B (or A ∗BT )and PT ∗A ∗ P (or R ∗A ∗RT ), are computational kernels inthe PETSc library. Recent addition of a geometric-algebraicmultigrid preconditioner requires these matrix operations to bescalable to tens of thousands processors cores, which forces usto take an innovated approach in algorithm design andimplementations for these operations, including some relevantdata structures. In this talk, we will present our newly developedscalable sparse matrix-matrix algorithms, implementations andperformance, along with lessons learned and experiences gainedfrom this work.Hong ZhangDept. of Computer Science Department, Illinois Institute of Technology,USA and Argonne National Laboratory, [email protected]

Barry Smith and PETSc TeamMathematics and Computer Science Division, Argonne NationalLaboratory, [email protected]

Talk 3. Hierarchical QR factorization algorithms formulti-core cluster systemsThis paper describes a new QR factorization algorithm which isespecially designed for massively parallel platforms combiningparallel distributed multi-core nodes. These platforms make thepresent and the foreseeable future of high-performancecomputing. Our new QR factorization algorithm falls in thecategory of the tile algorithms which naturally enables good datalocality for the sequential kernels executed by the cores (highsequential performance), low number of messages in a paralleldistributed setting (small latency term), and fine granularity(high parallelism).Julien LangouUniversity of Colorado Denver, USAsupported by the National Science Foundation grant # NSF [email protected]

Jack DongarraUniversity of Tennessee Knoxville, USAOak Ridge National Laboratory, USAManchester University, [email protected]

Mathieu FavergeUniversity of Tennessee Knoxville, [email protected]

Thomas HeraultUniversity of Tennessee Knoxville, [email protected]

Yves RobertUniversity of Tennessee Knoxville, USAEcole Normale Superieure de Lyon, [email protected]

Talk 4. Towards robust numerical algorithms for exascalesimulationThe advent of exascale machines will require the use of parallelresources at an unprecendent scale, leading to a high rate of

47 2012 SIAM Conference on Applied Linear Algebra

hardware faults. High Performance Computing applications thataim at exploiting all these resources will thus need to beresilient, i.e., being able to compute a correct output in presenceof faults. Contrary to checkpointing techniques or AlgorithmBased Fault Tolerant (ABFT) mechanisms, strategies based oninterpolation for recovering lost data do not require extra workor memory when no fault occurs. We apply this latter strategy toKrylov iterative solvers, which are often the most computationalintensive kernels in HPC simulation codes. Our maincontribution is the proposition and discussion of several variantscompared to previous works. For that, we propose a new variantfor recovering data, we study the occurrence of multiple faults,we consider the GMRES, CG and BICGSTAB solvers, and weinject faults according to an advanced model of faultdistribution. We assess the impact of the recovery method, thefault rate and the number of processors on resilience. Ratherthan implementing a particular actual parallel code, we assess allour strategies based on sequential Matlab implementations thatsimulate parallel executions.Emmanuel AgulloINRIA Bordeaux - Sud-Ouest, 351, Cours de la Liberation, BatimentA29 bis, 33405 Talence Cedex [email protected]

Luc GiraudINRIA Bordeaux - Sud-Ouest, [email protected]

Abdou GuermoucheINRIA Bordeaux - Sud-Ouest, [email protected]

Jean RomanINRIA, [email protected]

Mawussi ZounonINRIA, [email protected]

MS 53. Efficient preconditioners for real worldapplications - Part I of IITalk 1. A parallel factored preconditioner for non-symmetriclinear systemsThe efficient solution to non-symmetric linear systems is still anopen issue on parallel computers. In this communication wegeneralize to the non-symmetric case the Block FSAI (BFSAI)preconditioner which has already proved very effective onsymmetric problems arising from different applications. BFSAIis a hybrid approach combining an “inner” preconditioner, withthe aim of transforming the system in a block diagonal one, withan “outer” one, a block diagonal incomplete decomposition,intended to decrease the conditioning of each block. Theproposed algorithm is experimented with in a number of largesize problems showing both good robustness and scalability.Carlo JannaDept. of Mathematical Methods and Models for Scientific ApplicationsUniversity of Padova, [email protected]

Massimiliano FerronatoDept. of Mathematical Methods and Models for Scientific ApplicationsUniversity of Padova, [email protected]

Giorgio PiniDept. of Mathematical Methods and Models for Scientific ApplicationsUniversity of Padova, [email protected]

Talk 2. Preconditioning for linear least-squares problemsIn this talk we deal with iterative methods for solving large andsparse linear least squares problems. In particular we describetwo new preconditioning techniques for the CGLS method. Firstwe consider the strategy which is based on the LU factorization.Our approach includes a new reordering based on a specificweighting transversal problem. Direct preconditioning of thenormal equations by the balanced symmetric and positivedefinite factorization is our second approach. Numericalexperiments demonstrate effectiveness of the algorithmic andimplementational features of the new approaches.Miroslav TumaInstitute of Computer ScienceAcademy of Sciences of the Czech [email protected]

Rafael BruInstitut de Matematica MultidisciplinarUniversitat Politecnica de [email protected]

Jose MasInstitut de Matematica MultidisciplinarUniversitat Politecnica de [email protected]

Jose MarınInstitut de Matematica MultidisciplinarUniversitat Politecnica de [email protected]

Talk 3. Robust and parallel preconditioners for mechanicalproblemsWe consider the simulation of displacements under loading formechanical problems. Large discontinuities in materialproperties, lead to ill-conditioned systems of linear equations,which leads to slow convergence of the PreconditionedConjugate Gradient (PCG) method. This paper considers theRecursively Deflated Preconditioned Conjugate Gradient(RDPCG) method for solving such systems. Our deflationtechnique uses as deflation space the rigid body modes of sets ofelements with homogeneous material properties. We show thatin the deflated spectrum the small eigenvalues are mapped tozero and no longer negatively affect the convergence. We justifyour approach through mathematical analysis and we show withnumerical experiments on both academic and realistic testproblems that the convergence of our RDPCG method isindependent of discontinuities in the material properties.Kees VuikDelft Institute of Applied MathematicsTechnology University of Delft, The [email protected]

Tom JonsthovelDelft Institute of Applied MathematicsTechnology University of Delft, The [email protected]

Martin van GijzenDelft Institute of Applied MathematicsTechnology University of Delft, The [email protected]

Talk 4. Block factorized forms of SPAIIn this talk we present new results on block versions of sparseapproximate inverse preconditioners M for sparse matrix A. Weconsider the Frobenius norm minimization minM ‖AM − I‖F .Blocked versions are interesting because often the underlyingproblem introduces in a natural way a block structure.Furthermore, they allow a more efficient memory access, and

2012 SIAM Conference on Applied Linear Algebra 48

they reduce the number of least squares problems that have to beconsidered in the construction of the preconditioner M . We areinterested in determining appropriate block patterns for a generalsparse matrix. Therefore, we want to combine columns of Mwith similar least squares problems to blocks in order to reducethe number of least squares problems that have to be solved forconstructing the preconditioner. Furthermore, given an arbitraryblocking we also have to find the nonzero blocks that we have toinclude to derive a good preconditioner.Thomas HuckleInstitut fur InformatikTechnische Universitat Munchen, [email protected]

Matous SedlacekInstitut fur InformatikTechnische Universitat Munchen, [email protected]

MS 54. Solving ill-posed systems via signal-processingtechniques - Part I of IITalk 1. Sequential updates for L1 minimization: sparseKalman filtering, reweighted L1, and moreSparse signal recovery often involves solving an L1-regularizedoptimization problem. Most of the existing algorithms focus onthe static settings, where the goal is to recover a fixed signalfrom a fixed system of equations. In this talk, we present acollection of homotopy-based algorithms that dynamicallyupdate the solution of the underlying L1 problem as the systemchanges. The sparse Kalman filter solves an L1-regularizedKalman filter equation for a time-varying signal that follows alinear dynamical system. Our proposed algorithm sequentiallyupdates the solution as the new measurements are added and theold measurements are removed from the system.Justin RombergElectrical and Computer EngineeringGeorgia Institute of [email protected]

M. Salman AsifElectrical and Computer EngineeringGeorgia Institute of [email protected]

Talk 2. Solving basis pursuit: infeasible-point subgradientalgorithm, computational comparison, and improvementsWe propose a subgradient algorithm called ISAL1 for thel1-minimization (Basis Pursuit) problem which appliesapproximate projections via a truncated CG scheme. We willalso present results of an extensive computational comparison ofour method and various state-of-the-art l1-solvers on a largetestset. It turns out our algorithm compares favorably. Moreover,we show how integrating a new heuristic optimality check calledHSE can improve the solution speed and accuracy of several ofthese solvers.Andreas TillmannInstitute for Mathematical OptimizationTechnische Universitat [email protected]

Dirk LorenzInstitute for Analysis and AlgebraTechnische Universitat [email protected]

Marc PfetschInstitute for Mathematical OptimizationTechnische Universitat Braunschweig

[email protected]

Talk 3. Semismooth Newton methods with multi-dimensionalfilter globalization for l1 optimizationWe present a class of methods for l1-regularized optimizationproblems. They are based on a flexible combination ofsemismooth Newton steps and globally convergent descentmethods. A multidimensional filter framework is used to controlthe acceptance of semismooth Newton steps. We prove globalconvergence and transition to fast local convergence for both theconvex and the nonconvex case. Numerical results show theefficiency of the method.Andre MilzarekDepartment of Mathematics, TopMathTechnische Universitat [email protected]

Michael UlbrichChair of Mathematical OptimizationTechnische Universitat [email protected]

Talk 4. Improved first-order methods: how to handleconstraints, non-smoothness, and slow convergenceThere are many specialized solvers that solve specific convexprograms efficiently, but few algorithms can deal with generalcomplicated constraints and non-smooth functions. To addressthese difficult problems, we introduce a framework and softwarepackage called TFOCS (Becker/Candes/Grant). The methodrelies on two tricks: dualization and smoothing. This talkdescribes the framework and also discusses recent splittingmethods such as the method by Chambolle and Pock and byCombettes et al. We also cover recent progress in improving theconvergence of first-order algorithms by using non-diagonalpreconditioners (with J. Fadili).Stephen BeckerFSMP, laboratoire Jacques-Louis LionsUPMC (Paris-6)–[email protected]

Jalal FadiliCNRS-ENSICAENUniversite de [email protected]

Emmanuel CandesDepartments of Mathematics and StatisticsStanford [email protected]

Michael GrantCVX [email protected]

MS 55. Max-algebra - Part I of IITalk 1. Tropical bounds for the eigenvalues of structuredmatricesWe establish several inequalities of log-majorization type,relating the moduli of the eigenvalues of a complex matrix ormatrix polynomial with the tropical eigenvalues of auxiliarymatrix polynomials. This provides bounds which can becomputed by combinatorial means. We consider in particularstructured matrices and obtain bounds depending on the normsof block submatrices and on the pattern (graph structure) of thematrix.Marianne AkianINRIA Saclay–Ile-de-France and CMAP

49 2012 SIAM Conference on Applied Linear Algebra

CMAP, Ecole Polytechnique, Route de Saclay, 91128 Palaiseau Cedex,[email protected]

Stephane GaubertINRIA Saclay–Ile-de-France and CMAPCMAP, Ecole Polytechnique, Route de Saclay, 91128 Palaiseau Cedex,[email protected]

Meisam SharifyINRIA Saclay–Ile-de-France and LRILRI, Bat 650 Universite Paris-Sud 11 91405 Orsay Cedex, [email protected]

Talk 2. Sensitivity in extremal systems of linear equations andinequalitiesA survey of some recent results concerning the properties offinite systems of linear equations and inequalities in extremalalgebra will be presented. Problems connected with sensitivityand parametric analysis of such systems will be discussed.Possible applications of the results for post-optimal analysis ofoptimization problems, the set of feasible solutions of which isdescribed by the systems of (max, min) and (max, plus) linearsystems, will be shown. The objective functions of theoptimization problems are expressed as the maximum of finitelymany continuous functions, each depending on one variable.Karel ZimmermannFaculty of Mathematics and PhysicsCharles University in Prague, Czech [email protected]

Martin GavalecFaculty of Informatics and ManagementUniversity of Hradec Kralove, Czech [email protected]

Talk 3. Multiplicative structure of tropical matricesI shall report on a programme of research aiming to understandthe structure of tropical (max-plus) matrix semigroups. Thisstructure turns out to be intimately connected with the geometryof tropical convexity; indeed, it transpires that almost everyalgebraic property of the full n× n tropical matrix semigroupmanifests itself in some beautiful geometric phenomenoninvolving polytopes. Various parts of the programme are jointwork with people including Christopher Hollings, Zur Izhakianand Marianne Johnson.Mark KambitesSchool of MathematicsUniversity of [email protected]

Talk 4. Transience bounds for matrix powers in max algebraIn this talk we demonstrate how the concept of CSR expansionsdeveloped by Schneider and Sergeev helps to unify and comparethe bounds on periodicity transient existing in the literature.Unlike in the theory of graph exponents, these bounds are notstrongly polynomial. To this end, we also present somemax-algebraic extensions of polynomial bounds on graphexponents, due to Wielandt and Schwartz.Sergeı SergeevSchool of MathematicsUniversity of Birmingham, [email protected]

MS 56. Eigenvalue perturbations and pseudospectra -Part I of IITalk 1. Inclusion theorems for pseudospectra of blocktriangular matrices

The ε-pseudospectrum of A ∈ Cn×n, denoted by σε(A), is theunion of the spectra of the matrices A+ E, where E ∈ Cn×nand ‖E‖2 ≤ ε. In this talk we consider inclusion relations of theform

σf(ε)(A11) ∪ σf(ε)(A22) ⊆ σε

([A11 A12

0 A22

])⊆ σg(ε)(A11) ∪ σg(ε)(A22).

We derive formulae for f(ε) and g(ε) in terms ofsepλ(A11, A22) and ‖R‖2, where R is the solution of theSylvester equation A11R−RA22 = A12.Michael KarowDept. of MathematicsTechnical University of [email protected]

Talk 2. Conjectures on pseudospectra of matricesWe discuss some conjectures concerning coalescence points ofconnected components of pseudospectra of a square complexmatrix A. We call pseudospectrum of A of order j and level ε,Λε,j(A), to the set of eigenvalues of multiplicity ≥ j of matriceswhose distance to A is ≤ ε.• The coalescence points of the components of Λε,j(A) are

points where the pseudospectra Λε,j+1(A) arise from.

• The coalescence points of the components of Λε,1(A) arein line segments connecting eigenvalues of A, andanalogously for Λε,j(A).

Juan-Miguel GraciaDept. of Applied Mathematics and StatisticsUniversity of the Basque Country UPV/EHU, [email protected]

Talk 3. Sensitivity of eigenvalues of an unsymmetrictridiagonal matrixThe Wilkinson condition number ignores the tridiagonal formand so can be unduly pessimistic. We propose several relativecondition numbers that exploit the tridiagonal form. Some ofthese numbers are derived from different factored forms (orrepresentations) of the (possibly shifted) matrix and so they shedlight on which factored forms are best for computation. Weshow some interesting examples.Carla FerreiraMathematics and Applications DepartmentUniversity of Minho, [email protected]

Beresford ParlettUniversity of California, Berkeley

Froilan DopicoUniversidad Carlos III de Madrid, Spain

Talk 4. First order structured perturbation theory foreigenvalues of skew-adjoint matricesThe main goal of structured matrix perturbation theory is toidentify situations where eigenvalues are much less sensitive tostructured perturbations (i.e., those belonging to the same kindof matrices as the unperturbed one) than to unstructured ones. Inthis talk we analyze one such situation: the relevant structure isskew-adjointness with respect to an indefinite scalar product.Explicit formulas are obtained for both the leading exponent andthe leading coefficient of asymptotic perturbation expansionswhen the perturbations are taken to be also skew-adjoint. Using

2012 SIAM Conference on Applied Linear Algebra 50

the Newton diagram as the main tool, it is shown that the leadingcoefficient depends on both first (i.e., eigenvectors) and secondvectors in the longest Jordan chains associated with theeigenvalue under study.Julio MoroDepartamento de MatematicasUniversidad Carlos III de MadridMadrid (Spain)[email protected]

Marıa Jose PelaezDepartamento de MatematicasUniversidad Catolica del NorteAntofagasta (Chile)[email protected]

MS 57. Numerical linear algebra and optimization inimaging applications - Part I of IITalk 1. Some numerical linear algebra and optimizationproblems in spectral imagingIn this talk we overview some of our recent work on numericalalgorithms for spectral image analysis. Spectral imaging collectsand processes image information from across theelectromagnetic spectrum (often represented visually as a cube),and has a wide array of modern applications, for example inremote sensing for ecology and surveillance. Here we describesome of our work on the design and analysis of mathematicaltechniques for compressive sensing, processing, and analysis ofspectral data. Topics considered include: (1) random SVDmethods for dimensionality reduction, (2) joint reconstructionand classification of spectral images, and (3) applications ofunmixing, clustering and classification methods to targetidentification. Tests on real data are described.This represents is joint work with several people, including teammembers on projects funded by the U.S. Air Force Office ofScientific Research and the National Geospatial-IntelligenceAgency.Bob PlemmonsDepts. of Computer Science and MathematicsWake Forest [email protected]

Jennifer ErwayDept. of MathematicsWake Forest [email protected]

Nicolas GillisDept. of Combinatorics & OptimizationUniversity of [email protected]

Xiaofei HuDept. of MathematicsWake Forest [email protected]

Michael NgDept. of MathematicsHong Kong Baptist [email protected]

Paul PaucaDept. of Computer ScienceWake Forest [email protected]

Sudhakar PrasadDept. of Astronomy and PhysicsUniversity of New [email protected]

Jiani Zhang

Dept. of MathematicsWake Forest [email protected]

Qiang ZhangDept. of Health SciencesWake Forest [email protected]

Talk 2. Image restoration via constrained optimization: anapproach using feasible direction methodsImage restoration is an ill-posed inverse problem which requiresregularization. Regularization leads to reformulate the originalrestoration problem as a constrained optimization problemwhere the objective function is a regularization term and theconstraint imposes fidelity to the data.In this talk, we present a feasible direction method for obtainingrestored images as solutions of the optimization problem. Thepresented method computes feasible search directions byinexactly solving trust region subproblems whose radius isadjusted in order to maintain feasibility of the iterates.Numerical results are presented to illustrate the effectiveness andefficiency of the proposed method.Germana LandiDept. of MathematicsUniversity of [email protected]

Talk 3. On the solution of linear systems in Newton-typemethods for image reconstructionSome imaging applications are usually modeled as aminimization problem with nonnegative constraints on thesolution. For large size problems, projected-Newton methods arevery attractive because of their fast convergence. In order tomake them computationally competitive, the inner linear systemfor the search direction computation should be solved efficiently.In this talk we focus on the solution of the linear system whendifferent objective functions are considered. The objectivefunction is related to the application, to the noise affecting therecorded image and to the kind of regularization chosen.Elena PiccolominiDept. of MathematicsUniversity of [email protected]

Talk 4. Alternating direction optimization for convex inverseproblems. Application to imaging and hyperspectralunmixingIn this talk I will address a new class of fast of algorithms forsolving convex inverse problems where the objective function isa sum of convex terms with possibly convex constraints. Usually,one of terms in the objective function measures the data fidelity,while the others, jointly with the constraints, enforce some typeof regularization on the solution. Several particular features ofthese problems (huge dimensionality, nonsmoothness) precludethe use of off-the-shelf optimization tools and have stimulated aconsiderable amount of research. In this talk, I will present anew class of algorithms to handle convex inverse problemstailored to image recovery applications and to hyperspectralunmixing. The proposed class of algorithms is an instance of theso-called alternating direction method of multipliers (ADMM),for which convergence sufficient conditions are known. Weshow that these conditions are satisfied by the proposedalgorithms. The effectiveness of the proposed approach isillustrated in a series of imaging inverse problems, including

51 2012 SIAM Conference on Applied Linear Algebra

deconvolution and reconstruction from compressiveobservations, and hyperspectral unmixing problems, includingsparse unmixing and positive matrix factorization.Jose Bioucas-DiasInstituto de Telecomunicaes and Department of Electrical and ComputerEngineering,Instituto Superior Tcnico, [email protected]

Mario FigueiredoInstituto de Telecomunicaes and Department of Electrical and ComputerEngineering,Instituto Superior Tcnico, [email protected]

MS 58. Parametric eigenvalue problems - Part I of IITalk 1. Computing double eigenvalues via the two-parametereigenvalue problemA task of computing all values of the parameter λ, such that thematrix A+ λB has a double eigenvalue, can be interpreted as asingular (quadratic) two-parameter eigenvalue problem. Using anumerical method for the singular two-parameter eigenvalueproblem it is possible to obtain all solutions as eigenvalues of acertain generalized eigenvalue problem.Bor PlestenjakIMFM & Department of MathematicsUniversity of [email protected]

Andrej MuhicIMFM & Department of MathematicsUniversity of [email protected]

Talk 2. Lyapunov inverse iteration for identifying Hopfbifurcations in models of incompressible flowThe identification of instability in large-scale dynamical systemscaused by Hopf bifurcation is difficult because of the problem ofcomputing the rightmost pair of complex eigenvalues of largesparse generalised eigenvalue problems. A method developed in[Meerbergen & Spence, SIMAX (2010), pp.1982-1999] avoidsthis computation, instead performing an inverse iteration for acertain set of real eigenvalues that requires the solution of alarge-scale Lyapunov equation at each iteration. This talkdiscusses a refinement of the method of Meerbergen & Spenceand tests it on challenging problems arising from fluid dynamics.Alastair SpenceDept. of Mathematical SciencesUniversity of Bath, [email protected]

Howard ElmanDept. of Computer ScienceUniversity of Maryland, [email protected]

Karl MeerbergenDept. of Computer ScienceKatholieke Universiteit Leuven, [email protected]

Minghao WuDept. of Applied Mathematics & StatisticsUniversity of Maryland, [email protected]

Talk 3. A quadratically convergent algorithm for matrixdistance problemsWe discuss a method for the computation of the distance of astable matrix to the unstable matrices with respect to the open

left-half plane. First, we provide a fast algorithm to compute thecomplex unstructured stability radius. Second, based on aformula by Qiu et al. (Automatica, 31 (1995), pp. 879–890) wegive a new fast method to compute the real structured stabilityradius. Both algorithms are based on finding Jordan blockscorresponding to a pure imaginary eigenvalue in aparameter-dependent Hamiltonian eigenvalue problem.Numerical results show the performance of both algorithms forseveral examples.Melina A. FreitagDept. of Mathematical SciencesUniversity of [email protected]

Alastair SpenceDept. of Mathematical SciencesUniversity of [email protected]

Talk 4. A real Jacobi-Davidson algorithm for the2-real-parameter eigenvalue problemWe consider the nonlinear eigenvalue problem

(iωM +A+ e−iωτB)u = 0

for real matrices M,A,B ∈ Rn×n with invertible M . Soughtare triples (ω, τ, u) consisting of a complex eigenvector u ∈ Cnand two real eigenvalues ω and τ .Problems of this type appear e.g. in the search for critical delaysof linear time invariant delay differential equations (LTI-DDEs)

Mx(t) = −Ax(t)−Bx(t− τ).

In [Meerbergen, Schroder, Voss, 2010, submitted] this problemis discussed for complex M,A,B. Like there we are consideringa Jacobi-Davidson-like projection method. The main differencesin the real case are that a) the search space is kept real and b) aspecialized method for the small projected problem is used.Numerical experiments show the effectiveness of the method.Christian SchroderDept. of MathematicsTU [email protected]

MS 59. Structured matrix computations - Part I of IITalk 1. Factorization ofH2-matricesHierarchical matrices (H-matrices) have been shown to be veryuseful tools for a variety of applications, e.g., the construction ofefficient and robust preconditioners for the solution of ellipticpartial differential equations and integral equations. MostH-matrix algorithms are based on a recursive algorithm thatapproximates the product of twoH-matrices by a newH-matrix,since this fundamental algorithm can be combined with simplerecursive techniques to approximate the inverse or the LRfactorization.CombiningH-matrices with multilevel techniques leads toH2-matrices that can reach significantly higher compressionrates, particularly for very large matrices. Although mostalgorithms can take advantage of the rich multilevel structure tosignificantly improve efficiency, the construction ofH2-matricesposes a challenge since the connections between a large numberof blocks have to be taken into account at each step.The talk presents a new approach to the latter task: by adding asmall amount of book-keeping information to the usual

2012 SIAM Conference on Applied Linear Algebra 52

H2-matrix structure, it is possible to develop an algorithm thatcomputes a low-rank update of a submatrix G|t×s in anH2-matrix in O(k2(#t+ #s)) operations, where k is the localrank.With this new algorithm at our disposal, several importanthigher-level operations become very simple: Low-rankapproximations of submatrices, e.g., computed by popular crossapproximation schemes, can easily be merged into anH2-matrix, leading to an algorithm of complexity O(nk2 logn).The multiplication of twoH2-matrices can be implemented as asequence of low-rank updates, the resulting algorithm also has acomplexity of O(nk2 logn). Using the matrix multiplication,we can also construct the inverse and the LR factorization inO(nk2 logn) operations.Similar toH-matrix techniques, the new algorithm is fullyadaptive, e.g., it can reach any given accuracy, and it is able toeasily handle matrices resulting from the discretization of two-or three-dimensional geometries. Its main advantages overH-matrix algorithms are the significantly reduced storagerequirements and the higher efficiency for large matrices.Steffen BormInstitut fur InformatikChristian-Albrechts-Universitat zu Kiel24118 Kiel, [email protected]

Talk 2. The polynomial root finding problems andquasiseparable representations of unitary matricesThe effective tool to compute all the roots of a polynomial is todetermine the eigenvalues of the corresponding companionmatrix using the QR iteration method. The companion matrixbelongs to the class of upper Hessenberg matrices which arerank one perturbations of unitary matrices. This class is invariantunder QR iterations. Moreover it turns out that for every matrixin this class the corresponding unitary matrix has quasiseparablestructure. This structure may be used to develop fast algorithmsto compute eigenvalues of companion matrices. We discussimplicit fast QR eigenvalue algorithms solving this problem.The obtained algorithm is of complexity O(N) in contrast toO(N2) for non-structured methods.Yuli EidelmanSchool of Mathematical SciencesTel-Aviv UniversityRamat-Aviv 69978, [email protected]

Talk 3. A fast direct solver for structured matrices arisingfrom non-oscillatory integral equationsWe present a fast direct solver for structured dense matricesarising from non-oscillatory integral equations. The solver isbased on (1) multilevel matrix compression techniques thatexploit a complex hierarchical low-rank block structure, and (2)a sparse matrix embedding that allows fast and robust matrixfactorization and inverse application. For boundary integralequations in 2D, the solver has optimal O(N) complexity,where N is the system size; in 3D, it incurs an O(N3/2)precomputation cost, followed by O(N logN) solves.Numerical experiments suggest the utility of our method as botha direct solver and a preconditioner for complex problems.Kenneth L. HoCourant Institute of Mathematical SciencesNew York [email protected]

Leslie GreengardCourant Institute of Mathematical SciencesNew York [email protected]

Talk 4. Multivariate orthogonal polynomials and inverseeigenvalue problemsIt is well known that the computation of the recurrencecoefficients of orthogonal polynomials with respect to a discreteinner product is related to an inverse eigenvalue problem. In thistalk we present an algorithm to compute these recurrencecoefficients for multivariate orthogonal polynomials. Thisalgorithm generalizes previous results for the bivariate case anduses Givens transformations to solve the related inverseeigenvalue problem. We give a number of numerical examplesusing different configurations of points that define the discreteinner product.Matthias HumetDept. of Computer ScienceUniversity of [email protected]

Marc Van BarelDept. of Computer ScienceUniversity of [email protected]

MS 60. Numerical linear algebra libraries for high endcomputing - Part II of IITalk 1. Thick-restart Lanczos methods forsymmetric-indefinite generalized eigenproblems in SLEPcIn this talk we present results on a Lanczos method forgeneralized eigenvalue problems Ax = λBx where both A andB are symmetric matrices but the pair (A,B) is not definite. Inthis case, eigenvalues are not guaranteed to be real and also aneigenvalue can be defective. The standard B-Lanczos cannot beused, but still the symmetry of the matrices can be exploited tosome extent by means of the pseudo-Lanczos process. We showresults of a thick-restart variant implemented in the SLEPclibrary, and compare it with other solvers that totally ignoresymmetry.Jose E. RomanDept. de Sistemes Informatics i ComputacioUniversitat Politecnica de Valencia, [email protected]

Carmen CamposDept. de Sistemes Informatics i ComputacioUniversitat Politecnica de Valencia, [email protected]

Talk 2. Parametric approach to smart-tuning and auto-tuningof the DOE ACTS collectionThe Advanced CompuTational Software (ACTS) Collection is aset of computational tools and libraries developed primarily atDOE laboratories. Here we look at deriving parameters toautomatically identify, at run-time, the most suitable auto-tunedkernels to load with a given application. Additionally, theseparameters can be used in the context of ”Smart-Tuning” toselect the best algorithmic functionality and arguments to alibrary’s API.L. A. (Tony) DrummondComputational Research Division, Lawrence Berkeley NationalLaboratory, [email protected]

Osni Marques

53 2012 SIAM Conference on Applied Linear Algebra

Computational Research Division, Lawrence Berkeley NationalLaboratory, [email protected]

Talk 3. Trilinos: foundational libraries that enablenext-generation computingWith the availability and diversity of powerful computationalresources, including multi-core CPU and GPU technology, thereis significant interest in numerical software libraries that allow adeveloper to optimize the trade-off between effort and impact. Inthis talk we will discuss the current and ongoing efforts bywhich Trilinos is providing enabling technologies for thedevelopment of academic and industrial software targeted atnext-generation architectures. We will cover a wide variety ofTrilinos packages, from the foundational libraries for numericallinear algebra to the solver libraries that can leverage thesefundamental capabilities to develop next-generation algorithms.Heidi ThornquistElectrical Systems Modeling DepartmentSandia National [email protected]

Talk 4. Rethinking distributed dense linear algebraIt is a commonly held misconception that matrices must bedistributed by blocks in order to translate the local computationof classical dense matrix operations into level 3 BLASoperations. In this talk, the performance and programmabilityimplications of element-wise matrix distributions are discussedin the context of a recently introduced implementation,Elemental. In order to be able to effectively convey codesamples, both the FLAME methodology and its extension todistributed-memory computation will be briefly introduced.Jack PoulsonInstitute for Computational Engineering and SciencesThe University of Texas at [email protected]

MS 61. Efficient preconditioners for real worldapplications - Part II of IITalk 1. Relaxed mixed constraint preconditioners forill-conditioned symmetric saddle point linear systemsWe develop efficient preconditioners for generalized saddle

point linear system Ax = b, where A =

[A B>

B −C

]and

A > 0, C ≥ 0 and B a full-rank rectangular matrix. Given twopreconditioners for A (PA and PA) and a preconditioner (PS)for S = BPA

−1B> + C, the Relaxed Mixed Constraint

Preconditioners (RMCP) is denoted byM−1(ω) where

M(ω) =

[I 0

BP−1A I

] [PA 00 −ωPS

] [I P−1

A B>

0 I

].

Eigenanalysis ofM−1(ω)A shows that the optimal ω is relatedto the (cheaply estimated) spectral radius of P−1

A A and P−1S S.

Results regarding large linear systems arising fromdiscretizations of geomechanical problems as well as fluid flowin porous media, show that proper choice of ω improvesconsiderably the MCP performance.Luca BergamaschiDipartimento di Ingegneria Civile Edile AmbientaleUniversity of Padova, [email protected]

Angeles Martınez

Dipartimento di MatematicaUniversity of Padova, [email protected]

Talk 2. Chebychev acceleration of iterative refinementGaussian elimination with partial pivoting followed by iterativerefinement can compute approximate solutions of linear systemsof equations that are backward stable. In some situations, thenumber of refinement steps can be large and their costprohibitive. Limiting the number of steps is particularlyimportant on multicore architectures where the solve phase of asparse direct solver can represent a bottleneck.We propose variants of the Chebyshev algorithm that can beused to accelerate the refinement procedure without loss ofnumerical stability. Numerical experiments on sparse problemsfrom practical applications corroborate the theory and illustratethe potential savings offered by Chebyshev acceleration.Jennifer ScottComputational Science & Engineering DepartmentScience and Technology Faculties Council, [email protected]

Mario ArioliComputational Science & Engineering DepartmentScience and Technology Faculties Council, [email protected]

Talk 3. Parallel deflated GMRES with the Newton basisThe GMRES iterative method is widely used as a Krylovsubspace accelerator for solving sparse linear systems when thecoefficient matrix is nonsymmetric. The Newton basisimplementation has been proposed for distributed memorycomputers as an alternative of the Arnoldi-based approach toavoid low-grained commmunications. The aim of our work hereis to introduce a modification based on deflation techniques.This approach builds an augmented subspace in an adaptive wayto accelerate the convergence of the restarted formulation. In ournumerical experiments, we show the benefits of using thisimplementation in the PETSc package with Schwarzpreconditioners for solving large CFD linear systems.Desire Nuentsa WakamINRIA, Bordeaux, Francedesire.nuentsa [email protected]

Jocelyne ErhelINRIA, Rennes, [email protected]

Talk 4. Rank-k updates of incomplete Sherman-MorrisonpreconditionersLet B = A+ PQT be a large and sparse matrix where A is anonsingular matrix and PQT is a rank−k matrix. In this workwe are interested in solving the updated linear systemBx = b bypreconditioned iterations. We study the problem of updating analready existing preconditioner M for the matrix A. In particularwe consider how to update the incomplete LU factorizationcomputed by the BIF algorithm (SIAM J. Sci. Comput. Vol.30(5), pp. 2302-2318, (2008)). The results of the numericalexperiments with different types of problems will be presented.Jose MarınInstituto de Matematica MultidisciplinarUniversidad Politecnica de ValenciaCamino de Vera s/n, 46022 Valencia, [email protected]

Juana CerdanInstituto de Matematica MultidisciplinarUniversidad Politecnica de Valencia

2012 SIAM Conference on Applied Linear Algebra 54

Camino de Vera s/n, 46022 Valencia, [email protected]

Jose MasInstituto de Matematica MultidisciplinarUniversidad Politecnica de ValenciaCamino de Vera s/n, 46022 Valencia, [email protected]

MS 62. Solving ill-posed systems via signal-processingtechniques - Part II of IITalk 1. Effects of prox parameter selection strategies in exactand inexact first-order methods for compressed sensing andother composite optimization problemsWe will discuss theoretical and practical implications of variousstrategies for choosing the prox parameter in prox gradientmethods and related alternating direction methods. We will showextension of existing convergence rates for both accelerated andclassical first-order methods. Practical comparison based on atesting environment for L1 optimization will be presented.Katya ScheinbergDepartment of Industrial and Systems EngineeringLehigh [email protected]

Donald GoldfarbIEOR DepartmentColumbia [email protected]

Shiqian MaInstitute for Mathematics and its ApplicationsUniversity of [email protected]

Talk 2. An adaptive inverse scale space method forcompressed sensingIn this talk a novel adaptive approach for solving`1-minimization problems as frequently arising in compressedsensing is introduced, which is based on the recently introducedinverse scale space method. The scheme allows to efficientlycompute minimizers by solving a sequence of low-dimensionalnonnegative least-squares problems. Moreover, extensivecomparisons between the proposed method and the relatedorthogonal matching pursuit algorithm are presented.Martin BenningInstitute for Computational and Applied MathematicsUniversity of [email protected]

Michael MollerInstitute for Computational and Applied MathematicsUniversity of [email protected]

Pia HeinsInstitute for Computational and Applied MathematicsUniversity of [email protected]

Talk 3. CGSO for convex problems with polyhedralconstraintsConjugate Gradient with Subspace Optimization (CGSO) is avariant of conjugate gradient algorithm that achieves the optimalcomplexity bound of Nemirovski-Yudin’s algorithm for the classof strongly convex functions. In this talk we are extendingCGSO to constrained problems in which we are minimizing astrictly convex function over a convex polyhedron. We discussthe theoretical properties of CGSO for this class of problems aswell as its practical performance.

Sahar KarimiDepartment of Combinatorics and OptimizationUniversity of [email protected]

Stephen VavasisDepartment of Combinatorics and OptimizationUniversity of [email protected]

Talk 4. Phase-retrieval using explicit low-rank matrixfactorizationRecently, Candes et al. proposed a novel methodology for phaseretrieval from magnitude information by formulating it as amatrix-completion problem. In this work we develop analgorithm aimed at solving large-scale instances of this problem.We take advantage of the fact that the desired solution is of rankone and use low-rank matrix factorization techniques to attainconsiderable speed-up over existing approaches. We considerphase recovery in both the noisy and noiseless setting and studyhow various design choices affect the performance andreliability of the algorithm.Ewout van den BergDepartment of MathematicsStanford [email protected]

Emmanuel CandesDepartments of Mathematics and StatisticsStanford [email protected]

MS 63. Max-algebra - Part II of IITalk 1. Three-dimensional convex polyhedra tropicallyspanned by four pointsIn this talk we show how a 3–dimensional convex polyhedronTA is obtained from a 4× 4 normal tropically idempotentinteger matrix A (i.e., A = (aij), aij ∈ Z≤0, aii = 0,AA = A, with ⊕ = max, = +).Which polyhedra arise this way? TA has 20 vertices and 12facets, at most. By Euler’s formula, the f–vector is (20, 30, 12),at most. Facets have 6 vertices, at most.We show that a polyhedron TA combinatorially equivalent to theregular dodecahedron does not occur, for any A, i.e., thepolygon–vector of TA cannot be (0, 0, 12, 0). Thepolygon–vector of TA cannot be (0, 1, 10, 1), either, but we haveexamples where the polygon–vector of TA is (0, 2, 8, 2). Weprovide families of matrices with TA having polygon–vector(0, 3, 6, 3). Finally, certain families of circulant matrices yieldTA with polygon–vector (0, 4, 4, 4) i.e., the facets of TA are 4quadrilaterals, 4 pentagons and 4 hexagons. Previous work hasbeen done by Joswig and Kulas, Develin and Sturmfels.Marıa Jesus de la PuenteFacultad de MatematicasUniversidad Complutense (UCM)[email protected]

Adrian JimenezFacultad de MatematicasUniversidad Complutense (UCM)[email protected]

Talk 2. Algorithmic problems in tropical convexityWe present recent advances in tropical computational geometry,including the fundamental problem of computing the vertices ofa tropical polyhedron described as intersection of half-spaces, orinversely. We also discuss the connection of these problems with

55 2012 SIAM Conference on Applied Linear Algebra

hypergraph transversals, directed hypergraphs, and mean payoffgames. We finally point out applications of tropical convexity toother fields in computer science, such as formal verification.Xavier AllamigeonINRIA Saclay – Ile-de-France and CMAPEcole Polytechnique, [email protected]

Stephane GaubertINRIA Saclay – Ile-de-France and CMAPEcole Polytechnique, [email protected]

Eric GoubaultCEA Saclay Nano-INNOVInstitut Carnot CEA LIST, DILS/MeASI, [email protected]

Ricardo D. KatzCONICET. Instituto de Matematica “Beppo Levi”Universidad Nacional de Rosario, [email protected]

Talk 3. On the weak robustness of interval fuzzy matricesA fuzzy matrix A (i.e. matrix in a (max, min)-algebra) is calledweakly robust if every orbit sequence of A with starting vector xcontains no eigenvectors, unless x itself is an eigenvector of A.Weak robustness is extended to interval fuzzy matrices and theirproperties are studied. A characterization of weakly robustinterval fuzzy matrices is described and a quadratic algorithm forchecking the weak robustness of a given interval fuzzy matrix ispresented.Jan PlavkaDepartment of Mathematics and Theoretical InformaticsTechnical University in [email protected]

Martin GavalecDepartment of Information TechnologiesUniversity of Hradec [email protected]

Talk 4. Weakly stable matricesGiven a square matrix A and a vector x the sequenceAk ⊗ x

∞k=0

in the max-plus algebra is called the orbit of Awith starting vector x. For some matrices the orbit never reachesan eigenspace unless it starts in one; such matrices are calledweakly stable. We will characterise weakly stable matrices bothin the reducible and irreducible case. It turns out that irreducibleweakly stable matrices are exactly matrices whose critical graphis a Hamiltonian cycle in the associated graph. This talk is basedon joint work with S. Sergeev and H. Schneider.Peter ButkovicSchool of MathematicsUniversity of [email protected]

MS 64. Eigenvalue perturbations and pseudospectra -Part II of IITalk 1. Ritz value localization for non-Hermitian matricesThe Ritz values of Hermitian matrices have long been wellunderstood, thanks to the Cauchy Interlacing Theorem. Recentprogress has begun to uncover similar properties for Ritz valuesof non-Hermitian matrices. For example, the “inverse field ofvalues problem” asks whether a set of k points in the field ofvalues can be Ritz values from a k dimensional subspace. Wesurvey results on this problem, describe how majorization canlead to Ritz value containment regions, and provide a detailedanalysis for a Jordan block.

Mark EmbreeDepartment of Computational and Applied MathematicsRice [email protected]

Russell CardenDepartment of Computational and Applied MathematicsRice [email protected]

Talk 2. Optimization of eigenvalues of Hermitian matrixfunctionsThis work concerns a Hermitian matrix function depending onits parameters analytically. We introduce a numerical algorithmfor the global optimization of a specified eigenvalue of such aHermitian matrix function. The algorithm is based onconstructing piece-wise quadratic under-estimators for theeigenvalue function, and finding global minimizers of thesequadratic models. In the multi-dimensional case finding theglobal minimizers of the quadratic models is equivalent tosolving quadratic programs. The algorithm generates sequencesconverging to global optimizers (linearly in practice). Theapplications include the H-infinity norm of a linear system, andthe distance from a matrix to defectiveness.Emre MengiDept. of MathematicsKoc [email protected]

Mustafa KılıcDept. of MathematicsKoc [email protected]

E. Alper YıldırimDepartment of Industrial EngineeringKoc [email protected]

Talk 3. Algorithms for approximating the H∞ normH∞ norm methods are used in control theory to design optimalcontrollers. The controllers are designed so as to minimize theH∞ norm of the n× n closed-loop transfer matrix where n isthe system order. This optimization procedure necessitatesefficient methods for the computation of the H∞ norm itself.Existing methods compute the H∞ norm accurately but the costis multiple singular value decompositions and eigenvaluedecompositions of size n, making them impractical when n islarge. We present a novel method which provides a fastcomputation of the H∞ norm for large and sparse matrices, suchas the ones arising in the control of PDE’s. The method is anested fixed point iteration, where the outer iteration is a Newtonstep and the inner iteration is associated with the problem of thecomputation of the ε-pseudospectral abscissa, i.e. the real part ofa righmost point of the ε-pseudospectrum of a certain linearoperator. The fixed points of the iteration are characterized, locallinear convergence of the algorithm for small enough ε is givenand some applications to the control of PDE’s are discussed.

Michael L. OvertonCourant Institute, New York [email protected]

Mert GurbuzbalabanCourant Institute, New York [email protected]

Nicola GuglielmiUniversity of L’[email protected]

Talk 4. Reduced basis methods for computing pseudospectral

2012 SIAM Conference on Applied Linear Algebra 56

quantitiesReduced basis methods have been developed to efficiently solveparameter-dependent partial differential equations and, after aspatial discretization, the resulting linear systems. They typicallyconsist of two phases. In an offline phase, the linear system issolved for several parameter samples and a low-dimensionalsubspace containing these solutions (approximately) isconstructed. In an online phase, only the compression of thelinear system with respect to this subspace needs to be solved,leading to greatly reduced execution times. The effectivity ofreduced basis methods crucially depends on the regularity of theparameter dependence and the chosen sampling strategy.In this talk, we show how an approach inspired by reduced basismethods can be used to compute pseudospectra and associatedquantities of a matrix A. Instead of solving aparameter-dependent linear system, one needs to consider thecomputation of the singular vector(s) belong to the smallestsingular value(s) of A− zI for all values of the complexparameter z of interest. While the computation of pseudospectraitself is still under development, we show how these techniquecan already be used to speed up a recently proposed algorithmby Gulgielmi and Overton for computing the pseudospectralabscissa.Daniel KressnerMATHICSE-ANCHPEPF [email protected]

MS 65. Numerical linear algebra and optimization inimaging applications - Part II of IITalk 1. A recursion relation for solving L-BFGS systems withdiagonal updatesWe investigate a formula to solve limited-memory BFGSquasi-Newton Hessian systems with full-rank diagonal updates.Under some mild conditions, the system can be solved via arecursion that uses only vector inner products. This approach hasbroad applications in trust region and barrier methods inlarge-scale optimization.Jennifer B. ErwayDept. of MathematicsWake Forest [email protected]

Roummel F. MarciaSchool of Natural SciencesUniversity of California, [email protected]

Talk 2. Wavefront gradients reconstruction using l1 − lpmodelsImages of objects in outer space acquired by ground-basedtelescopes are usually blurred by atmospheric turbulence. Toimprove the quality of these images, the wavefront of the light isutilized to derive the point spread function (PSF). We proposedthe l1 − lp (p = 1, 2) model for reconstructing the wavefrontgradients and hence the wavefront itself. The model can give amore accurate PSF and therefore better restored images.Numerical results are given to illustrate the performance of theproposed models.Raymond H. ChanDept. of MathematicsThe Chinese University of Hong [email protected]

Xiaoming YuanDept. of MathematicsHong Kong Baptist [email protected]

Wenxing ZhangDept. of MathematicsNanjing [email protected]

Talk 3. Edge-preserving image enhancement via blinddeconvolution and upsampling operatorsWe consider the super-resolution model introduced by S.J. Osherand A. Marquina in J. Sci. Comput., 2008 volume 37, 367382,consisting of the solution of a total-variation based variationalproblem that solves a linear degradation model involving aconvolution operator and a down- sampling operator. In thisresearch work we explore different edge preservingup/down-sampling operators with different orders of spatialaccuracy and estimated convolution operators to removeunknown blur from degraded low resolved images to solve thisgenuine inverse problem. Some numerical examples areprovided to show the features of the proposed algorithm.Antonio MarquinaDepartment of Applied Mathematics,University of Valencia, [email protected]

Stanley J. OsherDepartment. of Mathematics,UCLA, [email protected]

Shantanu H. JoshiLONI,UCLA, [email protected]

Talk 4. A new hybrid-optimization method for large-scale,non-negative, full regularizationWe present a new method for solving the full regularizationproblem of computing both the regularization parameter and thecorresponding solution of a linear or nonlinear ill-posedproblem. The method is based on stochastic and fast,gradient-based optimization techniques, and computesnon-negative solutions to `1 or `2 regularization problems. Wedescribe the method and present numerical results for large-scaleimage restoration problems.Marielba RojasDelft Institute of Applied MathematicsDelft University of Technology, The [email protected]

Johana GuerreroDepartment of Computer ScienceUniversity of Carabobo, [email protected]

Marcos RaydanDepartment of Scientific Computing and StatisticsSimon Bolıvar University, [email protected]

MS 66. Parametric eigenvalue problems - Part II of IITalk 1. A subspace optimization technique for the generalizedminimal maximal eigenvalue problem.Consider the Generalized Minimal Maximal Eigenvalue

57 2012 SIAM Conference on Applied Linear Algebra

Problem:

λ∗max := minx

maxλ,~v

λ

st :

K(x)~v = λM(x)~v,

x ∈ H ⊂ Rp,~v ∈ Rn

where the matrices K(x) and M(x) are affine matrix functions,creating a symmetric positive definite pencil (M(x),M(x)) onthe hypercubeH. It is a quasi-convex, non-smooth optimizationproblem.We present a subspace iteration method for computing theminimal maximal eigenvalue of a large scale eigenvalueproblem. The idea is based on Kelley’s Classical cutting planemethod for convex problems. In each iteration λ1 and acorresponding eigenvector ~v are computed for a specific x. Theeigenvector ~v is then used to expand the subspace and the nextiterate x is determined from the minimal maximum eigenvalueof the small scale projected problem. Convergence theory of themethod relies on the fact that the maximal Ritz value is a supportto the maximal eigenvalue.Jeroen De VliegerDept. of Computer [email protected]

Karl Meerbergen (promotor)Dept. of Computer [email protected]

Talk 2. An iterative method for computing pseudospectralabscissa and stability radii for nonlinear eigenvalueproblemsWe consider the following class of nonlinear eigenvalueproblems, (

m∑i=1

Aipi(λ)

)v = 0,

where A1, . . . , Am are given n× n matrices and the functionsp1, . . . , pm are assumed to be entire. This does not only includepolynomial eigenvalue problems but also eigenvalue problemsarising from systems of delay differential equations. Our aim isto compute the ε-pseudospectral abscissa, i.e. the real part of therightmost point in the ε-pseudospectrum, which is the complexset obtained by joining all solutions of the eigenvalue problemunder perturbations δAimi=1, of norm at most ε, of the matricesAimi=1.In analogy to the linear eigenvalue problem we prove that it issufficient to restrict the analysis to rank-1 perturbations of theform δAi = βiuv

∗ where u ∈ Cn and v ∈ Cn with βi ∈ C forall i. Using this main - and unexpected - result we present newiterative algorithms which only require the computation of thespectral abscissa of a sequence of problems obtained by addingrank one updates to the matrices Ai. These provide lowerbounds to the pseudspectral abscissa and in most cases convergeto it. A detailed analysis of the convergence of the algorithms ismade.The methods available for the standard eigenvalue problem inthe literature provide a robust and reliable computation but at thecost of full eigenvalue decompositions of order 2n and singularvalue decompositions, making them unfeasible for largesystems. Moreover, these methods cannot be generalized to

nonlinear eigenvalue problems, as we shall explain. Therefore,the presented method is the first generally applicable method fornonlinear problems. In order to be applied it simply requires aprocedure to compute the rightmost eigenvalue and thecorresponding left and right eigenvectors. In addition, if thematrices Ai are large and sparse then the computation of therightmost eigenvalue can for many classes of nonlineareigenvalue problems be performed in an efficient way byiterative algorithms which only rely on matrix vectormultiplication and on solving systems of linear equations, wherethe structure of the matrices (original sparse matrices plus rankone updates) can be exploited. This feature, as well otherproperties of the presented numerical methods, are illustrated bymeans of the delay and polynomial eigenvalue problem.Wim MichielsDepartment of Computer ScienceKU [email protected]

Nicola GuglielmiDipartimento di Matematica Pura e ApplicataUniversita dell’[email protected]

Talk 3. Statistical pseudospectrum and eigenvalue robustnessto rank-one disturbanceIn this talk, we present a new statistical measure of therobustness (or sensitivity) of linear dynamic systems to rank-onerandom disturbances. The sensitivity assessment is a statisticalpseudospectrum: given the probability distribution of therandom disturbance magnitude, it measures the expectedfrequency region where the system eigenvalues are located . Wediscuss the properties of the robustness and sensitivity measure.We notably stress the existence of an invariant of the measurethat consequently shows that under certain conditions, the rate ofincrease of a rank-one pseudospectrum area is constant as afunction of the disturbance magnitude.Christophe LecomteInstitute of Sound and Vibration Research (ISVR) andSouthampton Statistical Sciences Research Institute (S3RI)University of [email protected]

Maryam Ghandchi TehraniInstitute of Sound and Vibration Research (ISVR) andUniversity of SouthamptonUniversity Road, [email protected]

MS 67. Structured matrix computations - Part II of IITalk 1. Randomized numerical matrix computations withapplicationsIt is long and well known that random matrices tend to be wellconditioned, but exploitation of this phenomenon in numericalmatrix computations is more recent and there are still variousnew directions to investigate. Some of them will be covered inthis talk. This includes pivoting-free but safe Gaussianelimination, randomized preconditioning of linear systems ofequations, computation of a basis for the null space of a singularmatrix, approximation by low-rank matrices, and applications topolynomial root-finding.Victor PanDepartment of Mathematics and Computer ScienceLehman College - City University of New YorkBronx, New York 10468

2012 SIAM Conference on Applied Linear Algebra 58

[email protected]

Talk 2. Massively parallel structured direct solver for theequations describing time-harmonic seismic wavesWe consider the discretization and approximate solutions ofequations describing time-harmonic seismic waves in 3D. Wediscuss scalar qP polarized waves, and multi-component elasticwaves in inhomogeneous anisotropic media. The anisotropycomprises general (tilted) TI and orthorhombic symmetries. Weare concerned with solving these equations for a large number ofdifferent sources on a large domain. We consider variable orderfinite difference schemes, to accommodate anisotropy on the onehand and allow higher order accuracy – to control sampling ratesfor relatively high frequencies – on the other hand.We make use of a nested dissection based domaindecomposition, with separators of variable thickness, andintroduce an approximate direct (multifrontal) solver bydeveloping a parallel Hierarchically SemiSeparable (HSS)matrix compression, factorization, and solution approach. Inparticular, we present elements of the following new parallelalgorithms and their scalability: The parallel construction of anHSS representation or approximation for a general dense matrix,the parallel ULV factorization of such a matrix, and the parallelsolution with multiple right-hand sides. The parallel HSSconstruction consists of three phases: Parallel rank revealing QR(RRQR) factorization based on a Modified Gram-Schmidt(MGS) method with column pivoting, parallel row compression,and parallel column compression. The parallel HSS factorizationinvolves the use of two children’s contexts for a given parentcontext. The communication patterns are composed ofintra-context and inter-context ones. Similar strategies areapplied to the HSS solution.We present various examples illustrating the performance of ouralgorithm as well as applications in so-called full waveforminversion.Maarten V. de HoopDepartment of Mathematics and Department of Earth and AtmosphericSciencesPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

Shen WangDepartment of Mathematics and Department of Earth and AtmosphericSciencesPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

Jianlin XiaDepartment of MathematicsPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

Xiaoye S. LiComputational Research DivisionLawrence Berkeley National LaboratoryBerkeley, CA [email protected]

Talk 3. On the conditioning of incomplete Choleskyfactorizations with orthogonal droppingWe consider incomplete Cholesky factorizations based onorthogonal dropping for the iterative solution of symmetricpositive definite linear systems. These methods becomeincreasingly popular tools for computing an approximate

factorization of large dense matrices, including update matricesand Schur complements that arise in spares solvers. For thesystem preconditioned with these incomplete factorizations wepresent an upper bound on the condition number which onlydepends on the accuracy of the individual approximation(dropping) steps. The analysis is illustrated with some existingfactorization algorithms in the context of discretized ellipticpartial differential equations.Artem NapovComputational Research DivisionLawrence Berkeley National LaboratoryBerkeley, CA 94720, [email protected]

Talk 4. Randomized direct solversWe propose some new structured direct solvers for large linearsystems, using randomization and other techniques. Our workinvolves new flexible methods to exploit structures in largematrix computations. Our randomized structured techniquesprovide both higher efficiency and better applicability than someexisting structured methods. New efficient ways are proposed toconveniently perform various complex operations which aredifficult in standard rank-structured solvers. Extension of thetechniques to least squares problems and eigenvalue problemswill also be shown.We also study the following issues:

1. Develop matrix-free structured solvers.2. Update a structured factorization when few matrix entries

change.3. Relaxed rank requirements in structured solvers. We

show the feasibility of our methods for solving variousdifficult problems, especially high dimensional ones.

4. Develop effective preconditioners for problems withoutsignificant rank structures. We analyze the criterion forcompressing off-diagonal blocks so as to achieve nearlyoptimal effectiveness and efficiency in our preconditioner.

Jianlin XiaDepartment of MathematicsPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

Maarten V. de HoopDepartment of Mathematics and Department of Earth and AtmosphericSciencesPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

Xiaoye S. LiComputational Research DivisionLawrence Berkeley National LaboratoryBerkeley, CA [email protected]

Shen WangDepartment of Mathematics and Department of Earth and AtmosphericSciencesPurdue UniversityWest Lafayette, Indiana 47907, [email protected]

MS 68. Linear algebra for structured eigenvaluecomputations arising from (matrix) polynomialsTalk 1. A QR algorithm with generator compression forstructured eigenvalue computation

59 2012 SIAM Conference on Applied Linear Algebra

In this talk we present a new structured implicit QR algorithmfor fast computation of matrix eigenvalues. The algorithm,which relies on the properties of quasiseparable structure,applies to unitary-plus-rank-one Hessenberg matrices and, inparticular, to Frobenius companion matrices. It computes theeigenvalues of an n× n matrix in O(n2) operations with O(n)memory. The introduction of a generator compression stepallows to reduce complexity – without spoiling accuracy – withrespect to a previous fast eigensolver for companion matrices([Bini, Boito, Eidelman, Gemignani, Gohberg, LAA 2010]).Numerical results will be presented for the single- and doubleshift strategies.Paola BoitoXLIM-DMIUniversity of [email protected]

Yuli EidelmanSchool of Mathematical SciencesRaymond and Beverly Sackler Faculty of Exact SciencesTel-Aviv [email protected]

Luca GemignaniDept. of MathematicsUniversity of [email protected]

Israel GohbergSchool of Mathematical SciencesRaymond and Beverly Sackler Faculty of Exact SciencesTel-Aviv University

Talk 2. Quadratic realizability for structured matrixpolynomialsWhich lists of elementary divisors L can be realized by somequadratic matrix polynomial Q(λ)? For regular Q(λ) over thefield C, this problem has recently been solved by severalresearchers. Indeed, if L can be realized at all by some regular Qover C, then it can always be realized by some upper triangularQ; several independent proofs are now known. This talk focuseson the analogous question for structured matrix polynomials. IfS is a class of structured polynomials, which elementary divisorlists L can be realized by some quadratic Q(λ) in S? We surveycurrent progress on this problem for various structure classes S,including palindromic, alternating, and Hermitian matrixpolynomials. As time permits we will also consider the quadraticrealizability of additional features, such as sign characteristicsand minimal indices, for these same structure classes.D. Steven MackeyDept. of MathematicsWestern Michigan [email protected]

Fernando De TeranDepto. de MatematicasUniversidad Carlos III de [email protected]

Froilan M. DopicoICMAT/Depto. de MatematicasUniversidad Carlos III de [email protected]

Francoise TisseurSchool of MathematicsThe University of [email protected]

Talk 3. Fast computation of zeros of a polynomialThe usual method for computing the zeros of a polynomial is to

form the companion matrix and compute its eigenvalues. Inrecent years several methods that do this computation in O(n2)time with O(n) memory by exploiting the structure of thecompanion matrix have been proposed. We propose a newmethod of this type that makes use of Fiedler’s factorization of acompanion matrix into a product of n− 1 essentially 2× 2matrices. Our method is a non-unitary variant of Francis’simplicitly-shifted QR algorithm that preserves this structure. Asa consequence the memory requirement is a very modest 4n, andthe flop count is O(n) per iteration (and O(n2) overall). We willpresent numerical results and compare our method with othermethods.David S. WatkinsDepartment of MathematicsWashington State [email protected]

Jared L. AurentzDepartment of MathematicsWashington State [email protected]

Raf VandebrilDepartment of Computer ScienceK. U. [email protected]

Talk 4. Eigenvector recovery of linearizations and thecondition number of eigenvalues of matrix polynomialsThe standard formula for the condition number of a simpleeigenvalue of a matrix polynomial involves the left and righteigenvectors associated with the eigenvalue [F. TISSEUR,Backward error and condition of polynomial eigenvalueproblems, Linear Algebra Appl., 309 (2000) 339–361]. Theusual way to solve polynomial eigenvalue problems is by usinglinearizations. In the past few years, different families oflinearizations have been introduced. In order to compare thecondition number for the eigenvalues of these linearizations, weneed formulas for the associated eigenvectors. We present in thistalk formulas for the eigenvectors of several families ofFiedler-lyke linearizations. These formulas are introduced usingthe notion of eigencolumn, which allows us to relate theeigenvectors of the linearizations with the eigenvectors of thepolynomial. This fact may allow also to compare the conditionnumber of the eigenvalue in the polynomial with the conditionnumber of the eigenvalue in the linearizations. Moreover, the useof eigencolumns allows us to express similar formulas forminimal bases of singular polynomials.Fernando De TeranDepto. de MatematicasUniversidad Carlos III de [email protected]

Maria I. BuenoDept. of MathematicsUniversity of California at Santa [email protected]

Froilan M. DopicoICMAT/Depto. de MatematicasUniversidad Carlos III de [email protected]

D. Steven MackeyDept. of MathematicsWestern Michigan [email protected]

MS 69. Advances in sparse matrix factorization

2012 SIAM Conference on Applied Linear Algebra 60

Talk 1. A sparse inertia-revealing factorizationWe show how to apply Wilkinson’s inertia-revealingfactorization to sparse matrices in a way that preserves sparsity.No other inertia-revealing factorization is guaranteed to preservesparsity. The input matrix A is factored row by row, therebyproducing the triangular factors of all its leading blocks. theinertia is derived from the number of sign changes in thesequence of determinants (revealed by the factors) of theseblocks. We show that the fill in the triangular factors is boundedby the fill in the QR factorization of A. Therefore,symmetrically pre-permuting A to a doubly-bordered formguarantees sparsity in our algorithm.Alex DruinskySchool of Computer ScienceTel-Aviv [email protected]

Sivan ToledoSchool of Computer ScienceTel-Aviv [email protected]

Talk 2. Multifrontal factorization on heterogeneous multicoresystemsWhen solving the sparse linear systems that arise in MCAEapplications, the multifrontal method is particularly attractive asit transforms the sparse matrix factorization into an eliminationtree of dense matrix factorizations. The vast majority of thefloating point operations can be performed with calls to highlytuned BLAS3 routines, and near peak throughput is expected.Such computations are performed today on clusters of multicoremicroprocessors, often accelerated by by graphics processingunits (GPUs). This talk discusses how concurrency in themultifrontal computation is processed with message passing(MPI), shared memory (OpenMP), and GPU accelerators(CUDA), exploiting the unique strengths of each.Bob LucasInformation Sciences InstituteUniversity of Southern [email protected]

Roger GrimesLivermore Software Technology [email protected]

John TranInformation Sciences InstituteUniversity of Southern [email protected]

Gene WagenbrethInformation Sciences InstituteUniversity of Southern [email protected]

Talk 3. Towards an optimal parallel approximate sparsefactorization algorithm using hierarchically semi-separablestructuresHierarchically semiseparable (HSS) matrix algorithms areemerging techniques in constructing the superfast direct solversfor both dense and sparse linear systems. We present a set ofnovel parallel algorithms for the key HSS operations that areneeded for solving large linear systems, including the parallelrank-revealing QR factorization, the HSS constructions withhierarchical compression, the ULV HSS factorization, and theHSS solutions. We have appplied our new parallelHSS-embedded multifrontal solver to the anisotropic Helmholtzequations for seismic imaging, and were able to solve a linear

system with 6.4 billion unknowns using 4096 processors, inabout 20 minutes.Xiaoye S. LiLawrence Berkeley National [email protected]

Shen WangPurdue [email protected]

Jianlin XiaPurdue [email protected]

Maarten V. de HoopPurdue [email protected]

Talk 4. Improving multifrontal methods by means oflow-rank approximation techniquesBy a careful exploitation of the low-rank property of discretizedelliptic PDEs, a substantial reduction of the flops and memoryconsumption can be achieved for many linear algebra operations.In this talk, we present how low-rank approximations can beused to significantly improve a sparse multifrontal solver. Weintroduce a blocked, low-rank storage format for compressingfrontal matrices and compare it to the HSS storage format.Finally, we present experimental results showing the reductionof flops and memory footprint achieved on the solution of largescale matrices from applications such as the acoustic waveequation and thermo-mechanics.Clement WeisbeckerINPT(ENSEEIHT)-IRITUniversity of [email protected]

Patrick AmestoyINPT(ENSEEIHT)-IRITUniversity of [email protected]

Cleve AshcraftLSTC, [email protected]

Olivier BoiteauEDF R&D, [email protected]

Alfredo [email protected]

Jean-Yves L’ExcellentINRIA-LIP(ENS Lyon)[email protected]

MS 70. Accurate algorithms and applicationsTalk 1. High-precision and accurate algorithms in Physics andMathematicsIn this talk we present a survey of recent applications in Physicsand Mathematics where high level of numeric precision isrequired. Such calculations are facilitated, on one hand, byhigh-precision software packages that include high-levellanguage translation modules to minimize the conversion effort,and on the other hand, by the use of theoretical error bounds andaccurate algorithms when available. These applications includesupernova simulations, planetary orbit calculations, Coulombn-body atomic systems, scattering amplitudes of quarks, gluonsand bosons, nonlinear oscillator theory, experimentalmathematics, evaluation of recurrence relations, numericalintegration of ODEs, computation of periodic orbits, studies ofthe splitting of separatrices, detection of strange non-chaotic

61 2012 SIAM Conference on Applied Linear Algebra

attractors, Ising theory, quantum field theory, and discretedynamical systems. We conclude that high-precision arithmeticfacilities are now an indispensable component of a modernlarge-scale scientific computing environment.Roberto BarrioDept. Matematica Aplicada and IUMA,Universidad de Zaragoza, [email protected]

David H. BaileyLawrence Berkeley National Laboratory,Berkeley, [email protected]

Jonathan M. BorweinCARMA, University of Newcastle,Callaghan, [email protected]

Sergio SerranoDept. Matematica Aplicada and IUMA,Universidad de Zaragoza, [email protected]

Talk 2. Accurate evaluation of 1D and 2D polynomials inBernstein formIn this talk we present some fast and accurate algorithms toevaluate the 1D and 2D polynomials expressed in the Bernsteinform in CAGD. As a well-known and stable algorithm, DeCasteljau algorithm may be still less accurate than expectedowing to cancelation in some circumstances. Our compensatedalgorithms, applying error-free transformation, can yield a resultas accurate as if computed by the De Casteljau algorithm intwice working precision. Numerical tests illustrate that ouralgorithms can run significantly faster than the De Casteljaualgorithm using double-double library.Hao JiangPEQUAN team, LIP6,Universite of Pierre et Marie [email protected]

Roberto BarrioDept. Matematica Aplicada and IUMA,Universidad de Zaragoza, [email protected]

Talk 3. Some issues related to double roundingsDouble rounding is a phenomenon that may occur whendifferent floating-point precisions are available on a samesystem, or when performing scaled operations whose final resultis subnormal. Although double rounding is, in general,innocuous, it may change the behavior of some useful smallfloating-point algorithms. We analyze the potential influence ofdouble roundings on the Fast2Sum and 2Sum algorithms, onsome summation algorithms, and Veltkamp’s splitting. We alsoshow how to handle possible double roundings when performingscaled Newton-Raphson division iterations (to avoid possibleunderflow problems).Jean-Michel MullerCNRS, Laboratoire LIPUniversite de Lyon, [email protected]

Erik Martin-DorelEcole Normale Superieure de Lyon, Laboratoire LIPUniversite de Lyon, [email protected]

Guillaume MelquiondINRIA, Proval team, LRIUniversite Paris Sud, [email protected]

Talk 4. Error bounds for floating-point summation and dotproductThe sum and dot product of vectors of floating-point numbersare ubiquitous in numerical calculations. Since four decades theerror is bounded by the classical Wilkinson estimates. However,those contain some nasty denominator covering higher orderterms. In this talk we show that the latter can be omitted. A keyto the (mostly) simple proofs is our ufp-concept denoting the”unit in the first place”. In contrast to the well-known ulp (unitin the last place) it is defined for real numbers and allows sharp,simple and nice error estimates for floating-point operations.The practical relevance of the new estimates is limited; however,they are aesthetically pleasing and confirm that it is true whatone may (hope or) expect.Siegfried M. RumpInstitute for Reliable ComputingHamburg University of TechnologyHamburg, Germany, andFaculty of Science and EngineeringWaseda University, Tokyo, [email protected]

MS 71. Theoretical and applied aspects of graphLaplaciansTalk 1. Potential theory for perturbed Laplacian of finitenetworksGiven a symmetric and irreducible M–matrix M, theoff-diagonal entries of M can be identified with the conductancefunction of a connected network Γ. In particular the matrixobtained by choosing di =

∑nj=1,j 6=i cij , where n is the order

of the network, is nothing but the combinatorial Laplacian of thenetwork. Therefore, any matrix with off-diagonal values givenby −cij can be considered as a perturbed Laplacian of ΓFrom the operator theory point of view, the perturbed Laplaciansare identified with the so-called discrete Schrodinger operatorsof the network Γ. Our main objective is the study of positivesemi–definite Schrodinger operators. In fact, many of ourtechniques and results appear as the discrete counterpart of thestandard treatment of the resolvent of elliptic operators onRiemannian manifolds. Joint work with E. Bendito, A. Carmonaand A.M. EncinasMargarida MitjanaDep. Matemtica Aplicada IUniversitat Politecnica de [email protected]

Enrique BenditoDep. Matematica Aplicada IIIUniversitat Politecnica de [email protected]

Angeles CarmonaDep. Matematica Aplicada IIIUniversitat Politecnica de [email protected]

Andres M. EncinasDep. Matematica Aplicada IIIUniversitat Politecnica de [email protected]

Talk 2. Subclasses of graphs with partial ordering withrespect to the spectral radius of generalized graphLaplacians.Brualdi and Solheid (1986) proposed the following generalproblem, which became a classical problem in spectral graph

2012 SIAM Conference on Applied Linear Algebra 62

theory: “Given a set G of graphs, find an upper bound for thespectral radius in this set and characterize the graphs in whichthe maximal spectral radius is attained.” Now there existsextensive literature that characterizes such extremal graphs forquite a couple of such sets. Moreover, the problem has beengeneralized to the (signless) Laplacian matrix of graphs andthere even exist a few contributions that provide results fornon-linear generalizations of these matrices, such as thep-Laplacian.Many of the proofs for these results apply graph perturbationsthat increase or decrease the spectral radius. In graph classes,like trees, one may eventually arrive at a graph with maximumor minimum spectral radius. As a by-product we get a partialordering of graphs in such classes. This procedure may alsowork for generalized graph Laplacians (symmetric matrices withnon-positive off-diagonal entries) like Dirichlet matrices.In this talk we want to find a general framework for derivingsuch results. We present sets of graphs and generalizedLaplacians where this procedure can be applied.This is a joint work with Turker Bıyıkoglu.Josef LeydoldInstitute for Statistics and MathematicsWU Vienna University of Economics and [email protected]

Turker BıyıkogluDepartment of MathematicsIsık University,[email protected]

Talk 3. Some new results on the signless Laplacian of graphsSince recently the signless Laplacian spectrum has attractedmuch attention in the literature. In this talk we will put focus onsome new results about the signless Laplacian spectrum whichare inspired by the results so far established for the adjacency orthe Laplacian spectrum.Slobodan K. SimicMathematical Institute of Serbian Academy of Science and [email protected]

Talk 4. Graph bisection from the principal normalizedLaplacian eigenvectorGraph bisection is the most often encountered form of graphpartitioning, which asks to divide the nodes of a network intotwo non-overlapping groups such that the number of linksbetween nodes in different groups is minimized. However, graphbisection may also be understood as the search for the largestbipartite subgraph in the complement: if two groups of G shouldhave sizes roughly equal to n/2, and if m1 and m2 are thenumbers of intra-group links and inter-group links, respectively,then the numbers of intra-group and inter-group links in thecomplement GC will be roughly equal to n2/4m1 and n2/4m2.Hence, the request to minimize m2 translates into the request ofmaximizing n2/4m2.Here we may relate to a well-known property of the spectrum ofthe normalized Laplacian matrix L∗ of G, saying that G isbipartite if and only if 2 is the largest eigenvalue of L∗. If G isbipartite, then the signs of the components of the eigenvectorcorresponding to eigenvalue 2 of L∗, yield the bipartition of G.In the more interesting case when G is not bipartite, one maystill expect that the signs of the components of the eigenvectorcorresponding to the largest eigenvalue of L∗ will yield a largebipartite subgraph of G. For example, it yields a perfectclassification of club members in the Zacharys karate club

network. More interestingly, the sizes of the bipartite subgraphsobtained in this way in instances of random networks show highsimilarity to those obtained by applying the old Erdos method,an iterative process which starts with a random partition in twogroups and, as long as there exists a node having more than halfof its neighbors in its own group, it moves such node to the othergroup. This similarity is explored in more detail in this talk.Dragan StevanovicUniversity of Primorska, UP IAM, Koper Solvenia andUniversity of Nis, [email protected]

MS 72. Linear techniques for solving nonlinearequationsTalk 1. A Gauss-Seidel process in iterative methods forsolving nonlinear equationsIn this talk, we present a process named “Gauss-Seidelization”for solving nonlinear equations . It is an iterative process basedon the well-known Gauss-Seidel method to numerically solve asystem of linear equations. Together with some convergenceresults, we show several numerical experiments in order toemphasize how the Gauss-Seidelization process influences onthe dynamical behaviour of an iterative method for solvingnonlinear equations.Jose M. GutierrezDept. Mathematics and Computer SciencesUniversity of La RiojaLogrono, [email protected]

A. MagrenanDept. Mathematics and Computer SciencesUniversity of La RiojaLogrono, [email protected]

J. L. VaronaDept. Mathematics and Computer SciencesUniversity of La RiojaLogrono, [email protected]

Talk 2. A greedy algorithm for the convergence of a fractionalblind deconvolutionIn this talk we present new results on a greedy algorithm to studythe convergence of the Fractional Blind Deconvolution based onfractional powers of the laplacian. Greedy algorithms perform atheoretical background to prove the convergence in a HilbertSpace. We will show the theoretical results and an application tobaroque paintings.Vicente F. CandelaDepartamento de Matematica AplicadaUniversidad de [email protected]

Pantaleon David Romero SanchezDepartamento de Ciencias Fısicas, Matematicas y de la ComputacionUniversidad CEU-Cardenal [email protected]

Talk 3. Overview of iterative methods using a variationalapproachIn this talk, we introduce a general framework for theapproximation of nonlinear equations in Banach spaces. Weadapt a variational perspective recently introduced for theanalysis of differential equations. In this new approach, someclassical iterative methods, including their convergence analysis,can be obtained. The method can be considered as a

63 2012 SIAM Conference on Applied Linear Algebra

generalization of the discrete least squares method, which is astandard approach to the approximate solution of many types ofproblems including overdetermined systems of nonlinearequations.S. BusquierDept. of Applied Mathematics and StatisticsU.P. [email protected]

S. AmatDept. of Applied Mathematics and StatisticsU. P. [email protected]

P. PedregalE.T.S. Ingenieros Industriales.Universidad de Castilla La [email protected]

Talk 4. Iterative methods for ill-conditioned problemsIn this talk we study some features related to ill-conditionednonlinear equations. A strategy to choose iterative methods forsolving these equations is developed. In particular, we analyze avariation of Newton method, the so called perturbed Newtonmethod. Important features to achieve good performance, suchas choice of pivots and parameters provide degrees of freedomthat can be suited for improvement of the method. We illustratethis analysis through examples.Rosa M. Peris

Dept. of Applied MathematicsUniversity of [email protected]

Vicente F. CandelaDept. of Applied MathematicsUniversity of [email protected]

MS 73. Algebraic Riccati equations associated withM-matrices: numerical solution and applicationsTalk 1. Monotone convergence of Newton-like methods forM-matrix algebraic Riccati equationsThe minimal nonnegative solution of an M-matrix algebraicRiccati equation can be obtained by Newton’s method. Here westudy Newton-like methods that have higher-order convergenceand are not much more expensive each iteration, and are thusmore efficient than Newton’s method. For the Riccati equation,these Newton-like methods are actually special cases of theNewton–Shamanskii method. We show that, starting with zeroinitial guess or some other suitable initial guess, the sequencegenerated by the Newton–Shamanskii method convergesmonotonically to the minimal nonnegative solution.Chun-Hua GuoDept. of Mathematics and StatisticsUniversity of [email protected]

Talk 2. Accurate solution of M -matrix algebraic Riccatiequation by ADDA: alternating-directional doublingalgorithmIt is known that that an M -matrix Algebraic Riccati Equation(MARE) XDX −AX −XB + C = 0 has a unique minimalnonnegative solution Φ. In this talk, we will discuss two recentdevelopments:

• a relative perturbation theory that will show that smallrelative perturbations to the entries of A, B, C, and Dintroduce small relative changes to the entries of the

nonnegative solution Φ, unlike the existing perturbationtheory for (general) Algebraic Riccati Equations.

• an efficient Alternating-Directional Doubling Algorithm(ADDA) that can compute such a solution Φ asaccurately as predicted by the relative perturbation theory.

Ren-Cang LiDept. of MathematicsUniversity of Texas at [email protected]

Wei-guo WangSchool of Mathematical SciencesOcean University of [email protected]

Wei-chao WangDept. of MathematicsUniversity of Texas at [email protected]

Shufang XuSchool of Mathematical SciencesPeking [email protected]

Jungong XueSchool of Mathematical ScienceFudan [email protected]

Talk 3. When fluid becomes Brownian: the morphing ofRiccati into quadratic equationsRamaswami (2011) approximates Brownian motion using aMarkov-modulated linear fluid model. This is extended toapproximate Markov-modulated Brownian motion (MMBM); inparticular, the Laplace matrix exponent of a Markov-modulatedlinear fluid model converges to that of an MMBM. When theMMBM is reflected at zero, its stationary distribution is the limitof that of the fluid model also reflected at zero. Proof ofconvergence combines probabilistic arguments and linearalgebra. Starting from the Wiener-Hopf factorization of themodulating Markov chain, expressed as an algebraic Riccatiequation, key matrices in the limiting stationary distribution areshown to be solutions of a new quadratic equation.Giang T. NguyenDepartement d’InformatiqueUniversite Libre de [email protected]

Guy LatoucheDepartement d’InformatiqueUniversite Libre de [email protected]

V. RamaswamiAT&T Labs, [email protected]

Talk 4. Analyzing multi-type queues with general customerimpatience using Riccati equationsWe consider a class of multi-type queueing systems withcustomer impatience, where the (phase-type) service time and(general) patience distribution is type dependent and the types ofconsecutive customers may be correlated.To obtain the per-type waiting time distribution and probabilityof abandonment, we construct a fluid queue with r thresholds,the steady state of which can be expressed via the solution of 2ralgebraic Riccati equations using matrix-analytic methods.Numerical examples indicate that thousands thresholds may berequired to obtain accurate results, indicating the need for fast

2012 SIAM Conference on Applied Linear Algebra 64

and numerically stable algorithms to solve large sets of Riccatiequations.Benny Van HoudtDept. of Mathematics and Computer ScienceUniversity of [email protected]

MS 74. Recent advances in the numerical solution oflarge scale matrix equationsTalk 1. Hierarchical and multigrid methods for matrix andtensor equationsHierarchical and Multigrid methods are among the most efficientmethods for the solution of large-scale systems that stem, e.g.from the discretization of partial differential equations (PDE). Inthis talk we will review the generalization of these methods tothe solution of matrix equations (L. Grasedyck, W. Hackbusch, AMultigrid Method to Solve Large Scale Sylvester Equations,SIAM J. Matrix Anal. Appl. 29, pp. 870–894, 2007; L.Grasedyck, Nonlinear multigrid for the solution of large scaleRiccati equations in low-rank and H-matrix format,Num.Lin.Alg.Appl. 15, pp. 779–807, 2008), and equations thatpossess a tensor structure (L. Grasedyck, Hierarchical SingularValue Decomposition of Tensors, SIAM J. Matrix Anal. Appl. 31,pp. 2029–2054, 2010.). The standard hierarchical and multigridmethods can perfectly be combined with low rank (matrix) andlow tensor rank representations. The benefit is that the solutionis computable in almost optimal complexity with respect to theamount of data needed for the representation of the solution. Asan example we consider a PDE posed in a product domainΩ× Ω, Ω ⊂ Rd and discretized with Nd basis functions for thedomain Ω. Under separability assumptions on the right-handside the system is solved in low rank form in O(Nd) complexity(instead of O(N2d) required for the full solution). For a PDE onthe product domain Ω× · · · × Ω one can even solve the systemin low tensor rank form in O(N · d) complexity (instead ofO(N2d) required for the full solution). The state of the art willbe shortly summarized.Lars GrasedyckInst. for Geometry and Practical MathematicsRWTH [email protected]

Talk 2. A survey on Newton-ADI based solvers for large scaleAREsNewton based solvers for the algebraic Riccati equation (ARE)have been arround for several decades. Only roughly one decadeago these have become applicable to large and sparse AREswhen combined with the low rank Cholesky factor alternatingdirections implicit (LRCF-ADI) iteration for solving theLyapunov equations arising in every Newton step. In thiscontribution we give a survey on accelerated variants of the lowrank Cholesky factor Newton method (LRCF-NM) developedover the recent years. These include projection based methodsand inexact Newton-like approaches.Jens SaakComputational Methods in Systems and Control TheoryMax Planck Institute for Dynamics of Complex Technical Systemsand Department of MathematicsChemnitz University [email protected]

Talk 3. An invariant subspace method for large-scalealgebraic Riccati and Bernoulli equations

We present a new family of low-rank approximations of thesolution of the algebraic Riccati equation based on stableinvariant subspaces of the Hamiltonian matrix ( L. Amodei andJ.-M. Buchot, An invariant subspace method for large-scalealgebraic Riccati equations, Appl. Numer. Math., 60 (11),1067-1082, 2010.). The approximations are given by afactorized form which preserves the positive semi-definiteness ofthe exact solution. By using the same general factorization, wededuce a reduced rank expression of the exact solution of theBernoulli equation. We illustrate the effectiveness of thisformulation by considering the stabilization of the nonstationaryincompressible Navier-Stokes equations by a boundary feedbackcontrol obtained from the solution of the Bernoulli equation (L.Amodei and J.-M. Buchot, A stabilization algorithm of theNavier-Stokes equations based on algebraic Bernoulli equation,Numer. Linear Algebra Appl., DOI: 10.1002/nla.799, 2011.).Luca AmodeiInstitut de Mathematiques de ToulouseUniversite Paul Sabatier118 route de Narbonne31062 Toulouse Cedex 9, [email protected]

Jean-Marie BuchotInstitut de Mathematiques de ToulouseUniversite Paul [email protected]

Talk 4. Delay Lyapunov equations and model order reductionof time delay systemsWe present a version of balanced truncation for model orderreduction of linear time-delay systems. The procedure is basedon a coordinate transformation of the position and preserves thedelay structure of the system. To every position we associatequantities representing energies for the controllability andobservability of the position. We show that these energies can beexpressed explicitly in terms of Gramians which are given assolutions to corresponding delay Lyapunov equations. Balancedtruncation can be based on these Gramians in the usual way.Tobias DammDept. of MathematicsUniversity of [email protected]

Elias JarlebringDept. Num. Analysis,KTH, [email protected]

Wim MichielsDepartment of Computer ScienceK.U. [email protected]

MS 75. Points that minimize potential functionsTalk 1. Discretizing compact manifolds with minimal energyOne approach to generate “good” point configurations isinspired by physics: points emulate repelling unit point chargesinteracting through a Coulomb potential 1/r, where r measuresthe Euclidean distance between points in the ambient space.Points on the sphere that minimize the corresponding energy(potential energy) are uniformly distributed even if a Rieszs-potential 1/rs governs the point interaction.On other compact manifolds minimal energy systems areuniformly distributed if s is greater than the dimension d of themanifold (“Poppy Seed Bagle” theorem).

65 2012 SIAM Conference on Applied Linear Algebra

This talk gives an introduction into the discrete minimal Rieszs-energy problem on compact manifolds.Johann S. BrauchartSchool of Mathematics and StatisticsUniversity of New South [email protected]

Talk 2. Well conditioned spherical designs and potentialfunctionsA spherical t-design is a set of N points on the unit sphere suchthat equal weighted numerical integration at these points is exactfor all spherical polynomials of degree at most t. This talk willlook at the calculation and properties of spherical designs withN between t2/2 and (t+ 1)2, their characterization by differentpotential functions, and using any degrees of freedom to alsominimize the condition number of the basis matrix or maximizethe determinant of the associated Gram matrix.Robert S. WomersleySchool of Mathematics and StatisticsUniversity of New South [email protected]

Talk 3. Probabilistic frames in the 2-Wasserstein metricIn this talk I will review the notion of probabilistic frames andindicate how it is a natural generalization of frames. In addition,I will show how within the framework of the 2-Wassersteinmetric, probabilistic frames can be constructed and analyzed.This is based on joint work with M. Ehler.Kasso OkoudjouDept. of MathematicsUniversity of [email protected]

Talk 4. Numerical minimization of potential energies onspecific manifoldsIn this talk we consider the problem of computing localminimizers of potential energies of the form

E(x1, . . . ,xM ) :=

M∑i,j=1;i6=j

K(xi,xj),

with xi ∈M, i = 1, . . . ,M, whereM⊂ Rn is ad-dimensional compact manifold and K :M×M→ R is agiven function. The optimization approach is based on anonlinear conjugate gradient method on Riemannian manifolds,which is a generalization of the CG-method in Euclidean space.This method was already successfully applied to thecomputation of spherical t-designs in [Numer. Math., 119:699 –724, 2011] and low discrepancy points for polynomial kernelsK, cf. [TU Chemnitz, Preprint 5, 2011]. Now, we presentinteresting numerical results for localized non-polynomialdiscrepancy kernels K.Manuel GrafDept. of MathematicsChemnitz University of [email protected]