%%% -*-BibTeX-*- %%% ==================================================================== %%% BibTeX-file{ %%% author = "Nelson H. F. Beebe", %%% version = "1.07", %%% date = "03 March 2014", %%% time = "10:37:22 MST", %%% filename = "stoc2010.bib", %%% address = "University of Utah %%% Department of Mathematics, 110 LCB %%% 155 S 1400 E RM 233 %%% Salt Lake City, UT 84112-0090 %%% USA", %%% telephone = "+1 801 581 5254", %%% FAX = "+1 801 581 4148", %%% URL = "http://www.math.utah.edu/~beebe", %%% checksum = "32326 10295 60046 545877", %%% email = "beebe at math.utah.edu, beebe at acm.org, %%% beebe at computer.org (Internet)", %%% codetable = "ISO/ASCII", %%% keywords = "ACM Symposium on Theory of Computing (STOC)", %%% license = "public domain", %%% supported = "yes", %%% docstring = "This is a COMPLETE bibliography of %%% publications in the ACM Symposium on Theory %%% of Computing (STOC) conference proceedings %%% for the decade 2010--2019. Companion %%% bibliographies stoc19xx.bib and stoc20xx.bib %%% cover other decades, and stoc.bib contains %%% entries for just the proceedings volumes %%% themselves. %%% %%% There is World-Wide Web sites for these %%% publications at %%% %%% http://dl.acm.org/event.cfm?id=RE224 %%% http://www.sigact.org/stoc.html?searchterm=STOC %%% http://www.acm.org/pubs/contents/proceedings/series/stoc/ %%% %%% At version 1.07, the year coverage looked %%% like this: %%% %%% 2006 ( 2) 2009 ( 0) 2012 ( 90) %%% 2007 ( 0) 2010 ( 83) 2013 ( 101) %%% 2008 ( 0) 2011 ( 85) %%% %%% InProceedings: 356 %%% Proceedings: 5 %%% %%% Total entries: 361 %%% %%% The checksum field above contains a CRC-16 %%% checksum as the first value, followed by the %%% equivalent of the standard UNIX wc (word %%% count) utility output of lines, words, and %%% characters. This is produced by Robert %%% Solovay's checksum utility.", %%% } %%% ==================================================================== @Preamble{ "\ifx \undefined \mathbb \def \mathbb #1{{\bf #1}}\fi" # "\ifx \undefined \norm \def \norm {{\rm norm}} \fi" # "\ifx \undefined \ocirc \def \ocirc #1{{\accent'27#1}}\fi" # "\ifx \undefined \poly \def \poly {{\rm poly}} \fi" # "\ifx \undefined \polylog \def \polylog {{\rm polylog}} \fi" # "\ifx \undefined \rank \def \rank {{\rm rank}} \fi" } %%% ==================================================================== %%% Acknowledgement abbreviations: @String{ack-nhfb= "Nelson H. F. Beebe, University of Utah, Department of Mathematics, 110 LCB, 155 S 1400 E RM 233, Salt Lake City, UT 84112-0090, USA, Tel: +1 801 581 5254, FAX: +1 801 581 4148, e-mail: \path|beebe@math.utah.edu|, \path|beebe@acm.org|, \path|beebe@computer.org| (Internet), URL: \path|http://www.math.utah.edu/~beebe/|"} %%% ==================================================================== %%% Publisher abbreviations: @String{pub-ACM= "ACM Press"} @String{pub-ACM:adr= "New York, NY, USA"} %%% ==================================================================== %%% Bibliography entries: @InProceedings{Akavia:2006:BOW, author = "Adi Akavia and Oded Goldreich and Shafi Goldwasser and Dana Moshkovitz", title = "On basing one-way functions on {NP}-hardness", crossref = "ACM:2006:SPT", pages = "701--710", year = "2006", bibdate = "Thu May 25 06:19:54 MDT 2006", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", note = "See erratum \cite{Akavia:2010:EBO}.", acknowledgement = ack-nhfb, } @InProceedings{Kannan:2010:SMM, author = "Ravindran Kannan", title = "Spectral methods for matrices and tensors", crossref = "ACM:2010:SPA", pages = "1--12", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Talagrand:2010:MSS, author = "Michel Talagrand", title = "Are many small sets explicitly small?", crossref = "ACM:2010:SPA", pages = "13--36", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Montanari:2010:MPA, author = "Andrea Montanari", title = "Message passing algorithms: a success looking for theoreticians", crossref = "ACM:2010:SPA", pages = "37--38", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Goel:2010:PML, author = "Ashish Goel and Michael Kapralov and Sanjeev Khanna", title = "Perfect matchings in $o(n \log n)$ time in regular bipartite graphs", crossref = "ACM:2010:SPA", pages = "39--46", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Leighton:2010:ELV, author = "F. Thomson Leighton and Ankur Moitra", title = "Extensions and limits to vertex sparsification", crossref = "ACM:2010:SPA", pages = "47--56", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kolla:2010:SSN, author = "Alexandra Kolla and Yury Makarychev and Amin Saberi and Shang-Hua Teng", title = "Subgraph sparsification and nearly optimal ultrasparsifiers", crossref = "ACM:2010:SPA", pages = "57--66", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Barak:2010:HCI, author = "Boaz Barak and Mark Braverman and Xi Chen and Anup Rao", title = "How to compress interactive communication", crossref = "ACM:2010:SPA", pages = "67--76", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Klauck:2010:SDP, author = "Hartmut Klauck", title = "A strong direct product theorem for disjointness", crossref = "ACM:2010:SPA", pages = "77--86", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Beame:2010:HAP, author = "Paul Beame and Trinh Huynh and Toniann Pitassi", title = "Hardness amplification in proof complexity", crossref = "ACM:2010:SPA", pages = "87--96", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Gao:2010:LBO, author = "Pu Gao and Nicholas C. Wormald", title = "Load balancing and orientability thresholds for random hypergraphs", crossref = "ACM:2010:SPA", pages = "97--104", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bayati:2010:CAI, author = "Mohsen Bayati and David Gamarnik and Prasad Tetali", title = "Combinatorial approach to the interpolation method and scaling limits in sparse random graphs", crossref = "ACM:2010:SPA", pages = "105--114", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Hirai:2010:MMP, author = "Hiroshi Hirai", title = "The maximum multiflow problems with bounded fractionality", crossref = "ACM:2010:SPA", pages = "115--120", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Madry:2010:FAS, author = "Aleksander Madry", title = "Faster approximation schemes for fractional multicommodity flow problems via dynamic graph algorithms", crossref = "ACM:2010:SPA", pages = "121--130", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Aaronson:2010:FCQ, author = "Scott Aaronson and Andrew Drucker", title = "A full characterization of quantum advice", crossref = "ACM:2010:SPA", pages = "131--140", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Aaronson:2010:BPH, author = "Scott Aaronson", title = "{BQP} and the polynomial hierarchy", crossref = "ACM:2010:SPA", pages = "141--150", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Ambainis:2010:QLL, author = "Andris Ambainis and Julia Kempe and Or Sattath", title = "A quantum {Lov{\'a}sz} local lemma", crossref = "ACM:2010:SPA", pages = "151--160", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{De:2010:NOE, author = "Anindya De and Thomas Vidick", title = "Near-optimal extractors against quantum storage", crossref = "ACM:2010:SPA", pages = "161--170", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Applebaum:2010:PKC, author = "Benny Applebaum and Boaz Barak and Avi Wigderson", title = "Public-key cryptography from different assumptions", crossref = "ACM:2010:SPA", pages = "171--180", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Ajtai:2010:ORC, author = "Mikl{\'o}s Ajtai", title = "Oblivious {RAM}s without cryptographic assumptions", crossref = "ACM:2010:SPA", pages = "181--190", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Goyal:2010:RCC, author = "Vipul Goyal and Abhishek Jain", title = "On the round complexity of covert computation", crossref = "ACM:2010:SPA", pages = "191--200", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bhaskara:2010:DHL, author = "Aditya Bhaskara and Moses Charikar and Eden Chlamtac and Uriel Feige and Aravindan Vijayaraghavan", title = "Detecting high log-densities: an {$O(n^{1/4})$} approximation for densest $k$-subgraph", crossref = "ACM:2010:SPA", pages = "201--210", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bateni:2010:ASS, author = "MohammadHossein Bateni and MohammadTaghi Hajiaghayi and D{\'a}niel Marx", title = "Approximation schemes for {Steiner} forest on planar graphs and graphs of bounded treewidth", crossref = "ACM:2010:SPA", pages = "211--220", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dey:2010:OHC, author = "Tamal K. Dey and Anil N. Hirani and Bala Krishnamoorthy", title = "Optimal homologous cycles, total unimodularity, and linear programming", crossref = "ACM:2010:SPA", pages = "221--230", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Williams:2010:IES, author = "Ryan Williams", title = "Improving exhaustive search implies superpolynomial lower bounds", crossref = "ACM:2010:SPA", pages = "231--240", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Paturi:2010:CCS, author = "Ramamohan Paturi and Pavel Pudlak", title = "On the complexity of circuit satisfiability", crossref = "ACM:2010:SPA", pages = "241--250", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dell:2010:SAN, author = "Holger Dell and Dieter van Melkebeek", title = "Satisfiability allows no nontrivial sparsification unless the polynomial-time hierarchy collapses", crossref = "ACM:2010:SPA", pages = "251--260", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Magniez:2010:RWP, author = "Fr{\'e}d{\'e}ric Magniez and Claire Mathieu and Ashwin Nayak", title = "Recognizing well-parenthesized expressions in the streaming model", crossref = "ACM:2010:SPA", pages = "261--270", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2010:MID, author = "Vladimir Braverman and Rafail Ostrovsky", title = "Measuring independence of datasets", crossref = "ACM:2010:SPA", pages = "271--280", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2010:ZOF, author = "Vladimir Braverman and Rafail Ostrovsky", title = "Zero-one frequency laws", crossref = "ACM:2010:SPA", pages = "281--290", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Orlin:2010:IAC, author = "James B. Orlin", title = "Improved algorithms for computing {Fisher}'s market clearing prices", crossref = "ACM:2010:SPA", pages = "291--300", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Hartline:2010:BAM, author = "Jason D. Hartline and Brendan Lucier", title = "{Bayesian} algorithmic mechanism design", crossref = "ACM:2010:SPA", pages = "301--310", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chawla:2010:MPM, author = "Shuchi Chawla and Jason D. Hartline and David L. Malec and Balasubramanian Sivan", title = "Multi-parameter mechanism design and sequential posted pricing", crossref = "ACM:2010:SPA", pages = "311--320", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lokshtanov:2010:SSA, author = "Daniel Lokshtanov and Jesper Nederlof", title = "Saving space by algebraization", crossref = "ACM:2010:SPA", pages = "321--330", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Haramaty:2010:SCQ, author = "Elad Haramaty and Amir Shpilka", title = "On the structure of cubic and quartic polynomials", crossref = "ACM:2010:SPA", pages = "331--340", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dasgupta:2010:SJL, author = "Anirban Dasgupta and Ravi Kumar and Tam{\'a}s Sarlos", title = "A sparse {Johnson--Lindenstrauss} transform", crossref = "ACM:2010:SPA", pages = "341--350", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Micciancio:2010:DSE, author = "Daniele Micciancio and Panagiotis Voulgaris", title = "A deterministic single exponential time algorithm for most lattice problems based on {Voronoi} cell computations", crossref = "ACM:2010:SPA", pages = "351--358", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Cardinal:2010:SUP, author = "Jean Cardinal and Samuel Fiorini and Gwena{\"e}l Joret and Rapha{\"e}l M. Jungers and J. Ian Munro", title = "Sorting under partial information (without the ellipsoid algorithm)", crossref = "ACM:2010:SPA", pages = "359--368", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lee:2010:MMP, author = "Jon Lee and Maxim Sviridenko and Jan Vondrak", title = "Matroid matching: the power of local search", crossref = "ACM:2010:SPA", pages = "369--378", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bhattacharya:2010:BCA, author = "Sayan Bhattacharya and Gagan Goel and Sreenivas Gollapudi and Kamesh Munagala", title = "Budget constrained auctions with heterogeneous items", crossref = "ACM:2010:SPA", pages = "379--388", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Fraigniaud:2010:SSW, author = "Pierre Fraigniaud and George Giakkoupis", title = "On the searchability of small-world networks with arbitrary underlying structure", crossref = "ACM:2010:SPA", pages = "389--398", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chierichetti:2010:ATB, author = "Flavio Chierichetti and Silvio Lattanzi and Alessandro Panconesi", title = "Almost tight bounds for rumour spreading with conductance", crossref = "ACM:2010:SPA", pages = "399--408", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Guruswami:2010:LDR, author = "Venkatesan Guruswami and Johan H{\aa}stad and Swastik Kopparty", title = "On the list-decodability of random linear codes", crossref = "ACM:2010:SPA", pages = "409--416", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kopparty:2010:LLD, author = "Swastik Kopparty and Shubhangi Saraf", title = "Local list-decoding and testing of random linear codes from high error", crossref = "ACM:2010:SPA", pages = "417--426", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Meka:2010:PGP, author = "Raghu Meka and David Zuckerman", title = "Pseudorandom generators for polynomial threshold functions", crossref = "ACM:2010:SPA", pages = "427--436", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Haitner:2010:EIC, author = "Iftach Haitner and Omer Reingold and Salil Vadhan", title = "Efficiency improvements in constructing pseudorandom generators from one-way functions", crossref = "ACM:2010:SPA", pages = "437--446", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Verbin:2010:LBT, author = "Elad Verbin and Qin Zhang", title = "The limits of buffering: a tight lower bound for dynamic membership in the external memory model", crossref = "ACM:2010:SPA", pages = "447--456", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Onak:2010:MLM, author = "Krzysztof Onak and Ronitt Rubinfeld", title = "Maintaining a large matching and a small vertex cover", crossref = "ACM:2010:SPA", pages = "457--464", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Duan:2010:COF, author = "Ran Duan and Seth Pettie", title = "Connectivity oracles for failure prone graphs", crossref = "ACM:2010:SPA", pages = "465--474", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Gilbert:2010:ASR, author = "Anna C. Gilbert and Yi Li and Ely Porat and Martin J. Strauss", title = "Approximate sparse recovery: optimizing time and measurements", crossref = "ACM:2010:SPA", pages = "475--484", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Godoy:2010:HPD, author = "Guillem Godoy and Omer Gim{\'e}nez and Lander Ramos and Carme {\`A}lvarez", title = "The {HOM} problem is decidable", crossref = "ACM:2010:SPA", pages = "485--494", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kawamura:2010:CTO, author = "Akitoshi Kawamura and Stephen Cook", title = "Complexity theory for operators in analysis", crossref = "ACM:2010:SPA", pages = "495--502", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Burgisser:2010:SPE, author = "Peter B{\"u}rgisser and Felipe Cucker", title = "Solving polynomial equations in smoothed polynomial time and a near solution to {Smale}'s 17th problem", crossref = "ACM:2010:SPA", pages = "503--512", year = "2010", DOI = "http://doi.acm.org/10.1145/1806689.1806759", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The 17th of the problems proposed by Steve Smale for the 21st century asks for the existence of a deterministic algorithm computing an approximate solution of a system of $n$ complex polynomials in $n$ unknowns in time polynomial, on the average, in the size $N$ of the input system. A partial solution to this problem was given by Carlos Beltran and Luis Miguel Pardo who exhibited a randomized algorithm, call it LV, doing so. In this paper we further extend this result in several directions. Firstly, we perform a smoothed analysis (in the sense of Spielman and Teng) of algorithm LV and prove that its smoothed complexity is polynomial in the input size and $\sigma - 1$, where $\sigma$ controls the size of the random perturbation of the input systems. Secondly, we perform a condition-based analysis of LV. That is, we give a bound, for each system $f$, of the expected running time of LV with input $f$. In addition to its dependence on $N$ this bound also depends on the condition of $f$. Thirdly, and to conclude, we return to Smale's 17th problem as originally formulated for deterministic algorithms. We exhibit such an algorithm and show that its average complexity is $N^{O(\log \log N)}$. This is nearly a solution to Smale's 17th problem.", acknowledgement = ack-nhfb, } @InProceedings{Kuhn:2010:DCD, author = "Fabian Kuhn and Nancy Lynch and Rotem Oshman", title = "Distributed computation in dynamic networks", crossref = "ACM:2010:SPA", pages = "513--522", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Sherstov:2010:OBS, author = "Alexander A. Sherstov", title = "Optimal bounds for sign-representing the intersection of two halfspaces by polynomials", crossref = "ACM:2010:SPA", pages = "523--532", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Diakonikolas:2010:BAS, author = "Ilias Diakonikolas and Prahladh Harsha and Adam Klivans and Raghu Meka and Prasad Raghavendra and Rocco A. Servedio and Li-Yang Tan", title = "Bounding the average sensitivity and noise sensitivity of polynomial threshold functions", crossref = "ACM:2010:SPA", pages = "533--542", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Harsha:2010:IPP, author = "Prahladh Harsha and Adam Klivans and Raghu Meka", title = "An invariance principle for polytopes", crossref = "ACM:2010:SPA", pages = "543--552", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kalai:2010:ELM, author = "Adam Tauman Kalai and Ankur Moitra and Gregory Valiant", title = "Efficiently learning mixtures of two {Gaussians}", crossref = "ACM:2010:SPA", pages = "553--562", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{VAgh:2010:AUN, author = "L{\'a}szl{\'o} A. V{\'e}gh", title = "Augmenting undirected node-connectivity by one", crossref = "ACM:2010:SPA", pages = "563--572", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Jain:2010:QP, author = "Rahul Jain and Zhengfeng Ji and Sarvagya Upadhyay and John Watrous", title = "{QIP $=$ PSPACE}", crossref = "ACM:2010:SPA", pages = "573--582", year = "2010", DOI = "http://dx.doi.org/10.1145/1806689.1806768", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove that the complexity class QIP, which consists of all problems having quantum interactive proof systems, is contained in PSPACE. This containment is proved by applying a parallelized form of the matrix multiplicative weights update method to a class of semidefinite programs that captures the computational power of quantum interactive proofs. As the containment of PSPACE in QIP follows immediately from the well-known equality IP $=$ PSPACE, the equality QIP $=$ PSPACE follows.", acknowledgement = ack-nhfb, remark = "This work won the conference's Best Paper Award. An updated version appears in Comm. ACM 53(12) 102--109 (December 2010), \url{http://dx.doi.org/10.1145/1859204.1859231}.", } @InProceedings{Byrka:2010:ILB, author = "Jaroslaw Byrka and Fabrizio Grandoni and Thomas Rothvo{\ss} and Laura Sanit{\`a}", title = "An improved {LP}-based approximation for {Steiner} tree", crossref = "ACM:2010:SPA", pages = "583--592", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dodis:2010:CBL, author = "Yevgeniy Dodis and Mihai Patrascu and Mikkel Thorup", title = "Changing base without losing space", crossref = "ACM:2010:SPA", pages = "593--602", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Patrascu:2010:TPL, author = "Mihai Patrascu", title = "Towards polynomial lower bounds for dynamic problems", crossref = "ACM:2010:SPA", pages = "603--610", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Fraigniaud:2010:OAS, author = "Pierre Fraigniaud and Amos Korman", title = "An optimal ancestry scheme and small universal posets", crossref = "ACM:2010:SPA", pages = "611--620", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lee:2010:BSM, author = "James R. Lee and Mohammad Moharrami", title = "Bilipschitz snowflakes and metrics of negative type", crossref = "ACM:2010:SPA", pages = "621--630", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Raghavendra:2010:AIS, author = "Prasad Raghavendra and David Steurer and Prasad Tetali", title = "Approximations for the isoperimetric and spectral profile of graphs and related parameters", crossref = "ACM:2010:SPA", pages = "631--640", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Varadarajan:2010:WGS, author = "Kasturi Varadarajan", title = "Weighted geometric set cover via quasi-uniform sampling", crossref = "ACM:2010:SPA", pages = "641--648", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Karnin:2010:DIT, author = "Zohar S. Karnin and Partha Mukhopadhyay and Amir Shpilka and Ilya Volkovich", title = "Deterministic identity testing of depth-4 multilinear circuits with bounded top fan-in", crossref = "ACM:2010:SPA", pages = "649--658", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Raz:2010:TRL, author = "Ran Raz", title = "Tensor-rank and lower bounds for arithmetic formulas", crossref = "ACM:2010:SPA", pages = "659--666", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{HrubeAa:2010:NCC, author = "Pavel Hrube and Avi Wigderson and Amir Yehudayoff", title = "Non-commutative circuits and the sum-of-squares problem", crossref = "ACM:2010:SPA", pages = "667--676", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Arvind:2010:HND, author = "Vikraman Arvind and Srikanth Srinivasan", title = "On the hardness of the noncommutative determinant", crossref = "ACM:2010:SPA", pages = "677--686", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kawarabayashi:2010:SPG, author = "Ken-ichi Kawarabayashi and Paul Wollan", title = "A shorter proof of the graph minor algorithm: the unique linkage theorem", crossref = "ACM:2010:SPA", pages = "687--694", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kawarabayashi:2010:OCP, author = "Ken-ichi Kawarabayashi and Bruce Reed", title = "Odd cycle packing", crossref = "ACM:2010:SPA", pages = "695--704", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Hardt:2010:GDP, author = "Moritz Hardt and Kunal Talwar", title = "On the geometry of differential privacy", crossref = "ACM:2010:SPA", pages = "705--714", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dwork:2010:DPU, author = "Cynthia Dwork and Moni Naor and Toniann Pitassi and Guy N. Rothblum", title = "Differential privacy under continual observation", crossref = "ACM:2010:SPA", pages = "715--724", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dyer:2010:CC, author = "Martin E. Dyer and David M. Richerby", title = "On the complexity of {\#CSP}", crossref = "ACM:2010:SPA", pages = "725--734", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Marx:2010:THP, author = "D{\'a}niel Marx", title = "Tractable hypergraph properties for constraint satisfaction and conjunctive queries", crossref = "ACM:2010:SPA", pages = "735--744", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Svensson:2010:CHP, author = "Ola Svensson", title = "Conditional hardness of precedence constrained scheduling on identical machines", crossref = "ACM:2010:SPA", pages = "745--754", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Raghavendra:2010:GEU, author = "Prasad Raghavendra and David Steurer", title = "Graph expansion and the unique games conjecture", crossref = "ACM:2010:SPA", pages = "755--764", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Roth:2010:IPM, author = "Aaron Roth and Tim Roughgarden", title = "Interactive privacy via the median mechanism", crossref = "ACM:2010:SPA", pages = "765--774", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kasiviswanathan:2010:PPR, author = "Shiva Prasad Kasiviswanathan and Mark Rudelson and Adam Smith and Jonathan Ullman", title = "The price of privately releasing contingency tables and the spectra of random matrices with correlated rows", crossref = "ACM:2010:SPA", pages = "775--784", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chandran:2010:PAA, author = "Nishanth Chandran and Bhavana Kanukurthi and Rafail Ostrovsky and Leonid Reyzin", title = "Privacy amplification with asymptotically optimal entropy loss", crossref = "ACM:2010:SPA", pages = "785--794", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Akavia:2010:EBO, author = "Adi Akavia and Oded Goldreich and Shafi Goldwasser and Dana Moshkovitz", title = "Erratum for: {{\em On basing one-way functions on NP-hardness}}", crossref = "ACM:2010:SPA", pages = "795--796", year = "2010", bibdate = "Wed Sep 1 10:42:57 MDT 2010", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", note = "See \cite{Akavia:2006:BOW}.", acknowledgement = ack-nhfb, } @InProceedings{Patrascu:2011:PST, author = "Mihai Patrascu and Mikkel Thorup", title = "The power of simple tabulation hashing", crossref = "ACM:2011:SPA", pages = "1--10", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993638", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lenzen:2011:TBP, author = "Christoph Lenzen and Roger Wattenhofer", title = "Tight bounds for parallel randomized load balancing: extended abstract", crossref = "ACM:2011:SPA", pages = "11--20", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993639", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Doerr:2011:SNS, author = "Benjamin Doerr and Mahmoud Fouz and Tobias Friedrich", title = "Social networks spread rumors in sublogarithmic time", crossref = "ACM:2011:SPA", pages = "21--30", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993640", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Regev:2011:QOW, author = "Oded Regev and Bo'az Klartag", title = "Quantum one-way communication can be exponentially stronger than classical communication", crossref = "ACM:2011:SPA", pages = "31--40", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993642", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Sherstov:2011:SDP, author = "Alexander A. Sherstov", title = "Strong direct product theorems for quantum communication and query complexity", crossref = "ACM:2011:SPA", pages = "41--50", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993643", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chakrabarti:2011:OLB, author = "Amit Chakrabarti and Oded Regev", title = "An optimal lower bound on the communication complexity of gap-{Hamming}-distance", crossref = "ACM:2011:SPA", pages = "51--60", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993644", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Ding:2011:CTB, author = "Jian Ding and James R. Lee and Yuval Peres", title = "Cover times, blanket times, and majorizing measures", crossref = "ACM:2011:SPA", pages = "61--70", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993646", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Fung:2011:GFG, author = "Wai Shing Fung and Ramesh Hariharan and Nicholas J. A. Harvey and Debmalya Panigrahi", title = "A general framework for graph sparsification", crossref = "ACM:2011:SPA", pages = "71--80", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993647", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kawarabayashi:2011:BAA, author = "Ken-ichi Kawarabayashi and Yusuke Kobayashi", title = "Breaking $o(n^{1/2})$-approximation algorithms for the edge-disjoint paths problem with congestion two", crossref = "ACM:2011:SPA", pages = "81--88", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993648", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Holenstein:2011:ERO, author = "Thomas Holenstein and Robin K{\"u}nzler and Stefano Tessaro", title = "The equivalence of the random oracle model and the ideal cipher model, revisited", crossref = "ACM:2011:SPA", pages = "89--98", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993650", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Gentry:2011:SSN, author = "Craig Gentry and Daniel Wichs", title = "Separating succinct non-interactive arguments from all falsifiable assumptions", crossref = "ACM:2011:SPA", pages = "99--108", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993651", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Pass:2011:LPS, author = "Rafael Pass", title = "Limits of provable security from standard assumptions", crossref = "ACM:2011:SPA", pages = "109--118", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993652", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Papadimitriou:2011:OSI, author = "Christos H. Papadimitriou and George Pierrakos", title = "On optimal single-item auctions", crossref = "ACM:2011:SPA", pages = "119--128", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993654", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dobzinski:2011:OAC, author = "Shahar Dobzinski and Hu Fu and Robert D. Kleinberg", title = "Optimal auctions with correlated bidders are easy", crossref = "ACM:2011:SPA", pages = "129--138", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993655", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dobzinski:2011:IRT, author = "Shahar Dobzinski", title = "An impossibility result for truthful combinatorial auctions with submodular valuations", crossref = "ACM:2011:SPA", pages = "139--148", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993656", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dughmi:2011:COR, author = "Shaddin Dughmi and Tim Roughgarden and Qiqi Yan", title = "From convex optimization to randomized mechanisms: toward optimal combinatorial auctions", crossref = "ACM:2011:SPA", pages = "149--158", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993657", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2011:TCM, author = "Mark Braverman and Anup Rao", title = "Towards coding for maximum errors in interactive communication", crossref = "ACM:2011:SPA", pages = "159--166", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993659", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kopparty:2011:HRC, author = "Swastik Kopparty and Shubhangi Saraf and Sergey Yekhanin", title = "High-rate codes with sublinear-time decoding", crossref = "ACM:2011:SPA", pages = "167--176", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993660", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Zewi:2011:ATS, author = "Noga Zewi and Eli Ben-Sasson", title = "From affine to two-source extractors via approximate duality", crossref = "ACM:2011:SPA", pages = "177--186", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993661", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Hatami:2011:CTA, author = "Hamed Hatami and Shachar Lovett", title = "Correlation testing for affine invariant properties on {$\mathbb{F}_p^n$} in the high error regime", crossref = "ACM:2011:SPA", pages = "187--194", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993662", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Adsul:2011:RBG, author = "Bharat Adsul and Jugal Garg and Ruta Mehta and Milind Sohoni", title = "Rank-1 bimatrix games: a homeomorphism and a polynomial time algorithm", crossref = "ACM:2011:SPA", pages = "195--204", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993664", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Hansen:2011:EAS, author = "Kristoffer Arnsfelt Hansen and Michal Koucky and Niels Lauritzen and Peter Bro Miltersen and Elias P. Tsigaridas", title = "Exact algorithms for solving stochastic games: extended abstract", crossref = "ACM:2011:SPA", pages = "205--214", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993665", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Immorlica:2011:DA, author = "Nicole Immorlica and Adam Tauman Kalai and Brendan Lucier and Ankur Moitra and Andrew Postlewaite and Moshe Tennenholtz", title = "Dueling algorithms", crossref = "ACM:2011:SPA", pages = "215--224", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993666", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Moitra:2011:POS, author = "Ankur Moitra and Ryan O'Donnell", title = "{Pareto} optimal solutions for smoothed analysts", crossref = "ACM:2011:SPA", pages = "225--234", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993667", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kolipaka:2011:MTM, author = "Kashyap Babu Rao Kolipaka and Mario Szegedy", title = "{Moser} and {Tardos} meet {Lov{\'a}sz}", crossref = "ACM:2011:SPA", pages = "235--244", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993669", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Moser:2011:FDS, author = "Robin A. Moser and Dominik Scheder", title = "A full derandomization of {Sch{\"o}ning}'s {$k$-SAT} algorithm", crossref = "ACM:2011:SPA", pages = "245--252", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993670", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Gopalan:2011:PGC, author = "Parikshit Gopalan and Raghu Meka and Omer Reingold and David Zuckerman", title = "Pseudorandom generators for combinatorial shapes", crossref = "ACM:2011:SPA", pages = "253--262", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993671", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Koucky:2011:PGG, author = "Michal Kouck{\'y} and Prajakta Nimbhorkar and Pavel Pudl{\'a}k", title = "Pseudorandom generators for group products: extended abstract", crossref = "ACM:2011:SPA", pages = "263--272", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993672", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Christiano:2011:EFL, author = "Paul Christiano and Jonathan A. Kelner and Aleksander Madry and Daniel A. Spielman and Shang-Hua Teng", title = "Electrical flows, {Laplacian} systems, and faster approximation of maximum flow in undirected graphs", crossref = "ACM:2011:SPA", pages = "273--282", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993674", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Friedmann:2011:SLB, author = "Oliver Friedmann and Thomas Dueholm Hansen and Uri Zwick", title = "Subexponential lower bounds for randomized pivoting rules for the simplex algorithm", crossref = "ACM:2011:SPA", pages = "283--292", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993675", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Haeupler:2011:ANC, author = "Bernhard Haeupler", title = "Analyzing network coding gossip made easy", crossref = "ACM:2011:SPA", pages = "293--302", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993676", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chuzhoy:2011:AGC, author = "Julia Chuzhoy", title = "An algorithm for the graph crossing number problem", crossref = "ACM:2011:SPA", pages = "303--312", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993678", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Italiano:2011:IAM, author = "Giuseppe F. Italiano and Yahav Nussbaum and Piotr Sankowski and Christian Wulff-Nilsen", title = "Improved algorithms for min cut and max flow in undirected planar graphs", crossref = "ACM:2011:SPA", pages = "313--322", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993679", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Dinitz:2011:DSF, author = "Michael Dinitz and Robert Krauthgamer", title = "Directed spanners via flow-based linear programs", crossref = "ACM:2011:SPA", pages = "323--332", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993680", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Aaronson:2011:CCL, author = "Scott Aaronson and Alex Arkhipov", title = "The computational complexity of linear optics", crossref = "ACM:2011:SPA", pages = "333--342", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993682", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Brandao:2011:QTA, author = "Fernando G. S. L. Brand{\~a}o and Matthias Christandl and Jon Yard", title = "A quasipolynomial-time algorithm for the quantum separability problem", crossref = "ACM:2011:SPA", pages = "343--352", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993683", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kempe:2011:PRE, author = "Julia Kempe and Thomas Vidick", title = "Parallel repetition of entangled games", crossref = "ACM:2011:SPA", pages = "353--362", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993684", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Sarma:2011:DVH, author = "Atish Das Sarma and Stephan Holzer and Liah Kor and Amos Korman and Danupon Nanongkai and Gopal Pandurangan and David Peleg and Roger Wattenhofer", title = "Distributed verification and hardness of distributed approximation", crossref = "ACM:2011:SPA", pages = "363--372", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993686", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Golab:2011:LID, author = "Wojciech Golab and Lisa Higham and Philipp Woelfel", title = "Linearizable implementations do not suffice for randomized distributed computation", crossref = "ACM:2011:SPA", pages = "373--382", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993687", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kantor:2011:TWC, author = "Erez Kantor and Zvi Lotker and Merav Parter and David Peleg", title = "The topology of wireless communication", crossref = "ACM:2011:SPA", pages = "383--392", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993688", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Giakkoupis:2011:OPS, author = "George Giakkoupis and Nicolas Schabanel", title = "Optimal path search in small worlds: dimension matters", crossref = "ACM:2011:SPA", pages = "393--402", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993689", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Novocin:2011:LRA, author = "Andrew Novocin and Damien Stehl{\'e} and Gilles Villard", title = "An {LLL}-reduction algorithm with quasi-linear time complexity: extended abstract", crossref = "ACM:2011:SPA", pages = "403--412", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993691", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Khot:2011:NHA, author = "Subhash Khot and Dana Moshkovitz", title = "{NP}-hardness of approximately solving linear equations over reals", crossref = "ACM:2011:SPA", pages = "413--420", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993692", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Saraf:2011:BBI, author = "Shubhangi Saraf and Ilya Volkovich", title = "Black-box identity testing of depth-4 multilinear circuits", crossref = "ACM:2011:SPA", pages = "421--430", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993693", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Saxena:2011:BIT, author = "Nitin Saxena and C. Seshadhri", title = "Blackbox identity testing for bounded top fanin depth-3 circuits: the field doesn't matter", crossref = "ACM:2011:SPA", pages = "431--440", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993694", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Demaine:2011:CDH, author = "Erik D. Demaine and MohammadTaghi Hajiaghayi and Ken-ichi Kawarabayashi", title = "Contraction decomposition in $h$-minor-free graphs and algorithmic applications", crossref = "ACM:2011:SPA", pages = "441--450", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993696", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kawarabayashi:2011:SAS, author = "Ken-ichi Kawarabayashi and Paul Wollan", title = "A simpler algorithm and shorter proof for the graph minor decomposition", crossref = "ACM:2011:SPA", pages = "451--458", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993697", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bousquet:2011:MF, author = "Nicolas Bousquet and Jean Daligault and St{\'e}phan Thomass{\'e}", title = "Multicut is {FPT}", crossref = "ACM:2011:SPA", pages = "459--468", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993698", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Marx:2011:FPT, author = "D{\'a}niel Marx and Igor Razgon", title = "Fixed-parameter tractability of multicut parameterized by the size of the cutset", crossref = "ACM:2011:SPA", pages = "469--478", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993699", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Grohe:2011:FTS, author = "Martin Grohe and Ken-ichi Kawarabayashi and D{\'a}niel Marx and Paul Wollan", title = "Finding topological subgraphs is fixed-parameter tractable", crossref = "ACM:2011:SPA", pages = "479--488", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993700", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kopparty:2011:CPF, author = "Swastik Kopparty", title = "On the complexity of powering in finite fields", crossref = "ACM:2011:SPA", pages = "489--498", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993702", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Chien:2011:ASH, author = "Steve Chien and Prahladh Harsha and Alistair Sinclair and Srikanth Srinivasan", title = "Almost settling the hardness of noncommutative determinant", crossref = "ACM:2011:SPA", pages = "499--508", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993703", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Burgisser:2011:GCT, author = "Peter B{\"u}rgisser and Christian Ikenmeyer", title = "Geometric complexity theory and tensor rank", crossref = "ACM:2011:SPA", pages = "509--518", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993704", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Barak:2011:RBD, author = "Boaz Barak and Zeev Dvir and Amir Yehudayoff and Avi Wigderson", title = "Rank bounds for design matrices with applications to combinatorial geometry and locally correctable codes", crossref = "ACM:2011:SPA", pages = "519--528", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993705", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kleinberg:2011:MMA, author = "Jon Kleinberg and Sigal Oren", title = "Mechanisms for (mis)allocating scientific credit", crossref = "ACM:2011:SPA", pages = "529--538", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993707", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Cole:2011:IPS, author = "Richard Cole and Jos{\'e} R. Correa and Vasilis Gkatzelis and Vahab Mirrokni and Neil Olver", title = "Inner product spaces for {MinSum} coordination mechanisms", crossref = "ACM:2011:SPA", pages = "539--548", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993708", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Feige:2011:MDU, author = "Uriel Feige and Moshe Tennenholtz", title = "Mechanism design with uncertain inputs: (to err is human, to forgive divine)", crossref = "ACM:2011:SPA", pages = "549--558", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993709", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Patrascu:2011:DRU, author = "Mihai P{\u{a}}tra{\c{s}}cu and Mikkel Thorup", title = "Don't rush into a union: take time to find your roots", crossref = "ACM:2011:SPA", pages = "559--568", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993711", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Feldman:2011:UFA, author = "Dan Feldman and Michael Langberg", title = "A unified framework for approximating and clustering data", crossref = "ACM:2011:SPA", pages = "569--578", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993712", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Arya:2011:APM, author = "Sunil Arya and Guilherme D. da Fonseca and David M. Mount", title = "Approximate polytope membership queries", crossref = "ACM:2011:SPA", pages = "579--586", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993713", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Karande:2011:OBM, author = "Chinmay Karande and Aranyak Mehta and Pushkar Tripathi", title = "Online bipartite matching with unknown distributions", crossref = "ACM:2011:SPA", pages = "587--596", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993715", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Mahdian:2011:OBM, author = "Mohammad Mahdian and Qiqi Yan", title = "Online bipartite matching with random arrivals: an approach based on strongly factor-revealing {LPs}", crossref = "ACM:2011:SPA", pages = "597--606", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993716", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Adamaszek:2011:ATB, author = "Anna Adamaszek and Artur Czumaj and Matthias Englert and Harald R{\"a}cke", title = "Almost tight bounds for reordering buffer management", crossref = "ACM:2011:SPA", pages = "607--616", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993717", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Svensson:2011:SCS, author = "Ola Svensson", title = "{Santa Claus} schedules jobs on unrelated machines", crossref = "ACM:2011:SPA", pages = "617--626", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993718", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Indyk:2011:KMC, author = "Piotr Indyk and Eric Price", title = "{$K$}-median clustering, model-based compressive sensing, and sparse recovery for earth mover distance", crossref = "ACM:2011:SPA", pages = "627--636", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993720", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bourgain:2011:BBE, author = "Jean Bourgain and Stephen J. Dilworth and Kevin Ford and Sergei V. Konyagin and Denka Kutzarova", title = "Breaking the $k^2$ barrier for explicit {RIP} matrices", crossref = "ACM:2011:SPA", pages = "637--644", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993721", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Karnin:2011:DCH, author = "Zohar S. Karnin", title = "Deterministic construction of a high dimensional $l_p$ section in $l_1^n$ for any $p < 2$", crossref = "ACM:2011:SPA", pages = "645--654", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993722", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Bodirsky:2011:STG, author = "Manuel Bodirsky and Michael Pinsker", title = "{Schaefer}'s theorem for graphs", crossref = "ACM:2011:SPA", pages = "655--664", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993724", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Yoshida:2011:OCT, author = "Yuichi Yoshida", title = "Optimal constant-time approximation algorithms and (unconditional) inapproximability results for every bounded-degree {CSP}", crossref = "ACM:2011:SPA", pages = "665--674", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993725", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Newman:2011:EPH, author = "Ilan Newman and Christian Sohler", title = "Every property of hyperfinite graphs is testable", crossref = "ACM:2011:SPA", pages = "675--684", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993726", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Valiant:2011:EUS, author = "Gregory Valiant and Paul Valiant", title = "Estimating the unseen: an $n / \log(n)$-sample estimator for entropy and support size, shown optimal via new {CLTs}", crossref = "ACM:2011:SPA", pages = "685--694", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993727", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Goyal:2011:CRN, author = "Vipul Goyal", title = "Constant round non-malleable protocols using one way functions", crossref = "ACM:2011:SPA", pages = "695--704", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993729", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lin:2011:CRN, author = "Huijia Lin and Rafael Pass", title = "Constant-round non-malleable commitments from any one-way function", crossref = "ACM:2011:SPA", pages = "705--714", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993730", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Ajtai:2011:SCI, author = "Miklos Ajtai", title = "Secure computation with information leaking to an adversary", crossref = "ACM:2011:SPA", pages = "715--724", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993731", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lewko:2011:HLK, author = "Allison Lewko and Mark Lewko and Brent Waters", title = "How to leak on key updates", crossref = "ACM:2011:SPA", pages = "725--734", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993732", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Woodruff:2011:NOP, author = "David P. Woodruff", title = "Near-optimal private approximation protocols via a black box transformation", crossref = "ACM:2011:SPA", pages = "735--744", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993733", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kane:2011:FME, author = "Daniel M. Kane and Jelani Nelson and Ely Porat and David P. Woodruff", title = "Fast moment estimation in data streams in optimal space", crossref = "ACM:2011:SPA", pages = "745--754", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993735", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Sohler:2011:SEN, author = "Christian Sohler and David P. Woodruff", title = "Subspace embeddings for the {$L_1$}-norm with applications", crossref = "ACM:2011:SPA", pages = "755--764", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993736", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Lee:2011:NOD, author = "James R. Lee and Anastasios Sidiropoulos", title = "Near-optimal distortion bounds for embedding doubling spaces into {$L_1$}", crossref = "ACM:2011:SPA", pages = "765--772", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993737", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Fawzi:2011:LDN, author = "Omar Fawzi and Patrick Hayden and Pranab Sen", title = "From low-distortion norm embeddings to explicit uncertainty relations and efficient information locking", crossref = "ACM:2011:SPA", pages = "773--782", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993738", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Vondrak:2011:SFM, author = "Jan Vondr{\'a}k and Chandra Chekuri and Rico Zenklusen", title = "Submodular function maximization via the multilinear relaxation and contention resolution schemes", crossref = "ACM:2011:SPA", pages = "783--792", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993740", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Balcan:2011:LSF, author = "Maria-Florina Balcan and Nicholas J. A. Harvey", title = "Learning submodular functions", crossref = "ACM:2011:SPA", pages = "793--802", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993741", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Gupta:2011:PRC, author = "Anupam Gupta and Moritz Hardt and Aaron Roth and Jonathan Ullman", title = "Privately releasing conjunctions and the statistical query barrier", crossref = "ACM:2011:SPA", pages = "803--812", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993742", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Smith:2011:PPS, author = "Adam Smith", title = "Privacy-preserving statistical estimation with optimal convergence rates", crossref = "ACM:2011:SPA", pages = "813--822", year = "2011", DOI = "http://dx.doi.org/10.1145/1993636.1993743", bibdate = "Tue Jun 7 18:53:27 MDT 2011", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, } @InProceedings{Kelner:2012:FAM, author = "Jonathan A. Kelner and Gary L. Miller and Richard Peng", title = "Faster approximate multicommodity flow using quadratically coupled flows", crossref = "ACM:2012:SPA", pages = "1--18", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213979", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The maximum multicommodity flow problem is a natural generalization of the maximum flow problem to route multiple distinct flows. Obtaining a $1 - \epsilon$ approximation to the multicommodity flow problem on graphs is a well-studied problem. In this paper we present an adaptation of recent advances in single-commodity flow algorithms to this problem. As the underlying linear systems in the electrical problems of multicommodity flow problems are no longer Laplacians, our approach is tailored to generate specialized systems which can be preconditioned and solved efficiently using Laplacians. Given an undirected graph with $m$ edges and $k$ commodities, we give algorithms that find $1 - \epsilon$ approximate solutions to the maximum concurrent flow problem and maximum weighted multicommodity flow problem in time $O(m^{4/3} \poly(k, \epsilon^{-1}))$.", acknowledgement = ack-nhfb, } @InProceedings{Chakrabarti:2012:WCC, author = "Amit Chakrabarti and Lisa Fleischer and Christophe Weibel", title = "When the cut condition is enough: a complete characterization for multiflow problems in series-parallel networks", crossref = "ACM:2012:SPA", pages = "19--26", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213980", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Let $G = (V, E)$ be a supply graph and $H = (V,F)$ a demand graph defined on the same set of vertices. An assignment of capacities to the edges of $G$ and demands to the edges of $H$ is said to satisfy the cut condition if for any cut in the graph, the total demand crossing the cut is no more than the total capacity crossing it. The pair $(G, H)$ is called cut-sufficient if for any assignment of capacities and demands that satisfy the cut condition, there is a multiflow routing the demands defined on $H$ within the network with capacities defined on $G$. We prove a previous conjecture, which states that when the supply graph $G$ is series-parallel, the pair (G,H) is cut-sufficient if and only if $(G, H)$ does not contain an odd spindle as a minor; that is, if it is impossible to contract edges of $G$ and delete edges of $G$ and $H$ so that $G$ becomes the complete bipartite graph $K_{2,p}$, with $p \geq 3$ odd, and $H$ is composed of a cycle connecting the $p$ vertices of degree $2$, and an edge connecting the two vertices of degree $p$. We further prove that if the instance is Eulerian --- that is, the demands and capacities are integers and the total of demands and capacities incident to each vertex is even --- then the multiflow problem has an integral solution. We provide a polynomial-time algorithm to find an integral solution in this case. In order to prove these results, we formulate properties of tight cuts (cuts for which the cut condition inequality is tight) in cut-sufficient pairs. We believe these properties might be useful in extending our results to planar graphs.", acknowledgement = ack-nhfb, } @InProceedings{Vegh:2012:SPA, author = "L{\'a}szl{\'o} A. V{\'e}gh", title = "Strongly polynomial algorithm for a class of minimum-cost flow problems with separable convex objectives", crossref = "ACM:2012:SPA", pages = "27--40", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213981", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A well-studied nonlinear extension of the minimum-cost flow problem is to minimize the objective $\Sigma_{i j \in E} C_{ij} (f_{ij})$ over feasible flows $f$, where on every arc $ij$ of the network, $C_{ij}$ is a convex function. We give a strongly polynomial algorithm for finding an exact optimal solution for a broad class of such problems. The key characteristic of this class is that an optimal solution can be computed exactly provided its support. This includes separable convex quadratic objectives and also certain market equilibria problems: Fisher's market with linear and with spending constraint utilities. We thereby give the first strongly polynomial algorithms for separable quadratic minimum-cost flows and for Fisher's market with spending constraint utilities, settling open questions posed e.g. in [15] and in [35], respectively. The running time is $O(m^4 \log m)$ for quadratic costs, $O(n^4 + n^2 (m + n \log n) \log n)$ for Fisher's markets with linear utilities and $O(m n^3 + m^2 (m + n \log n) \log m)$ for spending constraint utilities.", acknowledgement = ack-nhfb, } @InProceedings{Aaronson:2012:QMH, author = "Scott Aaronson and Paul Christiano", title = "Quantum money from hidden subspaces", crossref = "ACM:2012:SPA", pages = "41--60", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213983", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Forty years ago, Wiesner pointed out that quantum mechanics raises the striking possibility of money that cannot be counterfeited according to the laws of physics. We propose the first quantum money scheme that is (1) public-key --- meaning that anyone can verify a banknote as genuine, not only the bank that printed it, and (2) cryptographically secure, under a ``classical'' hardness assumption that has nothing to do with quantum money. Our scheme is based on hidden subspaces, encoded as the zero-sets of random multivariate polynomials. A main technical advance is to show that the ``black-box'' version of our scheme, where the polynomials are replaced by classical oracles, is unconditionally secure. Previously, such a result had only been known relative to a quantum oracle (and even there, the proof was never published). Even in Wiesner's original setting --- quantum money that can only be verified by the bank --- we are able to use our techniques to patch a major security hole in Wiesner's scheme. We give the first private-key quantum money scheme that allows unlimited verifications and that remains unconditionally secure, even if the counterfeiter can interact adaptively with the bank. Our money scheme is simpler than previous public-key quantum money schemes, including a knot-based scheme of Farhi et al. The verifier needs to perform only two tests, one in the standard basis and one in the Hadamard basis --- matching the original intuition for quantum money, based on the existence of complementary observables. Our security proofs use a new variant of Ambainis's quantum adversary method, and several other tools that might be of independent interest.", acknowledgement = ack-nhfb, } @InProceedings{Vazirani:2012:CQD, author = "Umesh Vazirani and Thomas Vidick", title = "Certifiable quantum dice: or, true random number generation secure against quantum adversaries", crossref = "ACM:2012:SPA", pages = "61--76", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213984", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We introduce a protocol through which a pair of quantum mechanical devices may be used to generate $n$ bits that are $\epsilon$-close in statistical distance from $n$ uniformly distributed bits, starting from a seed of $O(\log n \log 1 / \epsilon)$ uniform bits. The bits generated are certifiably random based only on a simple statistical test that can be performed by the user, and on the assumption that the devices do not communicate in the middle of each phase of the protocol. No other assumptions are placed on the devices' inner workings. A modified protocol uses a seed of $O(\log^3 n)$ uniformly random bits to generate $n$ bits that are $\poly^{-1}(n)$-indistinguishable from uniform even from the point of view of a quantum adversary who may have had prior access to the devices, and may be entangled with them.", acknowledgement = ack-nhfb, } @InProceedings{Belovs:2012:SPF, author = "Aleksandrs Belovs", title = "Span programs for functions with constant-sized $1$-certificates: extended abstract", crossref = "ACM:2012:SPA", pages = "77--84", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213985", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Besides the Hidden Subgroup Problem, the second large class of quantum speed-ups is for functions with constant-sized 1-certificates. This includes the OR function, solvable by the Grover algorithm, the element distinctness, the triangle and other problems. The usual way to solve them is by quantum walk on the Johnson graph. We propose a solution for the same problems using span programs. The span program is a computational model equivalent to the quantum query algorithm in its strength, and yet very different in its outfit. We prove the power of our approach by designing a quantum algorithm for the triangle problem with query complexity $O(n^{35/27})$ that is better than $O(n^{13/10})$ of the best previously known algorithm by Magniez et al.", acknowledgement = ack-nhfb, } @InProceedings{Larsen:2012:CPC, author = "Kasper Green Larsen", title = "The cell probe complexity of dynamic range counting", crossref = "ACM:2012:SPA", pages = "85--94", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213987", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In this paper we develop a new technique for proving lower bounds on the update time and query time of dynamic data structures in the cell probe model. With this technique, we prove the highest lower bound to date for any explicit problem, namely a lower bound of $t_q = \Omega((\lg n / \lg (wt_u))^2)$. Here $n$ is the number of update operations, $w$ the cell size, $t_q$ the query time and t$_u$ the update time. In the most natural setting of cell size $w = \Theta(\lg n)$, this gives a lower bound of $t_q = \Omega((\lg n / \lg \lg n)^2)$ for any polylogarithmic update time. This bound is almost a quadratic improvement over the highest previous lower bound of $\Omega(\lg n)$, due to Patrascu and Demaine [SICOMP'06]. We prove our lower bound for the fundamental problem of weighted orthogonal range counting. In this problem, we are to support insertions of two-dimensional points, each assigned a $\Theta(\lg n)$-bit integer weight. A query to this problem is specified by a point $q = (x, y)$, and the goal is to report the sum of the weights assigned to the points dominated by $q$, where a point $(x',y')$ is dominated by $q$ if $x' \leq x$ and $y' \leq y$. In addition to being the highest cell probe lower bound to date, our lower bound is also tight for data structures with update time $t_u = \Omega(\lg^{2 + \epsilon} n)$, where $\epsilon > 0$ is an arbitrarily small constant.", acknowledgement = ack-nhfb, } @InProceedings{Fiorini:2012:LVS, author = "Samuel Fiorini and Serge Massar and Sebastian Pokutta and Hans Raj Tiwary and Ronald de Wolf", title = "Linear vs. semidefinite extended formulations: exponential separation and strong lower bounds", crossref = "ACM:2012:SPA", pages = "95--106", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213988", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We solve a 20-year old problem posed by Yannakakis and prove that there exists no polynomial-size linear program (LP) whose associated polytope projects to the traveling salesman polytope, even if the LP is not required to be symmetric. Moreover, we prove that this holds also for the cut polytope and the stable set polytope. These results were discovered through a new connection that we make between one-way quantum communication protocols and semidefinite programming reformulations of LPs.", acknowledgement = ack-nhfb, } @InProceedings{Goel:2012:PCA, author = "Gagan Goel and Vahab Mirrokni and Renato Paes Leme", title = "Polyhedral clinching auctions and the {AdWords} polytope", crossref = "ACM:2012:SPA", pages = "107--122", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213990", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A central issue in applying auction theory in practice is the problem of dealing with budget-constrained agents. A desirable goal in practice is to design incentive compatible, individually rational, and Pareto optimal auctions while respecting the budget constraints. Achieving this goal is particularly challenging in the presence of nontrivial combinatorial constraints over the set of feasible allocations. Toward this goal and motivated by AdWords auctions, we present an auction for polymatroidal environments satisfying the above properties. Our auction employs a novel clinching technique with a clean geometric description and only needs an oracle access to the submodular function defining the polymatroid. As a result, this auction not only simplifies and generalizes all previous results, it applies to several new applications including AdWords Auctions, bandwidth markets, and video on demand. In particular, our characterization of the AdWords auction as polymatroidal constraints might be of independent interest. This allows us to design the first mechanism for Ad Auctions taking into account simultaneously budgets, multiple keywords and multiple slots. We show that it is impossible to extend this result to generic polyhedral constraints. This also implies an impossibility result for multi-unit auctions with decreasing marginal utilities in the presence of budget constraints.", acknowledgement = ack-nhfb, } @InProceedings{Kleinberg:2012:MPI, author = "Robert Kleinberg and Seth Matthew Weinberg", title = "Matroid prophet inequalities", crossref = "ACM:2012:SPA", pages = "123--136", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213991", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Consider a gambler who observes a sequence of independent, non-negative random numbers and is allowed to stop the sequence at any time, claiming a reward equal to the most recent observation. The famous prophet inequality of Krengel, Sucheston, and Garling asserts that a gambler who knows the distribution of each random variable can achieve at least half as much reward, in expectation, as a ``prophet'' who knows the sampled values of each random variable and can choose the largest one. We generalize this result to the setting in which the gambler and the prophet are allowed to make more than one selection, subject to a matroid constraint. We show that the gambler can still achieve at least half as much reward as the prophet; this result is the best possible, since it is known that the ratio cannot be improved even in the original prophet inequality, which corresponds to the special case of rank-one matroids. Generalizing the result still further, we show that under an intersection of $p$ matroid constraints, the prophet's reward exceeds the gambler's by a factor of at most $O(p)$, and this factor is also tight. Beyond their interest as theorems about pure online algorithms or optimal stopping rules, these results also have applications to mechanism design. Our results imply improved bounds on the ability of sequential posted-price mechanisms to approximate optimal mechanisms in both single-parameter and multi-parameter Bayesian settings. In particular, our results imply the first efficiently computable constant-factor approximations to the Bayesian optimal revenue in certain multi-parameter settings.", acknowledgement = ack-nhfb, } @InProceedings{Devanur:2012:OMC, author = "Nikhil R. Devanur and Kamal Jain", title = "Online matching with concave returns", crossref = "ACM:2012:SPA", pages = "137--144", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213992", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider a significant generalization of the AdWords problem by allowing arbitrary concave returns, and we characterize the optimal competitive ratio achievable. The problem considers a sequence of items arriving online that have to be allocated to agents, with different agents bidding different amounts. The objective function is the sum, over each agent i, of a monotonically non-decreasing concave function $M_i: R_? + \to R_+$ of the total amount allocated to $i$. All variants of online matching problems (including the AdWords problem) studied in the literature consider the special case of budgeted linear functions, that is, functions of the form $M_i(u_i) = \min\{u_i, B_i\}$ for some constant $B_i$. The distinguishing feature of this paper is in allowing arbitrary concave returns. The main result of this paper is that for each concave function $M$, there exists a constant $F(M) \leq 1$ such that: there exists an algorithm with competitive ratio of $\min_i F(M_i)$, independent of the sequence of items. No algorithm has a competitive ratio larger than $F(M)$ over all instances with $M_i = M$ for all $i$. Our algorithm is based on the primal-dual paradigm and makes use of convex programming duality. The upper bounds are obtained by formulating the task of finding the right counterexample as an optimization problem. This path takes us through the calculus of variations which deals with optimizing over continuous functions. The algorithm and the upper bound are related to each other via a set of differential equations, which points to a certain kind of duality between them.", acknowledgement = ack-nhfb, } @InProceedings{Arora:2012:CNM, author = "Sanjeev Arora and Rong Ge and Ravindran Kannan and Ankur Moitra", title = "Computing a nonnegative matrix factorization --- provably", crossref = "ACM:2012:SPA", pages = "145--162", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213994", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Nonnegative Matrix Factorization (NMF) problem has a rich history spanning quantum mechanics, probability theory, data analysis, polyhedral combinatorics, communication complexity, demography, chemometrics, etc. In the past decade NMF has become enormously popular in machine learning, where the factorization is computed using a variety of local search heuristics. Vavasis recently proved that this problem is NP-complete. We initiate a study of when this problem is solvable in polynomial time. Consider a nonnegative $m \times n$ matrix $M$ and a target inner-dimension $r$. Our results are the following: --- We give a polynomial-time algorithm for exact and approximate NMF for every constant $r$. Indeed NMF is most interesting in applications precisely when $r$ is small. We complement this with a hardness result, that if exact NMF can be solved in time $(n m)^{o(r)}$, 3-SAT has a sub-exponential time algorithm. Hence, substantial improvements to the above algorithm are unlikely. --- We give an algorithm that runs in time polynomial in n, $m$ and $r$ under the separability condition identified by Donoho and Stodden in 2003. The algorithm may be practical since it is simple and noise tolerant (under benign assumptions). Separability is believed to hold in many practical settings. To the best of our knowledge, this last result is the first polynomial-time algorithm that provably works under a non-trivial condition on the input matrix and we believe that this will be an interesting and important direction for future work.", acknowledgement = ack-nhfb, } @InProceedings{Forbes:2012:ITT, author = "Michael A. Forbes and Amir Shpilka", title = "On identity testing of tensors, low-rank recovery and compressed sensing", crossref = "ACM:2012:SPA", pages = "163--172", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213995", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the problem of obtaining efficient, deterministic, {\em black-box polynomial identity testing algorithms\/} for depth-$3$ set-multilinear circuits (over arbitrary fields). This class of circuits has an efficient, deterministic, white-box polynomial identity testing algorithm (due to Raz and Shpilka [36]), but has no known such black-box algorithm. We recast this problem as a question of finding a low-dimensional subspace $H$, spanned by rank $1$ tensors, such that any non-zero tensor in the dual space $\ker(H)$ has high rank. We obtain explicit constructions of essentially optimal-size hitting sets for tensors of degree $2$ (matrices), and obtain the first quasi-polynomial sized hitting sets for arbitrary tensors.\par We also show connections to the task of performing low-rank recovery of matrices, which is studied in the field of compressed sensing. Low-rank recovery asks (say, over $R$) to recover a matrix $M$ from few measurements, under the promise that $M$ is rank $\leq r$. In this work, we restrict our attention to recovering matrices that are exactly rank $\leq r$ using deterministic, non-adaptive, linear measurements, that are free from noise. Over $R$, we provide a set (of size $4 n r$) of such measurements, from which $M$ can be recovered in $O(r n^2 + r^3 n)$ field operations, and the number of measurements is essentially optimal. Further, the measurements can be taken to be all rank-$1$ matrices, or all sparse matrices. To the best of our knowledge no explicit constructions with those properties were known prior to this work.\par We also give a more formal connection between low-rank recovery and the task of {\em sparse (vector) recovery\/}: any sparse-recovery algorithm that exactly recovers vectors of length $n$ and sparsity $2 r$, using $m$ non-adaptive measurements, yields a low-rank recovery scheme for exactly recovering $n \times n$ matrices of rank $\leq r$, making $2 n m$ non-adaptive measurements. Furthermore, if the sparse-recovery algorithm runs in time $\tau$, then the low-rank recovery algorithm runs in time $O(r n^2 + n \tau)$. We obtain this reduction using linear-algebraic techniques, and not using convex optimization, which is more commonly seen in compressed sensing algorithms.\par Finally, we also make a connection to {\em rank-metric codes}, as studied in coding theory. These are codes with codewords consisting of matrices (or tensors) where the distance of matrices $M$ and $N$ is rank $(M - N)$, as opposed to the usual Hamming metric. We obtain essentially optimal-rate codes over matrices, and provide an efficient decoding algorithm. We obtain codes over tensors as well, with poorer rate, but still with efficient decoding.", acknowledgement = ack-nhfb, } @InProceedings{Grohe:2012:STI, author = "Martin Grohe and D{\'a}niel Marx", title = "Structure theorem and isomorphism test for graphs with excluded topological subgraphs", crossref = "ACM:2012:SPA", pages = "173--192", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213996", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We generalize the structure theorem of Robertson and Seymour for graphs excluding a fixed graph $H$ as a minor to graphs excluding $H$ as a topological subgraph. We prove that for a fixed H, every graph excluding $H$ as a topological subgraph has a tree decomposition where each part is either ``almost embeddable'' to a fixed surface or has bounded degree with the exception of a bounded number of vertices. Furthermore, such a decomposition is computable by an algorithm that is fixed-parameter tractable with parameter $|H|$. We present two algorithmic applications of our structure theorem. To illustrate the mechanics of a ``typical'' application of the structure theorem, we show that on graphs excluding $H$ as a topological subgraph, Partial Dominating Set (find $k$ vertices whose closed neighborhood has maximum size) can be solved in time $f(H,k) \cdot n^{O(1)}$ time. More significantly, we show that on graphs excluding $H$ as a topological subgraph, Graph Isomorphism can be solved in time $n^{f(H)}$. This result unifies and generalizes two previously known important polynomial-time solvable cases of Graph Isomorphism: bounded-degree graphs and H-minor free graphs. The proof of this result needs a generalization of our structure theorem to the context of invariant treelike decomposition.", acknowledgement = ack-nhfb, } @InProceedings{Hrubes:2012:SPD, author = "Pavel Hrubes and Iddo Tzameret", title = "Short proofs for the determinant identities", crossref = "ACM:2012:SPA", pages = "193--212", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213998", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study arithmetic proof systems $P_c(F)$ and $P_f(F)$ operating with arithmetic circuits and arithmetic formulas, respectively, that prove polynomial identities over a field $F$. We establish a series of structural theorems about these proof systems, the main one stating that $P_c(F)$ proofs can be balanced: if a polynomial identity of syntactic degree $d$ and depth $k$ has a $P_c(F)$ proof of size $s$, then it also has a $P_c(F)$ proof of size $\poly(s,d)$ and depth $O(k + \log^2 d + \log d \cdot \log s)$. As a corollary, we obtain a quasipolynomial simulation of $P_c(F)$ by $P_f(F)$, for identities of a polynomial syntactic degree.\par Using these results we obtain the following: consider the identities:\par $$\det(X Y) = \det(X) \cdot \det(Y) \quad {\rm and} \quad \det(Z) = z_{11} \cdots z_{nn},$$\par where $X$, $Y$ and $Z$ are $n \times n$ square matrices and $Z$ is a triangular matrix with $z_{11}, \ldots{}, z_{nn}$ on the diagonal (and $\det$ is the determinant polynomial). Then we can construct a polynomial-size arithmetic circuit $\det$ such that the above identities have $P_c(F)$ proofs of polynomial-size and $O(\log^2 n)$ depth. Moreover, there exists an arithmetic formula det of size $n^{O(\log n)}$ such that the above identities have $P_f(F)$ proofs of size $n^{O(\log n)}$.\par This yields a solution to a basic open problem in propositional proof complexity, namely, whether there are polynomial-size NC$^2$-Frege proofs for the determinant identities and the hard matrix identities, as considered, e.g. in Soltys and Cook (2004) (cf., Beame and Pitassi (1998)). We show that matrix identities like $A B = I \to B A = I$ (for matrices over the two element field) as well as basic properties of the determinant have polynomial-size NC$^2$-Frege proofs, and quasipolynomial-size Frege proofs.", acknowledgement = ack-nhfb, } @InProceedings{Beame:2012:TST, author = "Paul Beame and Christopher Beck and Russell Impagliazzo", title = "Time-space tradeoffs in resolution: superpolynomial lower bounds for superlinear space", crossref = "ACM:2012:SPA", pages = "213--232", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2213999", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give the first time-space tradeoff lower bounds for Resolution proofs that apply to superlinear space. In particular, we show that there are formulas of size $N$ that have Resolution refutations of space and size each roughly $N^{\log 2 N}$ (and like all formulas have Resolution refutations of space $N$) for which any Resolution refutation using space $S$ and length $T$ requires $T \geq (N^{0.58 \log 2 N} /S)^{\Omega(\log \log N/\log \log \log N)}$. By downward translation, a similar tradeoff applies to all smaller space bounds. We also show somewhat stronger time-space tradeoff lower bounds for Regular Resolution, which are also the first to apply to superlinear space. Namely, for any space bound $S$ at most $2^{o(N^{1/4})}$ there are formulas of size $N$, having clauses of width $4$, that have Regular Resolution proofs of space $S$ and slightly larger size $T = O(NS)$, but for which any Regular Resolution proof of space $S^{1 - \epsilon}$ requires length $T^{\Omega(\log \log N / \log \log \log N)}$.", acknowledgement = ack-nhfb, } @InProceedings{Huynh:2012:VSP, author = "Trinh Huynh and Jakob Nordstrom", title = "On the virtue of succinct proofs: amplifying communication complexity hardness to time-space trade-offs in proof complexity", crossref = "ACM:2012:SPA", pages = "233--248", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214000", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "An active line of research in proof complexity over the last decade has been the study of proof space and trade-offs between size and space. Such questions were originally motivated by practical SAT solving, but have also led to the development of new theoretical concepts in proof complexity of intrinsic interest and to results establishing nontrivial relations between space and other proof complexity measures. By now, the resolution proof system is fairly well understood in this regard, as witnessed by a sequence of papers leading up to [Ben-Sasson and Nordstrom 2008, 2011] and [Beame, Beck, and Impagliazzo 2012]. However, for other relevant proof systems in the context of SAT solving, such as polynomial calculus (PC) and cutting planes (CP), very little has been known. Inspired by [BN08, BN11], we consider CNF encodings of so-called pebble games played on graphs and the approach of making such pebbling formulas harder by simple syntactic modifications. We use this paradigm of hardness amplification to make progress on the relatively longstanding open question of proving time-space trade-offs for PC and CP. Namely, we exhibit a family of modified pebbling formulas $\{F_n\}$ such that: --- The formulas $F_n$ have size $O(n)$ and width $O(1)$. --- They have proofs in length $O(n)$ in resolution, which generalize to both PC and CP. --- Any refutation in CP or PCR (a generalization of PC) in length $L$ and space $s$ must satisfy $s \log L > \approx \sqrt [4]{n}$. A crucial technical ingredient in these results is a new two-player communication complexity lower bound for composed search problems in terms of block sensitivity, a contribution that we believe to be of independent interest.", acknowledgement = ack-nhfb, } @InProceedings{Ajtai:2012:DVN, author = "Mikl{\'o}s Ajtai", title = "Determinism versus nondeterminism with arithmetic tests and computation: extended abstract", crossref = "ACM:2012:SPA", pages = "249--268", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214001", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "For each natural number $d$ we consider a finite structure $M_d$ whose universe is the set of all $0, 1$-sequence of length $n = 2^d$, each representing a natural number in the set $\{0, 1, \ldots{}, 2^n - 1\}$ in binary form. The operations included in the structure are the four constants $0$, $1$, $2^n - 1$, $n$, multiplication and addition modulo $2^n$, the unary function $\min\{2^x, 2^n - 1\}$, the binary functions $\lfloor x / y \rfloor$ (with $\lfloor x / 0 \rfloor = 0$), $\max(x, y)$, $\min(x, y)$, and the boolean vector operations $\wedge$, $\vee$, [??not??] defined on $0, 1$ sequences of length $n$ by performing the operations on all components simultaneously. These are essentially the arithmetic operations that can be performed on a RAM, with wordlength $n$, by a single instruction. We show that there exists a term (that is, an algebraic expression) $F(x,y)$ built up from the mentioned operations, with the only free variables $x$, $y$, such that for all terms $G(y)$, which is also built up from the mentioned operations, the following holds. For infinitely many positive integers $d$, there exists an $a \in M_d$ such that the following two statements are not equivalent: (i) $M_d \models \exists x, F(x,a)$, (ii) $M_d \models G(a) = 0$. In other words, the question whether an existential statement, depending on the parameter $a \in M_d$ is true or not, cannot be decided by evaluating an algebraic expression at $a$.\par Another way of formulating the theorem, in a slightly stronger form, is, that over the structures $M_d$, quantifier elimination is not possible in the following sense. Let $\cal M$ be a first-order language with equality, containing function symbols for all of the mentioned arithmetic operations. Then there exists an existential first-order formula $\phi(y)$ of $\cal M$, containing a single existential quantifier and the only free variable $y$, such that for each propositional formula $P(y)$ of $\cal M$, we have that for infinitely many positive integers $d$, $\phi(y)$ and $P(y)$ are not equivalent on $M_d$, that is, $M_d \models {\em [not]} \forall y$, $\phi(y) \leftrightarrow P(y)$.\par We also show that the theorem, in both forms, remains true if the binary operation $\min\{x^y, 2^n - 1\}$ is added to the structure $M_d$. A general theorem is proved as well, which describes sufficient conditions for a set of operations on a sequence of structures $K_d$, $d = 1, 2, \ldots{}$ which guarantees that the analogues of the mentioned theorems holds for the structures $K_d$ too.", acknowledgement = ack-nhfb, } @InProceedings{Heilman:2012:SPC, author = "Steven Heilman and Aukosh Jagannath and Assaf Naor", title = "Solution of the propeller conjecture in {$\mathbb{R}^3$}", crossref = "ACM:2012:SPA", pages = "269--276", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214003", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "It is shown that every measurable partition $\{A_1, \ldots{}, A_k\}$ of $\mathbb{R}^3$ satisfies:\par $$\sum_{i = 1}^k || \int_{A i} x e^{-1/2||x||^2_2} \, dx||_2^2 \leq 9 \pi ^2$$.\par Let $\{P_1, P_2, P_3\}$ be the partition of $\mathbb{R}^2$ into $120^\circ$ sectors centered at the origin. The bound (1) is sharp, with equality holding if $A_i = P_i \times \mathbb{R}$ for $i \in \{1, 2, 3\}$ and $A_i = \emptyset$ for $i \in \{4, \ldots{}, k\}$. This settles positively the 3-dimensional Propeller Conjecture of Khot and Naor (FOCS 2008). The proof of (1) reduces the problem to a finite set of numerical inequalities which are then verified with full rigor in a computer-assisted fashion. The main consequence (and motivation) of (1) is complexity-theoretic: the Unique Games hardness threshold of the Kernel Clustering problem with $4 \times 4$ centered and spherical hypothesis matrix equals $2 \pi /3$.", acknowledgement = ack-nhfb, } @InProceedings{Khot:2012:LHC, author = "Subhash A. Khot and Preyas Popat and Nisheeth K. Vishnoi", title = "$2^{\log^{1 - \epsilon} n}$ hardness for the closest vector problem with preprocessing", crossref = "ACM:2012:SPA", pages = "277--288", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214004", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove that for an arbitrarily small constant $\epsilon > 0$, assuming NP $\not\subseteq$ DTIME $(2^{\log^{O(1 / \epsilon) n}})$, the preprocessing versions of the closest vector problem and the nearest codeword problem are hard to approximate within a factor better than $2^{\log^{1 - \epsilon} n}$. This improves upon the previous hardness factor of $(\log n)^\delta$ for some $\delta > 0$ due to [AKKV05].", acknowledgement = ack-nhfb, } @InProceedings{ODonnell:2012:NPN, author = "Ryan O'Donnell and John Wright", title = "A new point of {NP}-hardness for unique games", crossref = "ACM:2012:SPA", pages = "289--306", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214005", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show that distinguishing $1/2$-satisfiable Unique-Games instances from $(3/8 + \epsilon)$-satisfiable instances is NP-hard (for all $\epsilon > 0$). A consequence is that we match or improve the best known $c$ vs. $s$ NP-hardness result for Unique-Games for all values of $c$ (except for $c$ very close to $0$). For these $c$, ours is the first hardness result showing that it helps to take the alphabet size larger than $2$. Our NP-hardness reductions are quasilinear-size and thus show nearly full exponential time is required, assuming the ETH.", acknowledgement = ack-nhfb, } @InProceedings{Barak:2012:HSS, author = "Boaz Barak and Fernando G. S. L. Brandao and Aram W. Harrow and Jonathan Kelner and David Steurer and Yuan Zhou", title = "Hypercontractivity, sum-of-squares proofs, and their applications", crossref = "ACM:2012:SPA", pages = "307--326", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214006", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the computational complexity of approximating the $2$-to-$q$ norm of linear operators (defined as $||A||_{2 \to q} = \max_{v \neq 0} ||A v||_q / ||v||_2$) for $q > 2$, as well as connections between this question and issues arising in quantum information theory and the study of Khot's Unique Games Conjecture (UGC). We show the following: For any constant even integer $q \geq 4$, a graph $G$ is a small-set expander if and only if the projector into the span of the top eigenvectors of $G$'s adjacency matrix has bounded $2 \to q$ norm. As a corollary, a good approximation to the $2 \to q$ norm will refute the Small-Set Expansion Conjecture --- a close variant of the UGC. We also show that such a good approximation can be obtained in $\exp(n^{2 / q})$ time, thus obtaining a different proof of the known subexponential algorithm for Small-Set-Expansion. Constant rounds of the ``Sum of Squares'' semidefinite programming hierarchy certify an upper bound on the $2 \to 4$ norm of the projector to low degree polynomials over the Boolean cube, as well certify the unsatisfiability of the ``noisy cube'' and ``short code'' based instances of Unique-Games considered by prior works. This improves on the previous upper bound of $\exp(\log^{O(1)} n)$ rounds (for the ``short code''), as well as separates the ``Sum of Squares'' / ``Lasserre'' hierarchy from weaker hierarchies that were known to require $\omega(1)$ rounds. We show reductions between computing the $2 \to 4$ norm and computing the injective tensor norm of a tensor, a problem with connections to quantum information theory. Three corollaries are: (i) the $2 \to 4$ norm is NP-hard to approximate to precision inverse-polynomial in the dimension, (ii) the $2 \to 4$ norm does not have a good approximation (in the sense above) unless 3-SAT can be solved in time $\exp(\sqrt n \poly \log (n))$, and (iii) known algorithms for the quantum separability problem imply a non-trivial additive approximation for the $2 \to 4$ norm.", acknowledgement = ack-nhfb, } @InProceedings{Efremenko:2012:IRL, author = "Klim Efremenko", title = "From irreducible representations to locally decodable codes", crossref = "ACM:2012:SPA", pages = "327--338", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214008", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A $q$-query Locally Decodable Code (LDC) is an error-correcting code that allows to read any particular symbol of the message by reading only $q$ symbols of the codeword even if the codeword is adversary corrupted. In this paper we present a new approach for the construction of LDCs. We show that if there exists an irreducible representation $(\rho, V)$ of $G$ and $q$ elements $g_1$, $g_2$, \ldots{}, $g_q$ in $G$ such that there exists a linear combination of matrices $\rho(g_i)$ that is of rank one, then we can construct a $q$-query Locally Decodable Code $C: V \to F^G$. We show the potential of this approach by constructing constant query LDCs of sub-exponential length matching the best known constructions.", acknowledgement = ack-nhfb, } @InProceedings{Guruswami:2012:FCF, author = "Venkatesan Guruswami and Chaoping Xing", title = "Folded codes from function field towers and improved optimal rate list decoding", crossref = "ACM:2012:SPA", pages = "339--350", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214009", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give a new construction of algebraic codes which are efficiently list decodable from a fraction $1 - R - \epsilon$ of adversarial errors where $R$ is the rate of the code, for any desired positive constant $\epsilon$. The worst-case list size output by the algorithm is $O(1 / \epsilon)$, matching the existential bound for random codes up to constant factors. Further, the alphabet size of the codes is a constant depending only on $\epsilon$ --- it can be made $\exp(\tilde{O}(1 / \epsilon^2))$ which is not much worse than the non-constructive $\exp(1 / \epsilon)$ bound of random codes. The code construction is Monte Carlo and has the claimed list decoding property with high probability. Once the code is (efficiently) sampled, the encoding/decoding algorithms are deterministic with a running time $O_{\epsilon} (N^c)$ for an absolute constant $c$, where $N$ is the code's block length. Our construction is based on a careful combination of a linear-algebraic approach to list decoding folded codes from towers of function fields, with a special form of subspace-evasive sets. Instantiating this with the explicit ``asymptotically good'' Garcia--Stichtenoth (GS for short) tower of function fields yields the above parameters. To illustrate the method in a simpler setting, we also present a construction based on Hermitian function fields, which offers similar guarantees with a list-size and alphabet size polylogarithmic in the block length $N$. In comparison, algebraic codes achieving the optimal trade-off between list decodability and rate based on folded Reed--Solomon codes have a decoding complexity of $N^{\Omega(1 / \epsilon)}$, an alphabet size of $N^{\Omega(1 / \epsilon 2)}$, and a list size of $O(1 / \epsilon^2)$ (even after combination with subspace-evasive sets). Thus we get an improvement over the previous best bounds in all three aspects simultaneously, and are quite close to the existential random coding bounds. Along the way, we shed light on how to use automorphisms of certain function fields to enable list decoding of the folded version of the associated algebraic-geometric codes.", acknowledgement = ack-nhfb, } @InProceedings{Dvir:2012:SES, author = "Zeev Dvir and Shachar Lovett", title = "Subspace evasive sets", crossref = "ACM:2012:SPA", pages = "351--358", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214010", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We construct explicit subspace-evasive sets. These are subsets of $F^n$ of size $|F|^{(1- \epsilon)n}$ whose intersection with any $k$-dimensional subspace is bounded by a constant $c(k, \epsilon)$. This problem was raised by Guruswami (CCC 2011) as it leads to optimal rate list-decodable codes of constant list size. The main technical ingredient is the construction of $k$ low-degree polynomials whose common set of zeros has small intersection with any $k$-dimensional subspace.", acknowledgement = ack-nhfb, } @InProceedings{Kaufman:2012:ETR, author = "Tali Kaufman and Alexander Lubotzky", title = "Edge transitive {Ramanujan} graphs and symmetric {LDPC} good codes", crossref = "ACM:2012:SPA", pages = "359--366", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214011", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present the first explicit construction of a binary symmetric code with constant rate and constant distance (i.e., good code). Moreover, the code is LDPC and its constraint space is generated by the orbit of one constant weight constraint under the group action. Our construction provides the first symmetric LDPC good codes. In particular, it solves the main open problem raised by Kaufman and Wigderson {8}.", acknowledgement = ack-nhfb, } @InProceedings{Makarychev:2012:AAS, author = "Konstantin Makarychev and Yury Makarychev and Aravindan Vijayaraghavan", title = "Approximation algorithms for semi-random partitioning problems", crossref = "ACM:2012:SPA", pages = "367--384", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214013", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In this paper, we propose and study a new semi-random model for graph partitioning problems. We believe that it captures many properties of real-world instances. The model is more flexible than the semi-random model of Feige and Kilian and planted random model of Bui, Chaudhuri, Leighton and Sipser. We develop a general framework for solving semi-random instances and apply it to several problems of interest. We present constant factor bi-criteria approximation algorithms for semi-random instances of the Balanced Cut, Multicut, Min Uncut, Sparsest Cut and Small Set Expansion problems. We also show how to almost recover the optimal solution if the instance satisfies an additional expanding condition. Our algorithms work in a wider range of parameters than most algorithms for previously studied random and semi-random models. Additionally, we study a new planted algebraic expander model and develop constant factor bi-criteria approximation algorithms for graph partitioning problems in this model.", acknowledgement = ack-nhfb, } @InProceedings{Sharathkumar:2012:NLT, author = "R. Sharathkumar and Pankaj K. Agarwal", title = "A near-linear time $\epsilon$-approximation algorithm for geometric bipartite matching", crossref = "ACM:2012:SPA", pages = "385--394", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214014", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "For point sets $A, B \subset \mathbb{R}^d$, $|A| = |B| = n$, and for a parameter $\epsilon > 0$, we present an algorithm that computes, in $O(n \poly(\log n, 1 / \epsilon))$ time, an $\epsilon$-approximate perfect matching of $A$ and $B$ with high probability; the previously best known algorithm takes $\Omega(n^{3/2})$ time. We approximate the L$_p$-norm using a distance function, $d(\cdot,\cdot)$ based on a randomly shifted quad-tree. The algorithm iteratively generates an approximate minimum-cost augmenting path under $d(\cdot,\cdot)$ in time proportional to the length of the path. We show that the total length of the augmenting paths generated by the algorithm is $O((n / \epsilon)\log n)$, implying that the running time of our algorithm is $O(n \poly(\log n, 1 / \epsilon))$.", acknowledgement = ack-nhfb, } @InProceedings{Abraham:2012:UPD, author = "Ittai Abraham and Ofer Neiman", title = "Using petal-decompositions to build a low stretch spanning tree", crossref = "ACM:2012:SPA", pages = "395--406", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214015", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove that any graph $G = (V, E)$ with $n$ points and $m$ edges has a spanning tree $T$ such that $\sum_{(u,v) \in E(G)} d_T(u,v) = O(m \log n \log \log n)$. Moreover such a tree can be found in time $O(m \log n \log \log n)$. Our result is obtained using a new petal-decomposition approach which guarantees that the radius of each cluster in the tree is at most $4$ times the radius of the induced subgraph of the cluster in the original graph.", acknowledgement = ack-nhfb, } @InProceedings{Brunsch:2012:ISA, author = "Tobias Brunsch and Heiko R{\"o}glin", title = "Improved smoothed analysis of multiobjective optimization", crossref = "ACM:2012:SPA", pages = "407--426", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214016", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present several new results about smoothed analysis of multiobjective optimization problems. Motivated by the discrepancy between worst-case analysis and practical experience, this line of research has gained a lot of attention in the last decade. We consider problems in which $d$ linear and one arbitrary objective function are to be optimized over a set $S \subseteq \{0, 1\}^n$ of feasible solutions. We improve the previously best known bound for the smoothed number of Pareto-optimal solutions to $O(n^{2d} \phi^d)$, where $\phi$ denotes the perturbation parameter. Additionally, we show that for any constant $c$ the $c$-th moment of the smoothed number of Pareto-optimal solutions is bounded by $O((n^{2d} \phi^d)^c)$. This improves the previously best known bounds significantly. Furthermore, we address the criticism that the perturbations in smoothed analysis destroy the zero-structure of problems by showing that the smoothed number of Pareto-optimal solutions remains polynomially bounded even for zero-preserving perturbations. This broadens the class of problems captured by smoothed analysis and it has consequences for non-linear objective functions. One corollary of our result is that the smoothed number of Pareto-optimal solutions is polynomially bounded for polynomial objective functions.", acknowledgement = ack-nhfb, } @InProceedings{Leonardi:2012:PFA, author = "Stefano Leonardi and Tim Roughgarden", title = "Prior-free auctions with ordered bidders", crossref = "ACM:2012:SPA", pages = "427--434", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214018", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Prior-free auctions are robust auctions that assume no distribution over bidders' valuations and provide worst-case (input-by-input) approximation guarantees. In contrast to previous work on this topic, we pursue good prior-free auctions with non-identical bidders. Prior-free auctions can approximate meaningful benchmarks for non-identical bidders only when ``sufficient qualitative information'' about the bidder asymmetry is publicly known. We consider digital goods auctions where there is a total ordering of the bidders that is known to the seller, where earlier bidders are in some sense thought to have higher valuations. We use the framework of Hartline and Roughgarden (STOC '08) to define an appropriate revenue benchmark: the maximum revenue that can be obtained from a bid vector using prices that are nonincreasing in the bidder ordering and bounded above by the second-highest bid. This monotone-price benchmark is always as large as the well-known fixed-price benchmark F$^{(2)}$, so designing prior-free auctions with good approximation guarantees is only harder. By design, an auction that approximates the monotone-price benchmark satisfies a very strong guarantee: it is, in particular, simultaneously near-optimal for essentially every Bayesian environment in which bidders' valuation distributions have nonincreasing monopoly prices, or in which the distribution of each bidder stochastically dominates that of the next. Of course, even if there is no distribution over bidders' valuations, such an auction still provides a quantifiable input-by-input performance guarantee. In this paper, we design a simple prior-free auction for digital goods with ordered bidders, the Random Price Restriction (RPR) auction. We prove that its expected revenue on every bid profile $b$ is $\Omega(M(b) / \log^*n)$, where $M$ denotes the monotone-price benchmark and $\log^*n$ denotes the number of times that the $\log_2$ operator can be applied to $n$ before the result drops below a fixed constant.", acknowledgement = ack-nhfb, } @InProceedings{Chawla:2012:LBB, author = "Shuchi Chawla and Nicole Immorlica and Brendan Lucier", title = "On the limits of black-box reductions in mechanism design", crossref = "ACM:2012:SPA", pages = "435--448", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214019", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider the problem of converting an arbitrary approximation algorithm for a single-parameter optimization problem into a computationally efficient truthful mechanism. We ask for reductions that are black-box, meaning that they require only oracle access to the given algorithm and in particular do not require explicit knowledge of the problem constraints. Such a reduction is known to be possible, for example, for the social welfare objective when the goal is to achieve Bayesian truthfulness and preserve social welfare in expectation. We show that a black-box reduction for the social welfare objective is not possible if the resulting mechanism is required to be truthful in expectation and to preserve the worst-case approximation ratio of the algorithm to within a subpolynomial factor. Further, we prove that for other objectives such as makespan, no black-box reduction is possible even if we only require Bayesian truthfulness and an average-case performance guarantee.", acknowledgement = ack-nhfb, } @InProceedings{Bei:2012:BFM, author = "Xiaohui Bei and Ning Chen and Nick Gravin and Pinyan Lu", title = "Budget feasible mechanism design: from prior-free to {Bayesian}", crossref = "ACM:2012:SPA", pages = "449--458", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214020", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Budget feasible mechanism design studies procurement combinatorial auctions in which the sellers have private costs to produce items, and the buyer (auctioneer) aims to maximize a social valuation function on subsets of items, under the budget constraint on the total payment. One of the most important questions in the field is ``which valuation domains admit truthful budget feasible mechanisms with 'small' approximations (compared to the social optimum)?'' Singer [35] showed that additive and submodular functions have a constant approximation mechanism. Recently, Dobzinski, Papadimitriou, and Singer [20] gave an $O(\log^2 n)$ approximation mechanism for subadditive functions; further, they remarked that: ``A fundamental question is whether, regardless of computational constraints, a constant-factor budget feasible mechanism exists for subadditive functions.'' In this paper, we address this question from two viewpoints: prior-free worst case analysis and Bayesian analysis, which are two standard approaches from computer science and economics, respectively. --- For the prior-free framework, we use a linear program (LP) that describes the fractional cover of the valuation function; the LP is also connected to the concept of approximate core in cooperative game theory. We provide a mechanism for subadditive functions whose approximation is $O(I)$, via the worst case integrality gap $I$ of this LP. This implies an $O(\log n)$-approximation for subadditive valuations, $O(1)$-approximation for XOS valuations, as well as for valuations having a constant integrality gap. XOS valuations are an important class of functions and lie between the submodular and the subadditive classes of valuations. We further give another polynomial time $O(\log n / (\log \log n))$ sub-logarithmic approximation mechanism for subadditive functions. Both of our mechanisms improve the best known approximation ratio $O(\log^2 n)$. --- For the Bayesian framework, we provide a constant approximation mechanism for all subadditive functions, using the above prior-free mechanism for XOS valuations as a subroutine. Our mechanism allows correlations in the distribution of private information and is universally truthful.", acknowledgement = ack-nhfb, } @InProceedings{Cai:2012:ACM, author = "Yang Cai and Constantinos Daskalakis and S. Matthew Weinberg", title = "An algorithmic characterization of multi-dimensional mechanisms", crossref = "ACM:2012:SPA", pages = "459--478", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214021", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show that every feasible, Bayesian, multi-item multi-bidder mechanism for independent, additive bidders can be implemented as a mechanism that: (a) allocates every item independently of the other items; (b) for the allocation of each item it uses a strict ordering of all bidders' types; and allocates the item using a distribution over hierarchical mechanisms that iron this ordering into a non-strict ordering, and give the item uniformly at random to the bidders whose reported types dominate all other reported types according to the non-strict ordering. Combined with cyclic-monotonicity our results provide a characterization of feasible, Bayesian Incentive Compatible mechanisms in this setting. Our characterization is enabled by a new, constructive proof of Border's theorem [Border 1991], and a new generalization of this theorem to independent (but not necessarily identically distributed) bidders, improving upon the results of [Border 2007, Che--Kim--Mierendorf 2011]. For a single item and independent bidders, we show that every feasible reduced form auction can be implemented as a distribution over hierarchical mechanisms that are consistent with the same strict ordering of all bidders' types, which every mechanism in the support of the distribution irons to a non-strict ordering. We also give a polynomial-time algorithm for determining feasibility of a reduced form auction, or providing a separation hyperplane from the set of feasible reduced forms. To complete the picture, we provide polynomial-time algorithms to find and exactly sample from a distribution over hierarchical mechanisms consistent with a given feasible reduced form. All these results generalize to multi-item reduced form auctions for independent, additive bidders. Finally, for multiple items, additive bidders with hard demand constraints, and arbitrary value correlation across items or bidders, we give a proper generalization of Border's Theorem, and characterize feasible reduced form auctions as multi-commodity flows in related multi-commodity flow instances. We also show that our generalization holds for a broader class of feasibility constraints, including the intersection of any two matroids. As a corollary of our results we compute revenue-optimal, Bayesian Incentive Compatible (BIC) mechanisms in multi-item multi-bidder settings, when each bidder has arbitrarily correlated values over the items and additive valuations over bundles of items, and the bidders are independent. Our mechanisms run in time polynomial in the total number of bidder types (and {not} type profiles). This running time is polynomial in the number of bidders, but potentially exponential in the number of items. We improve the running time to polynomial in both the number of items and the number of bidders by using recent structural results on optimal BIC auctions in item-symmetric settings [Daskalakis--Weinberg 2011].", acknowledgement = ack-nhfb, } @InProceedings{Gal:2012:TBC, author = "Anna G{\'a}l and Kristoffer Arnsfelt Hansen and Michal Kouck{\'y} and Pavel Pudl{\'a}k and Emanuele Viola", title = "Tight bounds on computing error-correcting codes by bounded-depth circuits with arbitrary gates", crossref = "ACM:2012:SPA", pages = "479--494", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214023", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We bound the minimum number $w$ of wires needed to compute any (asymptotically good) error-correcting code $C: \{0, 1\}^{\Omega(n)} \to \{0, 1\}^n$ with minimum distance $\Omega(n)$, using unbounded fan-in circuits of depth $d$ with arbitrary gates. Our main results are: (1) If $d = 2$ then $w = \Theta(n({\log n/ \log \log n})^2)$. (2) If $d = 3$ then $w = \Theta(n \lg \lg n)$. (3) If $d = 2k$ or $d = 2k + 1$ for some integer $k \geq 2$ then $w = \Theta(n \lambda_k (n))$, where $\lambda_1(n) = \lceil \log n \rceil$, $\lambda_{i + 1}(n) = \lambda_i*(n)$, and the $*$ operation gives how many times one has to iterate the function $\lambda_i$ to reach a value at most $1$ from the argument $n$. (4) If $d = \log * n$ then $w = O(n)$. For depth $d = 2$, our $\Omega(n (\log n/\log \log n)^2)$ lower bound gives the largest known lower bound for computing any linear map. Using a result by Ishai, Kushilevitz, Ostrovsky, and Sahai (2008), we also obtain similar bounds for computing pairwise-independent hash functions. Our lower bounds are based on a superconcentrator-like condition that the graphs of circuits computing good codes must satisfy. This condition is provably intermediate between superconcentrators and their weakenings considered before.", acknowledgement = ack-nhfb, } @InProceedings{Chan:2012:TBM, author = "Siu Man Chan and Aaron Potechin", title = "Tight bounds for monotone switching networks via {Fourier} analysis", crossref = "ACM:2012:SPA", pages = "495--504", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214024", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove tight size bounds on monotone switching networks for the $k$-clique problem, and for an explicit monotone problem by analyzing the generation problem with a pyramid structure of height $h$. This gives alternative proofs of the separations of $m$-NC from $m$-P and of $m$-NC$^i$ from $m$-NC$^{i + 1}$, different from Raz--McKenzie (Combinatorica '99). The enumerative-combinatorial and Fourier analytic techniques in this work are very different from a large body of work on circuit depth lower bounds, and may be of independent interest.", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2012:IIC, author = "Mark Braverman", title = "Interactive information complexity", crossref = "ACM:2012:SPA", pages = "505--524", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214025", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The primary goal of this paper is to define and study the interactive information complexity of functions. Let $f(x,y)$ be a function, and suppose Alice is given $x$ and Bob is given $y$. Informally, the interactive information complexity ${\rm IC}(f)$ of $f$ is the least amount of information Alice and Bob need to reveal to each other to compute $f$. Previously, information complexity has been defined with respect to a prior distribution on the input pairs $(x,y)$. Our first goal is to give a definition that is independent of the prior distribution. We show that several possible definitions are essentially equivalent. We establish some basic properties of the interactive information complexity ${\rm IC}(f)$. In particular, we show that ${\rm IC}(f)$ is equal to the amortized (randomized) communication complexity of $f$. We also show a direct sum theorem for ${\rm IC}(f)$ and give the first general connection between information complexity and (non-amortized) communication complexity. This connection implies that a non-trivial exchange of information is required when solving problems that have non-trivial communication complexity. We explore the information complexity of two specific problems --- Equality and Disjointness. We show that only a constant amount of information needs to be exchanged when solving Equality with no errors, while solving Disjointness with a constant error probability requires the parties to reveal a linear amount of information to each other.", acknowledgement = ack-nhfb, } @InProceedings{Sherstov:2012:MCC, author = "Alexander A. Sherstov", title = "The multiparty communication complexity of set disjointness", crossref = "ACM:2012:SPA", pages = "525--548", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214026", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the set disjointness problem in the number-on-the-forehead model of multiparty communication.\par (i) We prove that $k$-party set disjointness has communication complexity $\Omega(n/4^k)^{1/4}$ in the randomized and nondeterministic models and $\Omega(n/4^k)^{1/8}$ in the Merlin--Arthur model. These lower bounds are close to tight. Previous lower bounds (2007-2008) for $k \geq 3$ parties were weaker than $\Omega(n/2^{k^3})^{1 / (k + 1)}$ in all three models.\par (ii) We prove that solving $\ell$ instances of set disjointness requires $\ell \cdot \Omega(n/4^k)^{1/4}$ bits of communication, even to achieve correctness probability exponentially close to $1/2$. This gives the first direct-product result for multiparty set disjointness, solving an open problem due to Beame, Pitassi, Segerlind, and Wigderson (2005).\par (iii) We construct a read-once $\{\wedge, \vee\}$-circuit of depth 3 with exponentially small discrepancy for up to $k \approx (1/2)\log n$ parties. This result is optimal with respect to depth and solves an open problem due to Beame and Huynh-Ngoc (FOCS '09), who gave a depth-$6$ construction. Applications to circuit complexity are given.", acknowledgement = ack-nhfb, } @InProceedings{Cheung:2012:FMR, author = "Ho Yee Cheung and Tsz Chiu Kwok and Lap Chi Lau", title = "Fast matrix rank algorithms and applications", crossref = "ACM:2012:SPA", pages = "549--562", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214028", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider the problem of computing the rank of an $m \times n$ matrix $A$ over a field. We present a randomized algorithm to find a set of $r = \rank(A)$ linearly independent columns in $O(|A| + r^w)$ field operations, where $|A|$ denotes the number of nonzero entries in $A$ and $w < 2.38$ is the matrix multiplication exponent. Previously the best known algorithm to find a set of $r$ linearly independent columns is by Gaussian elimination, with running time $O(m n r^w)$. Our algorithm is faster when $r < \max\{m, n\}$, for instance when the matrix is rectangular. We also consider the problem of computing the rank of a matrix dynamically, supporting the operations of rank one updates and additions and deletions of rows and columns. We present an algorithm that updates the rank in $O(m n)$ field operations. We show that these algorithms can be used to obtain faster algorithms for various problems in numerical linear algebra, combinatorial optimization and dynamic data structure.", acknowledgement = ack-nhfb, } @InProceedings{Hassanieh:2012:NOS, author = "Haitham Hassanieh and Piotr Indyk and Dina Katabi and Eric Price", title = "Nearly optimal sparse {Fourier} transform", crossref = "ACM:2012:SPA", pages = "563--578", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214029", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider the problem of computing the $k$-sparse approximation to the discrete Fourier transform of an $n$-dimensional signal. We show:\par (1) An $O(k \log n)$-time randomized algorithm for the case where the input signal has at most $k$ non-zero Fourier coefficients, and\par (2) An $O(k \log n \log (n/k))$-time randomized algorithm for general input signals.\par Both algorithms achieve $o(n \log n)$ time, and thus improve over the Fast Fourier Transform, for any $k = o(n)$. They are the first known algorithms that satisfy this property. Also, if one assumes that the Fast Fourier Transform is optimal, the algorithm for the exactly $k$-sparse case is optimal for any $k = n^{\Omega(1)}$.\par We complement our algorithmic results by showing that any algorithm for computing the sparse Fourier transform of a general signal must use at least $\Omega(k \log (n / k) / \log \log n)$ signal samples, even if it is allowed to perform adaptive sampling.", acknowledgement = ack-nhfb, } @InProceedings{Etessami:2012:PTA, author = "Kousha Etessami and Alistair Stewart and Mihalis Yannakakis", title = "Polynomial time algorithms for multi-type branching processes and stochastic context-free grammars", crossref = "ACM:2012:SPA", pages = "579--588", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214030", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show that one can approximate the least fixed point solution for a multivariate system of monotone probabilistic polynomial equations in time polynomial in both the encoding size of the system of equations and in $\log (1 / \epsilon)$, where $\epsilon > 0$ is the desired additive error bound of the solution. (The model of computation is the standard Turing machine model.) We use this result to resolve several open problems regarding the computational complexity of computing key quantities associated with some classic and heavily studied stochastic processes, including multi-type branching processes and stochastic context-free grammars.", acknowledgement = ack-nhfb, } @InProceedings{Adamaszek:2012:OOB, author = "Anna Adamaszek and Artur Czumaj and Matthias Englert and Harald R{\"a}cke", title = "Optimal online buffer scheduling for block devices", crossref = "ACM:2012:SPA", pages = "589--598", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214031", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We introduce a buffer scheduling problem for block operation devices in an online setting. We consider a stream of items of different types to be processed by a block device. The block device can process all items of the same type in a single step. To improve the performance of the system a buffer of size $k$ is used to store items in order to reduce the number of operations required. Whenever the buffer becomes full a buffer scheduling strategy has to select one type and then a block operation on all elements with this type that are currently in the buffer is performed. The goal is to design a scheduling strategy that minimizes the number of block operations required. In this paper we consider the online version of this problem, where the buffer scheduling strategy must make decisions without knowing the future items that appear in the input stream. Our main result is the design of an $O(\log \log k)$-competitive online randomized buffer scheduling strategy. The bound is asymptotically tight. As a byproduct of our LP-based techniques, we obtain a randomized offline algorithm that approximates the optimal number of block operations to within a constant factor.", acknowledgement = ack-nhfb, } @InProceedings{Agrawal:2012:JHC, author = "Manindra Agrawal and Chandan Saha and Ramprasad Saptharishi and Nitin Saxena", title = "{Jacobian} hits circuits: hitting-sets, lower bounds for depth-{D} occur-$k$ formulas \& depth-$3$ transcendence degree-$k$ circuits", crossref = "ACM:2012:SPA", pages = "599--614", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214033", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present a single common tool to strictly subsume all known cases of polynomial time blackbox polynomial identity testing (PIT), that have been hitherto solved using diverse tools and techniques, over fields of zero or large characteristic. In particular, we show that polynomial time hitting-set generators for identity testing of the two seemingly different and well studied models --- depth-3 circuits with bounded top fanin, and constant-depth constant-read multilinear formulas --- can be constructed using one common algebraic-geometry theme: Jacobian captures algebraic independence. By exploiting the Jacobian, we design the first efficient hitting-set generators for broad generalizations of the above-mentioned models, namely:\par (1) depth-3 $(\Omega \Pi \Omega)$ circuits with constant transcendence degree of the polynomials computed by the product gates (no bounded top fanin restriction), and\par (2) constant-depth constant-{\em occur\/} formulas (no multilinear restriction).\par Constant-{\rm occur} of a variable, as we define it, is a much more general concept than constant-read. Also, earlier work on the latter model assumed that the formula is multilinear. Thus, our work goes further beyond the related results obtained by Saxena \& Seshadhri (STOC 2011), Saraf \& Volkovich (STOC 2011), Anderson et al. (CCC 2011), Beecken et al. (ICALP 2011) and Grenet et al. (FSTTCS 2011), and brings them under one unifying technique.\par In addition, using the same Jacobian based approach, we prove exponential lower bounds for the immanant (which includes permanent and determinant) on the same depth-$3$ and depth-$4$ models for which we give efficient PIT algorithms. Our results reinforce the intimate connection between identity testing and lower bounds by exhibiting a concrete mathematical tool --- the Jacobian --- that is equally effective in solving both the problems on certain interesting and previously well-investigated (but not well understood) models of computation.", acknowledgement = ack-nhfb, } @InProceedings{Dvir:2012:SMB, author = "Zeev Dvir and Guillaume Malod and Sylvain Perifel and Amir Yehudayoff", title = "Separating multilinear branching programs and formulas", crossref = "ACM:2012:SPA", pages = "615--624", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214034", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "This work deals with the power of linear algebra in the context of multilinear computation. By linear algebra we mean algebraic branching programs (ABPs) which are known to be computationally equivalent to two basic tools in linear algebra: iterated matrix multiplication and the determinant. We compare the computational power of multilinear ABPs to that of multilinear arithmetic formulas, and prove a tight super-polynomial separation between the two models. Specifically, we describe an explicit $n$-variate polynomial $F$ that is computed by a linear-size multilinear ABP but every multilinear formula computing $F$ must be of size n$^{ \Omega(\log n)}$.", acknowledgement = ack-nhfb, } @InProceedings{Gupta:2012:RDM, author = "Ankit Gupta and Neeraj Kayal and Satya Lokam", title = "Reconstruction of depth-$4$ multilinear circuits with top fan-in $2$", crossref = "ACM:2012:SPA", pages = "625--642", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214035", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present a randomized algorithm for reconstructing multilinear $\Sigma \Pi \Sigma \Pi (2)$ circuits, i.e., multilinear depth-$4$ circuits with fan-in $2$ at the top $+$ gate. The algorithm is given blackbox access to a polynomial $f \in F[x_1,\ldots{},x_n]$ computable by a multilinear $\Sigma \Pi \Sigma \Pi (2)$ circuit of size $s$ and outputs an equivalent multilinear $\Sigma \Pi \Sigma \Pi (2)$ circuit, runs in time $\poly(n,s)$, and works over any field $F$. This is the first reconstruction result for any model of depth-$4$ arithmetic circuits. Prior to our work, reconstruction results for bounded depth circuits were known only for depth-$2$ arithmetic circuits (Klivans \& Spielman, STOC 2001), $\Sigma \Pi \Sigma (2)$ circuits (depth-$3$ arithmetic circuits with top fan-in $2$) (Shpilka, STOC 2007), and $\Sigma \Pi \Sigma (k)$ with $k = O(1)$ (Karnin \& Shpilka, CCC 2009). Moreover, the running times of these algorithms have a polynomial dependence on $4|F|$ and hence do not work for infinite fields such as $Q$. Our techniques are quite different from the previous ones for depth-$3$ reconstruction and rely on a polynomial operator introduced by Karnin et al. (STOC 2010) and Saraf \& Volkovich (STOC 2011) for devising blackbox identity tests for multilinear $\Sigma \Pi \Sigma \Pi (k)$ circuits. Some other ingredients of our algorithm include the classical multivariate blackbox factoring algorithm by Kaltofen \& Trager (FOCS 1988) and an average-case algorithm for reconstructing $\Sigma \Pi \Sigma (2)$ circuits by Kayal.", acknowledgement = ack-nhfb, } @InProceedings{Kayal:2012:APP, author = "Neeraj Kayal", title = "Affine projections of polynomials: extended abstract", crossref = "ACM:2012:SPA", pages = "643--662", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214036", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "An $m$-variate polynomial $f$ is said to be an affine projection of some $n$-variate polynomial $g$ if there exists an $n m$ matrix $A$ and an $n$-dimensional vector $b$ such that $f(x) = g(Ax + b)$. In other words, if $f$ can be obtained by replacing each variable of $g$ by an affine combination of the variables occurring in $f$, then it is said to be an affine projection of $g$. Some well known problems (such as the determinant versus permanent and matrix multiplication for example) are instances of this problem. Given $f$ and $g$ can we determine whether $f$ is an affine projection of $g$? The intention of this paper is to understand the complexity of the corresponding computational problem: given polynomials $f$ and $g$ find $A$ and $b$ such that $f = g(A x + b)$, if such an $(A b)$ exists. We first show that this is an NP-hard problem. We then focus our attention on instances where $g$ is a member of some fixed, well known family of polynomials so that the input consists only of the polynomial $f(x)$ having $m$ variables and degree $d$. We consider the situation where $f(x)$ is given to us as a blackbox (i.e. for any point $aFm$ we can query the blackbox and obtain $f(a)$ in one step) and devise randomized algorithms with running time $\poly(m n d)$ in the following special cases. Firstly where $g$ is the Permanent (respectively the Determinant) of an $n \times n$ matrix and $A$ is of rank $n^2$. Secondly where $g$ is the sum of powers polynomial (respectively the sum of products polynomial), and $A$ is a random matrix of the appropriate dimensions (also $d$ should not be too small).", acknowledgement = ack-nhfb, } @InProceedings{Bartal:2012:TSP, author = "Yair Bartal and Lee-Ad Gottlieb and Robert Krauthgamer", title = "The traveling salesman problem: low-dimensionality implies a polynomial time approximation scheme", crossref = "ACM:2012:SPA", pages = "663--672", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214038", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Traveling Salesman Problem (TSP) is among the most famous NP-hard optimization problems. We design for this problem a randomized polynomial-time algorithm that computes a $(1 + \mu)$-approximation to the optimal tour, for any fixed $\mu > 0$, in TSP instances that form an arbitrary metric space with bounded intrinsic dimension. The celebrated results of Arora [Aro98] and Mitchell [Mit99] prove that the above result holds in the special case of TSP in a fixed-dimensional Euclidean space. Thus, our algorithm demonstrates that the algorithmic tractability of metric TSP depends on the dimensionality of the space and not on its specific geometry. This result resolves a problem that has been open since the quasi-polynomial time algorithm of Talwar [Tal04].", acknowledgement = ack-nhfb, } @InProceedings{Chuzhoy:2012:VSS, author = "Julia Chuzhoy", title = "On vertex sparsifiers with {Steiner} nodes", crossref = "ACM:2012:SPA", pages = "673--688", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214039", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Given an undirected graph $G = (V, E)$ with edge capacities $c_e \geq 1$ for $e \in E$ and a subset $T$ of $k$ vertices called terminals, we say that a graph $H$ is a quality-$q$ cut sparsifier for $G$ iff $T \subseteq V(H)$, and for any partition $(A, B)$ of $T$, the values of the minimum cuts separating $A$ and $B$ in graphs $G$ and $H$ are within a factor $q$ from each other. We say that $H$ is a quality-$q$ flow sparsifier for $G$ iff $T \subseteq V(H)$, and for any set $D$ of demands over the terminals, the values of the minimum edge congestion incurred by fractionally routing the demands in $D$ in graphs $G$ and $H$ are within a factor $q$ from each other.\par So far vertex sparsifiers have been studied in a restricted setting where the sparsifier $H$ is not allowed to contain any non-terminal vertices, that is $V(H) = {\cal T}$. For this setting, efficient algorithms are known for constructing quality-$O(\log k/\log \log k)$ cut and flow vertex sparsifiers, as well as a lower bound of $\tilde{\Omega}(\sqrt{\log k})$ on the quality of any flow or cut sparsifier.\par We study flow and cut sparsifiers in the more general setting where Steiner vertices are allowed, that is, we no longer require that $V(H) = {\cal T}$. We show algorithms to construct constant-quality cut sparsifiers of size $O(C^3)$ in time $\poly(n) \cdot 2^C$, and constant-quality flow sparsifiers of size $C^{O(\log \log C)}$ in time $n^{O(\log C)} \cdot 2^C$, where $C$ is the total capacity of the edges incident on the terminals.", acknowledgement = ack-nhfb, } @InProceedings{Chalermsook:2012:AAH, author = "Parinya Chalermsook and Julia Chuzhoy and Alina Ene and Shi Li", title = "Approximation algorithms and hardness of integral concurrent flow", crossref = "ACM:2012:SPA", pages = "689--708", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214040", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study an integral counterpart of the classical Maximum Concurrent Flow problem, that we call Integral Concurrent Flow (ICF). In the basic version of this problem (basic-ICF), we are given an undirected $n$-vertex graph $G$ with edge capacities $c(e)$, a subset $T$ of vertices called terminals, and a demand $D(t,t')$ for every pair $(t,t')$ of the terminals. The goal is to find a maximum value $\lambda$, and a collection $P$ of paths, such that every pair $(t,t')$ of terminals is connected by $\lfloor \lambda \cdot D(t,t') \rfloor$ paths in $P$, and the number of paths containing any edge $e$ is at most $c(e)$. We show an algorithm that achieves a $\poly \log n$-approximation for basic-ICF, while violating the edge capacities by only a constant factor. We complement this result by proving that no efficient algorithm can achieve a factor $\alpha$-approximation with congestion $c$ for any values $\alpha$, $c$ satisfying $\alpha \cdot c = O(\log \log n / \log \log \log n)$, unless NP $\subseteq$ ZPTIME(n$^{\poly \log n}$). We then turn to study the more general group version of the problem (group = ICF), in which we are given a collection $(S_1, T_1), \ldots{}, (S_k, T_k)$ of pairs of vertex subsets, and for each $1 \leq $i$ \leq k$, a demand D$_i$ is specified. The goal is to find a maximum value $\lambda$ and a collection $P$ of paths, such that for each $i$, at least $\lfloor \lambda \cdot D_i \rfloor$ paths connect the vertices of $S_i$ to the vertices of $T_i$, while respecting the edge capacities. We show that for any $1 \leq $c$ \leq O(\log \log n)$, no efficient algorithm can achieve a factor $O(n^{1/(2 2c + 3)})$-approximation with congestion $c$ for the problem, unless NP $\subseteq$ DTIME($n^{O(\log \log n)}$). On the other hand, we show an efficient randomized algorithm that finds a $\poly \log n$-approximate solution with a constant congestion, if we are guaranteed that the optimal solution contains at least $D \geq $k$ \poly \log n$ paths connecting every pair (S$_i$, T$_i$).", acknowledgement = ack-nhfb, } @InProceedings{Daskalakis:2012:LPB, author = "Constantinos Daskalakis and Ilias Diakonikolas and Rocco A. Servedio", title = "Learning {Poisson} binomial distributions", crossref = "ACM:2012:SPA", pages = "709--728", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214042", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider a basic problem in unsupervised learning: learning an unknown Poisson Binomial Distribution. A Poisson Binomial Distribution (PBD) over $\{0, 1,\ldots{}, n\}$ is the distribution of a sum of $n$ independent Bernoulli random variables which may have arbitrary, potentially non-equal, expectations. These distributions were first studied by S. Poisson in 1837 and are a natural $n$-parameter generalization of the familiar Binomial Distribution. Surprisingly, prior to our work this basic learning problem was poorly understood, and known results for it were far from optimal. We essentially settle the complexity of the learning problem for this basic class of distributions. As our main result we give a highly efficient algorithm which learns to $\epsilon$-accuracy using $O(1 / \epsilon^3)$ samples independent of $n$. The running time of the algorithm is quasilinear in the size of its input data, i.e. $\tilde{O}(\log (n)/ \epsilon^3)$ bit-operations (observe that each draw from the distribution is a $\log(n)$-bit string). This is nearly optimal since any algorithm must use $\Omega(1 / \epsilon^2)$ samples. We also give positive and negative results for some extensions of this learning problem.", acknowledgement = ack-nhfb, } @InProceedings{De:2012:NOS, author = "Anindya De and Ilias Diakonikolas and Vitaly Feldman and Rocco A. Servedio", title = "Nearly optimal solutions for the chow parameters problem and low-weight approximation of halfspaces", crossref = "ACM:2012:SPA", pages = "729--746", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214043", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Chow parameters of a Boolean function $f: \{-1, 1\}^n \to \{-1, 1\}$ are its $n + 1$ degree-$0$ and degree-$1$ Fourier coefficients. It has been known since 1961 [Cho61, Tan61] that the (exact values of the) Chow parameters of any linear threshold function $f$ uniquely specify $f$ within the space of all Boolean functions, but until recently [OS11] nothing was known about efficient algorithms for reconstructing $f$ (exactly or approximately) from exact or approximate values of its Chow parameters. We refer to this reconstruction problem as the Chow Parameters Problem. Our main result is a new algorithm for the Chow Parameters Problem which, given (sufficiently accurate approximations to) the Chow parameters of any linear threshold function f, runs in time $\tilde{O}(n^2) o(1/ \epsilon)^{O(\log 2 (1 / \epsilon))}$ and with high probability outputs a representation of an LTF $f'$ that is $\epsilon$-close to $f$. The only previous algorithm [OS11] had running time $\poly(n) \cdot 2^{2 \tilde{O}(1 / \epsilon 2)}$. As a byproduct of our approach, we show that for any linear threshold function $f$ over ${-1,1}^n$, there is a linear threshold function $f'$ which is $\epsilon$-close to $f$ and has all weights that are integers at most $\sqrt n o(1 / \epsilon)^{O(\log 2 (1 / \epsilon))}$. This significantly improves the best previous result of [Serv09] which gave a $\poly(n) o 2^{O(1 / \epsilon 2/3)}$ weight bound, and is close to the known lower bound of $\max\{\sqrt n, (1 / \epsilon)^{\Omega(\log \log (1 / \epsilon))}\}$ [Gol06,Serv07]. Our techniques also yield improved algorithms for related problems in learning theory. In addition to being significantly stronger than previous work, our results are obtained using conceptually simpler proofs. The two main ingredients underlying our results are (1) a new structural result showing that for $f$ any linear threshold function and g any bounded function, if the Chow parameters of $f$ are close to the Chow parameters of $g$ then $f$ is close to g; (2) a new boosting-like algorithm that given approximations to the Chow parameters of a linear threshold function outputs a bounded function whose Chow parameters are close to those of $f$.", } @InProceedings{Sherstov:2012:MPR, author = "Alexander A. Sherstov", title = "Making polynomials robust to noise", crossref = "ACM:2012:SPA", pages = "747--758", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214044", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A basic question in any computational model is how to reliably compute a given function when the inputs or intermediate computations are subject to noise at a constant rate. Ideally, one would like to use at most a constant factor more resources compared to the noise-free case. This question has been studied for decision trees, circuits, automata, data structures, broadcast networks, communication protocols, and other models. Buhrman et al. (2003) posed the noisy computation problem for real polynomials. We give a complete solution to this problem. For any polynomial $p : \{0, 1\}^n \to [-1, 1]$, we construct a polynomial $p_{\rm robust}: R^n \to R$ of degree $O(\deg p + \log(1 / \epsilon))$ that $\epsilon$-approximates $p$ and is robust to noise in the inputs: $|p(x) - p_{\rm robust} (x + \delta)| < \epsilon$ for all $x \in \{0, 1\}^n$ and all $\delta \in [-1/3, 1/3]^n$. This result is optimal with respect to all parameters. We construct $p_{\rm robust}$ explicitly for each $p$. Previously, it was open to give such a construction even for $p = x_1 \oplus x_2 \oplus \ldots{} \oplus x_n$ (Buhrman et al., 2003). The proof contributes a technique of independent interest, which allows one to force partial cancellation of error terms in a polynomial.", acknowledgement = ack-nhfb, } @InProceedings{Goyal:2012:CCN, author = "Sanjeev Goyal and Michael Kearns", title = "Competitive contagion in networks", crossref = "ACM:2012:SPA", pages = "759--774", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214046", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We develop a game-theoretic framework for the study of competition between firms who have budgets to ``seed'' the initial adoption of their products by consumers located in a social network. The payoffs to the firms are the eventual number of adoptions of their product through a competitive stochastic diffusion process in the network. This framework yields a rich class of competitive strategies, which depend in subtle ways on the stochastic dynamics of adoption, the relative budgets of the players, and the underlying structure of the social network. We identify a general property of the adoption dynamics --- namely, decreasing returns to local adoption --- for which the inefficiency of resource use at equilibrium (the Price of Anarchy) is uniformly bounded above, across all networks. We also show that if this property is violated the Price of Anarchy can be unbounded, thus yielding sharp threshold behavior for a broad class of dynamics. We also introduce a new notion, the Budget Multiplier, that measures the extent that imbalances in player budgets can be amplified at equilibrium. We again identify a general property of the adoption dynamics --- namely, proportional local adoption between competitors --- for which the (pure strategy) Budget Multiplier is uniformly bounded above, across all networks. We show that a violation of this property can lead to unbounded Budget Multiplier, again yielding sharp threshold behavior for a broad class of dynamics.", acknowledgement = ack-nhfb, } @InProceedings{Cebrian:2012:FRB, author = "Manuel Cebrian and Lorenzo Coviello and Andrea Vattani and Panagiotis Voulgaris", title = "Finding red balloons with split contracts: robustness to individuals' selfishness", crossref = "ACM:2012:SPA", pages = "775--788", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214047", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The present work deals with the problem of information acquisition in a strategic networked environment. To study this problem, Kleinberg and Raghavan (FOCS 2005) introduced the model of {\em query incentive networks}, where the root of a binomial branching process wishes to retrieve an information --- known by each node independently with probability 1/n --- by investing as little as possible. The authors considered {\em fixed-payment contracts\/} in which every node strategically chooses an amount to offer its children paid upon information retrieval to convince them to seek the information in their subtrees. Kleinberg and Raghavan discovered that the investment needed at the root exhibits an unexpected threshold behavior that depends on the branching parameter b. For b > 2, the investment is linear in the expected distance to the closest information (logarithmic in $n$, the rarity of the information), while, for $1 < b < 2$, it becomes exponential in the same distance (i.e., polynomial in $n$). Arcaute et al. (EC 2007) later observed the same threshold behavior for arbitrary Galton--Watson branching processes.\par The DARPA Network Challenge --- retrieving the locations of ten balloons placed at undisclosed positions in the US --- has recently brought practical attention to the problems of social mobilization and information acquisition in a networked environment. The MIT Media Laboratory team won the challenge by acting as the root of a query incentive network that unfolded all over the world. However, rather than adopting a {\em fixed-payment strategy}, the team implemented a different incentive scheme based on {\em $1/2$-split contracts}. Under such incentive scheme, a node $u$ who does not possess the information can recruit a friend $v$ through a contract stipulating that if the information is found in the subtree rooted at $v$, then $v$ has to give half of her own reward back to $u$.\par Motivated by its empirical success, we present a comprehensive theoretical study of this scheme in the game theoretical setting of query incentive networks. Our main result is that split contracts are robust --- as opposed to fixed-payment contracts --- to nodes' selfishness. Surprisingly, when nodes determine the splits to offer their children based on the contracts received from their recruiters, the threshold behavior observed in the previous work vanishes, and an investment linear in the expected distance to the closest information is sufficient to retrieve the information in {\em any arbitrary\/} Galton--Watson process with $b > 1$. Finally, while previous analyses considered the parameters of the branching process as constants, we are able to characterize the rate of the investment in terms of the branching process and the desired probability of success. This allows us to show improvements even in other special cases.", acknowledgement = ack-nhfb, } @InProceedings{Brandt:2012:AOD, author = "Christina Brandt and Nicole Immorlica and Gautam Kamath and Robert Kleinberg", title = "An analysis of one-dimensional {Schelling} segregation", crossref = "ACM:2012:SPA", pages = "789--804", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214048", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We analyze the Schelling model of segregation in which a society of $n$ individuals live in a ring. Each individual is one of two races and is only satisfied with his location so long as at least half his 2w nearest neighbors are of the same race as him. In the dynamics, randomly-chosen unhappy individuals successively swap locations. We consider the average size of monochromatic neighborhoods in the final stable state. Our analysis is the first rigorous analysis of the Schelling dynamics. We note that, in contrast to prior approximate analyses, the final state is nearly integrated: the average size of monochromatic neighborhoods is independent of $n$ and polynomial in w.", acknowledgement = ack-nhfb, } @InProceedings{Applebaum:2012:PGL, author = "Benny Applebaum", title = "Pseudorandom generators with long stretch and low locality from random local one-way functions", crossref = "ACM:2012:SPA", pages = "805--816", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214050", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We continue the study of {\em locally-computable\/} pseudorandom generators (PRG) $G : \{0, 1\}^n \to \{0, 1\}^m$ that each of their outputs depend on a small number of $d$ input bits. While it is known that such generators are likely to exist for the case of small sub-linear stretch $m = n + n^{1 - \delta}$, it is less clear whether achieving larger stretch such as $m = n + \Omega(n)$, or even $m = n^{1 + \delta}$ is possible. The existence of such PRGs, which was posed as an open question in previous works, has recently gained an additional motivation due to several interesting applications.\par We make progress towards resolving this question by obtaining several local constructions based on the one-wayness of ``random'' local functions --- a variant of an assumption made by Goldreich (ECCC 2000). Specifically, we construct collections of PRGs with the following parameters:\par 1. Linear stretch $m = n + \Omega(n)$ and constant locality $d = O(1)$.\par 2. Polynomial stretch $m = n^{1 + \delta}$ and {\em any\/} (arbitrarily slowly growing) super-constant locality $d = \omega(1)$, e.g., $\log^*n$.\par 3. Polynomial stretch $m = n^{1 + \delta}$, constant locality $d = O(1)$, and inverse polynomial distinguishing advantage (as opposed to the standard case of $n^{-\omega(1)}$).\par As an additional contribution, we show that our constructions give rise to strong inapproximability results for the densest-subgraph problem in $d$-uniform hypergraphs for constant $d$. This allows us to improve the previous bounds of Feige (STOC 2002) and Khot (FOCS 2004) from constant inapproximability factor to $n^\epsilon$-inapproximability, at the expense of relying on stronger assumptions.", acknowledgement = ack-nhfb, } @InProceedings{Vadhan:2012:CPS, author = "Salil Vadhan and Colin Jia Zheng", title = "Characterizing pseudoentropy and simplifying pseudorandom generator constructions", crossref = "ACM:2012:SPA", pages = "817--836", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214051", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We provide a characterization of pseudoentropy in terms of hardness of sampling: Let $(X, B)$ be jointly distributed random variables such that $B$ takes values in a polynomial-sized set. We show that $B$ is computationally indistinguishable from a random variable of higher Shannon entropy given $X$ if and only if there is no probabilistic polynomial-time $S$ such that $(X, S(X))$ has small KL divergence from $(X, B)$. This can be viewed as an analogue of the Impagliazzo Hardcore Theorem (FOCS '95) for Shannon entropy (rather than min-entropy).\par Using this characterization, we show that if $f$ is a one-way function, then $(f(U_n), U_n)$ has ``next-bit pseudoentropy'' at least $n + \log n$, establishing a conjecture of Haitner, Reingold, and Vadhan (STOC '10). Plugging this into the construction of Haitner et al., this yields a simpler construction of pseudorandom generators from one-way functions. In particular, the construction only performs hashing once, and only needs the hash functions that are randomness extractors (e.g. universal hash functions) rather than needing them to support ``local list-decoding'' (as in the Goldreich--Levin hardcore predicate, STOC '89).\par With an additional idea, we also show how to improve the seed length of the pseudorandom generator to $\tilde{O}(n^3)$, compared to $\tilde{O}(n^4)$ in the construction of Haitner et al.", acknowledgement = ack-nhfb, } @InProceedings{Li:2012:DEN, author = "Xin Li", title = "Design extractors, non-malleable condensers and privacy amplification", crossref = "ACM:2012:SPA", pages = "837--854", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214052", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We introduce a new combinatorial object, called a design extractor, that has both the properties of a design and an extractor. We give efficient constructions of such objects and show that they can be used in several applications.\par 1. {\bf Improving the output length of known non-malleable extractors.} Non-malleable extractors were introduced in [DW09] to study the problem of privacy amplification with an active adversary. Currently, only two explicit constructions are known [DLWZ11, CRS11]. Both constructions work for $n$ bit sources with min-entropy $k > n / 2$. However, in both constructions the output length is smaller than the seed length, while the probabilistic method shows that to achieve error $\epsilon$, one can use $O(\log n + \log (1 / \epsilon))$ bits to extract up to $k/2$ output bits. In this paper, we use our design extractor to give an explicit non-malleable extractor for min-entropy $k > n / 2$, that has seed length $O(\log n + \log (1 / \epsilon))$ and output length $\Omega(k)$.\par 2. {\bf Non-malleable condensers.} We introduce and define the notion of a {\em non-malleable condenser}. A non-malleable condenser is a generalization and relaxation of a non-malleable extractor. We show that similar as extractors and condensers, non-malleable condensers can be used to construct non-malleable extractors. We then show that our design extractor already gives a non-malleable condenser for min-entropy $k > n / 2$, with error $\epsilon$ and seed length $O(\log (1 / \epsilon))$.\par 3. {\bf A new optimal protocol for privacy amplification.} More surprisingly, we show that non-malleable condensers themselves give optimal privacy amplification protocols with an active adversary. In fact, the non-malleable condensers used in these protocols are much weaker compared to non-malleable extractors, in the sense that the entropy rate of the condenser's output does not need to increase at all. This suggests that one promising next step to achieve better privacy amplification protocols may be to construct non-malleable condensers for smaller min-entropy. As a by-product, we also obtain a new explicit $2$-round privacy amplification protocol with optimal entropy loss and optimal communication complexity for min-entropy $k > n / 2$, without using non-malleable extractors.", acknowledgement = ack-nhfb, } @InProceedings{Chuzhoy:2012:RUG, author = "Julia Chuzhoy", title = "Routing in undirected graphs with constant congestion", crossref = "ACM:2012:SPA", pages = "855--874", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214054", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Given an undirected graph $G = (V,E)$, a collection $(s_1, t_1)$, \ldots{}, $(s_k, t_k)$ of $k$ demand pairs, and an integer $c$, the goal in the Edge Disjoint Paths with Congestion problem is to connect maximum possible number of the demand pairs by paths, so that the maximum load on any edge (called edge congestion) does not exceed $c$. We show an efficient randomized algorithm that routes $\Omega({\rm OPT} / \poly \log k)$ demand pairs with congestion at most $14$, where OPT is the maximum number of pairs that can be simultaneously routed on edge-disjoint paths. The best previous algorithm that routed $\Omega({\rm OPT} / \poly \log n)$ pairs required congestion $\poly(\log \log n)$, and for the setting where the maximum allowed congestion is bounded by a constant $c$, the best previous algorithms could only guarantee the routing of OPT / $n^{O(1/c)}$ pairs. We also introduce a new type of vertex sparsifiers that we call integral flow sparsifiers, which approximately preserve both fractional and integral routings, and show an algorithm to construct such sparsifiers.", acknowledgement = ack-nhfb, } @InProceedings{An:2012:ICA, author = "Hyung-Chan An and Robert Kleinberg and David B. Shmoys", title = "Improving {Christofides}' algorithm for the $s$-$t$ path {TSP}", crossref = "ACM:2012:SPA", pages = "875--886", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214055", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present a deterministic $(1 + \sqrt 5/2)$-approximation algorithm for the $s$-$t$ path TSP for an arbitrary metric. Given a symmetric metric cost on $n$ vertices including two prespecified endpoints, the problem is to find a shortest Hamiltonian path between the two endpoints; Hoogeveen showed that the natural variant of Christofides' algorithm is a $5/3$-approximation algorithm for this problem, and this asymptotically tight bound in fact had been the best approximation ratio known until now. We modify this algorithm so that it chooses the initial spanning tree based on an optimal solution to the Held--Karp relaxation rather than a minimum spanning tree; we prove this simple but crucial modification leads to an improved approximation ratio, surpassing the 20-year-old barrier set by the natural Christofides' algorithm variant. Our algorithm also proves an upper bound of $1 + \sqrt 5/2$ on the integrality gap of the path-variant Held--Karp relaxation. The techniques devised in this paper can be applied to other optimization problems as well: these applications include improved approximation algorithms and improved LP integrality gap upper bounds for the prize-collecting $s$-$t$ path problem and the unit-weight graphical metric $s$-$t$ path TSP.", acknowledgement = ack-nhfb, } @InProceedings{Williams:2012:MMF, author = "Virginia Vassilevska Williams", title = "Multiplying matrices faster than {Coppersmith--Winograd}", crossref = "ACM:2012:SPA", pages = "887--898", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214056", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We develop an automated approach for designing matrix multiplication algorithms based on constructions similar to the Coppersmith--Winograd construction. Using this approach we obtain a new improved bound on the matrix multiplication exponent $\omega < 2.3727$.", acknowledgement = ack-nhfb, keywords = "fast matrix multiplication", } @InProceedings{Coja-Oglan:2012:CKN, author = "Amin Coja-Oglan and Konstantinos Panagiotou", title = "Catching the {$k$-NAESAT} threshold", crossref = "ACM:2012:SPA", pages = "899--908", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214058", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The best current estimates of the thresholds for the existence of solutions in random constraint satisfaction problems ('CSPs') mostly derive from the first and the second moment method. Yet apart from a very few exceptional cases these methods do not quite yield matching upper and lower bounds. According to deep but non-rigorous arguments from statistical mechanics, this discrepancy is due to a change in the geometry of the set of solutions called condensation that occurs shortly before the actual threshold for the existence of solutions (Krzakala, Montanari, Ricci-Tersenghi, Semerjian, Zdeborova: PNAS~2007). To cope with condensation, physicists have developed a sophisticated but non-rigorous formalism called Survey Propagation (Mezard, Parisi, Zecchina: Science 2002). This formalism yields precise conjectures on the threshold values of many random CSPs. Here we develop a new Survey Propagation inspired second moment method for the random $k$-NAESAT problem, which is one of the standard benchmark problems in the theory of random CSPs. This new technique allows us to overcome the barrier posed by condensation rigorously. We prove that the threshold for the existence of solutions in random $k$-NAESAT is $2^{k-1} \ln 2 - (\ln / 2 2 + 1/4) + \epsilon_k$, where $|\epsilon_k| \leq 2^{-(1 -o k(1))k}$, thereby verifying the statistical mechanics conjecture for this problem.", acknowledgement = ack-nhfb, } @InProceedings{Cai:2012:CCC, author = "Jin-Yi Cai and Xi Chen", title = "Complexity of counting {CSP} with complex weights", crossref = "ACM:2012:SPA", pages = "909--920", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214059", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give a complexity dichotomy theorem for the counting constraint satisfaction problem (\#CSP in short) with algebraic complex weights. To this end, we give three conditions for its tractability. Let $F$ be any finite set of complex-valued functions. We show that \#CSP($F$) is solvable in polynomial time if all three conditions are satisfied; and is \#P-hard otherwise. Our dichotomy theorem generalizes a long series of important results on counting problems: (a) the problem of counting graph homomorphisms is the special case when $F$ has a single symmetric binary function; (b) the problem of counting directed graph homomorphisms is the special case when $F$ has a single but not-necessarily-symmetric binary function; and (c) the unweighted form of \#CSP is when all functions in $F$ take values in $\{0, 1\}$.", acknowledgement = ack-nhfb, } @InProceedings{Molloy:2012:FTK, author = "Michael Molloy", title = "The freezing threshold for $k$-colourings of a random graph", crossref = "ACM:2012:SPA", pages = "921--930", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214060", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We rigorously determine the exact freezing threshold, $r_k^f$, for $k$-colourings of a random graph. We prove that for random graphs with density above $r_k^f$, almost every colouring is such that a linear number of variables are {\em frozen}, meaning that their colours cannot be changed by a sequence of alterations whereby we change the colours of $o(n)$ vertices at a time, always obtaining another proper colouring. When the density is below $r_k^f$, then almost every colouring has at most $o(n)$ frozen variables. This confirms hypotheses made using the non-rigorous cavity method.\par It has been hypothesized that the freezing threshold is the cause of the ``algorithmic barrier'', the long observed phenomenon that when the edge-density of a random graph exceeds $(1/2) k \ln k(1 + o_k(1))$, no algorithms are known to find $k$-colourings, despite the fact that this density is only half the $k$-colourability threshold.\par We also show that $r_k^f$ is the threshold of a strong form of reconstruction for $k$-colourings of the Galton--Watson tree, and of the graphical model.", acknowledgement = ack-nhfb, } @InProceedings{Barto:2012:RSC, author = "Libor Barto and Marcin Kozik", title = "Robust satisfiability of constraint satisfaction problems", crossref = "ACM:2012:SPA", pages = "931--940", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214061", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "An algorithm for a constraint satisfaction problem is called robust if it outputs an assignment satisfying at least $(1 - g(\epsilon))$-fraction of the constraints given a $(1 - \epsilon)$-satisfiable instance, where $g(\epsilon) \to 0$ as $\epsilon \to 0$, $g(0) = 0$. Guruswami and Zhou conjectured a characterization of constraint languages for which the corresponding constraint satisfaction problem admits an efficient robust algorithm. This paper confirms their conjecture.", acknowledgement = ack-nhfb, } @InProceedings{Woodruff:2012:TBD, author = "David P. Woodruff and Qin Zhang", title = "Tight bounds for distributed functional monitoring", crossref = "ACM:2012:SPA", pages = "941--960", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214063", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We resolve several fundamental questions in the area of distributed functional monitoring, initiated by Cormode, Muthukrishnan, and Yi (SODA, 2008), and receiving recent attention. In this model there are $k$ sites each tracking their input streams and communicating with a central coordinator. The coordinator's task is to continuously maintain an approximate output to a function computed over the union of the $k$ streams. The goal is to minimize the number of bits communicated. Let the $p$-th frequency moment be defined as $F_p = \sum_i f_i^p$, where $f_i$ is the frequency of element $i$. We show the randomized communication complexity of estimating the number of distinct elements (that is, $F_0$) up to a $1 + \epsilon$ factor is $\Omega(k / \epsilon^2)$, improving upon the previous $\Omega(k + 1/ \epsilon^2)$ bound and matching known upper bounds. For $F_p$, $p > 1$, we improve the previous $\Omega(k + 1/ \epsilon^2)$ communication bound to $\Omega(k^{p - 1} / \epsilon^2)$. We obtain similar improvements for heavy hitters, empirical entropy, and other problems. Our lower bounds are the first of any kind in distributed functional monitoring to depend on the product of $k$ and $1 / \epsilon^2$. Moreover, the lower bounds are for the static version of the distributed functional monitoring model where the coordinator only needs to compute the function at the time when all $k$ input streams end; surprisingly they almost match what is achievable in the (dynamic version of) distributed functional monitoring model where the coordinator needs to keep track of the function continuously at any time step. We also show that we can estimate $F_p$, for any $p > 1$, using $O(k^{p - 1} \poly(\epsilon^{-1}))$ communication. This drastically improves upon the previous $O(k^{2 p + 1} N^{1 - 2/p} \poly(\epsilon^{-1}))$ bound of Cormode, Muthukrishnan, and Yi for general $p$, and their $O(k^2 / \epsilon + k^{1.5} / \epsilon^3)$ bound for $p = 2$. For $p = 2$, our bound resolves their main open question. Our lower bounds are based on new direct sum theorems for approximate majority, and yield improvements to classical problems in the standard data stream model. First, we improve the known lower bound for estimating $F_p$, $p > 2$, in $t$ passes from $\Omega(n^{1 - 2 / p} /(\epsilon^{2 / p} t))$ to $\Omega(n^{1 - 2 / p} /(\epsilon^{4 / p} t))$, giving the first bound that matches what we expect when $p = 2$ for any constant number of passes. Second, we give the first lower bound for estimating $F_0$ in $t$ passes with $\Omega(1 / (\epsilon^2 t))$ bits of space that does not use the hardness of the gap-Hamming problem.", acknowledgement = ack-nhfb, } @InProceedings{Censor-Hillel:2012:GCP, author = "Keren Censor-Hillel and Bernhard Haeupler and Jonathan Kelner and Petar Maymounkov", title = "Global computation in a poorly connected world: fast rumor spreading with no dependence on conductance", crossref = "ACM:2012:SPA", pages = "961--970", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214064", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In this paper, we study the question of how efficiently a collection of interconnected nodes can perform a global computation in the GOSSIP model of communication. In this model, nodes do not know the global topology of the network, and they may only initiate contact with a single neighbor in each round. This model contrasts with the much less restrictive LOCAL model, where a node may simultaneously communicate with all of its neighbors in a single round. A basic question in this setting is how many rounds of communication are required for the information dissemination problem, in which each node has some piece of information and is required to collect all others. In the LOCAL model, this is quite simple: each node broadcasts all of its information in each round, and the number of rounds required will be equal to the diameter of the underlying communication graph. In the GOSSIP model, each node must independently choose a single neighbor to contact, and the lack of global information makes it difficult to make any sort of principled choice. As such, researchers have focused on the uniform gossip algorithm, in which each node independently selects a neighbor uniformly at random. When the graph is well-connected, this works quite well. In a string of beautiful papers, researchers proved a sequence of successively stronger bounds on the number of rounds required in terms of the conductance $\phi$ and graph size $n$, culminating in a bound of $O(\phi^{-1} \log n)$. In this paper, we show that a fairly simple modification of the protocol gives an algorithm that solves the information dissemination problem in at most $O(D + \polylog(n))$ rounds in a network of diameter $D$, with no dependence on the conductance. This is at most an additive polylogarithmic factor from the trivial lower bound of $D$, which applies even in the LOCAL model. In fact, we prove that something stronger is true: any algorithm that requires $T$ rounds in the LOCAL model can be simulated in $O(T + \polylog(n))$ rounds in the GOSSIP model. We thus prove that these two models of distributed computation are essentially equivalent.", acknowledgement = ack-nhfb, } @InProceedings{Bansal:2012:TTS, author = "Nikhil Bansal and Vibhor Bhatt and Prasad Jayanti and Ranganath Kondapally", title = "Tight time-space tradeoff for mutual exclusion", crossref = "ACM:2012:SPA", pages = "971--982", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214065", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Mutual Exclusion is a fundamental problem in distributed computing, and the problem of proving upper and lower bounds on the RMR complexity of this problem has been extensively studied. Here, we give matching lower and upper bounds on how RMR complexity trades off with space. Two implications of our results are that constant RMR complexity is impossible with subpolynomial space and subpolynomial RMR complexity is impossible with constant space for cache-coherent multiprocessors, regardless of how strong the hardware synchronization operations are. To prove these results we show that the complexity of mutual exclusion, which can be ``messy'' to analyze because of system details such as asynchrony and cache coherence, is captured precisely by a simple and purely combinatorial game that we design. We then derive lower and upper bounds for this game, thereby obtaining corresponding bounds for mutual exclusion. The lower bounds for the game are proved using potential functions.", acknowledgement = ack-nhfb, } @InProceedings{Giakkoupis:2012:TRL, author = "George Giakkoupis and Philipp Woelfel", title = "A tight {RMR} lower bound for randomized mutual exclusion", crossref = "ACM:2012:SPA", pages = "983--1002", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214066", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Cache Coherent (CC) and the Distributed Shared Memory (DSM) models are standard shared memory models, and the Remote Memory Reference (RMR) complexity is considered to accurately predict the actual performance of mutual exclusion algorithms in shared memory systems. In this paper we prove a tight lower bound for the RMR complexity of deadlock-free randomized mutual exclusion algorithms in both the CC and the DSM model with atomic registers and compare \& swap objects and an adaptive adversary. Our lower bound establishes that an adaptive adversary can schedule $n$ processes in such a way that each enters the critical section once, and the total number of RMRs is $\Omega(n \log n/\log \log n)$ in expectation. This matches an upper bound of Hendler and Woelfel (2011).", acknowledgement = ack-nhfb, } @InProceedings{Garg:2012:CPA, author = "Jugal Garg and Ruta Mehta and Milind Sohoni and Vijay V. Vazirani", title = "A complementary pivot algorithm for markets under separable, piecewise-linear concave utilities", crossref = "ACM:2012:SPA", pages = "1003--1016", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214068", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Using the powerful machinery of the linear complementarity problem and Lemke's algorithm, we give a practical algorithm for computing an equilibrium for Arrow--Debreu markets under separable, piecewise-linear concave (SPLC) utilities, despite the PPAD-completeness of this case. As a corollary, we obtain the first elementary proof of existence of equilibrium for this case, i.e., without using fixed point theorems. In 1975, Eaves [10] had given such an algorithm for the case of linear utilities and had asked for an extension to the piecewise-linear, concave utilities. Our result settles the relevant subcase of his problem as well as the problem of Vazirani and Yannakakis of obtaining a path following algorithm for SPLC markets, thereby giving a direct proof of membership of this case in PPAD. We also prove that SPLC markets have an odd number of equilibria (up to scaling), hence matching the classical result of Shapley about 2-Nash equilibria [24], which was based on the Lemke--Howson algorithm. For the linear case, Eaves had asked for a combinatorial interpretation of his algorithm. We provide this and it yields a particularly simple proof of the fact that the set of equilibrium prices is convex.", acknowledgement = ack-nhfb, } @InProceedings{Azar:2012:RP, author = "Pablo Daniel Azar and Silvio Micali", title = "Rational proofs", crossref = "ACM:2012:SPA", pages = "1017--1028", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214069", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study a new type of proof system, where an unbounded prover and a polynomial time verifier interact, on inputs a string $x$ and a function $f$, so that the Verifier may learn $f(x)$. The novelty of our setting is that there no longer are ``good'' or ``malicious'' provers, but only rational ones. In essence, the Verifier has a budget $c$ and gives the Prover a reward $r \in [0,c]$ determined by the transcript of their interaction; the prover wishes to maximize his expected reward; and his reward is maximized only if he the verifier correctly learns $f(x)$. Rational proof systems are as powerful as their classical counterparts for polynomially many rounds of interaction, but are much more powerful when we only allow a constant number of rounds. Indeed, we prove that if $f \in \#P$, then $f$ is computable by a one-round rational Merlin--Arthur game, where, on input $x$, Merlin's single message actually consists of sending just the value $f(x)$. Further, we prove that CH, the counting hierarchy, coincides with the class of languages computable by a constant-round rational Merlin--Arthur game. Our results rely on a basic and crucial connection between rational proof systems and proper scoring rules, a tool developed to elicit truthful information from experts.", acknowledgement = ack-nhfb, } @InProceedings{Abernethy:2012:MOP, author = "Jacob Abernethy and Rafael M. Frongillo and Andre Wibisono", title = "Minimax option pricing meets {Black--Scholes} in the limit", crossref = "ACM:2012:SPA", pages = "1029--1040", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214070", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Option contracts are a type of financial derivative that allow investors to hedge risk and speculate on the variation of an asset's future market price. In short, an option has a particular payout that is based on the market price for an asset on a given date in the future. In 1973, Black and Scholes proposed a valuation model for options that essentially estimates the tail risk of the asset price under the assumption that the price will fluctuate according to geometric Brownian motion. A key element of their analysis is that the investor can ``hedge'' the payout of the option by continuously buying and selling the asset depending on the price fluctuations. More recently, DeMarzo et al. proposed a more robust valuation scheme which does not require any assumption on the price path; indeed, in their model the asset's price can even be chosen adversarially. This framework can be considered as a sequential two-player zero-sum game between the investor and Nature. We analyze the value of this game in the limit, where the investor can trade at smaller and smaller time intervals. Under weak assumptions on the actions of Nature (an adversary), we show that the minimax option price asymptotically approaches exactly the Black--Scholes valuation. The key piece of our analysis is showing that Nature's minimax optimal dual strategy converges to geometric Brownian motion in the limit.", acknowledgement = ack-nhfb, } @InProceedings{Mossel:2012:QGS, author = "Elchanan Mossel and Mikl{\'o}s Z. R{\'a}cz", title = "A quantitative {Gibbard--Satterthwaite} theorem without neutrality", crossref = "ACM:2012:SPA", pages = "1041--1060", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214071", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Recently, quantitative versions of the Gibbard--Satterthwaite theorem were proven for $k = 3$ alternatives by Friedgut, Kalai, Keller and Nisan and for neutral functions on $k \geq 4$ alternatives by Isaksson, Kindler and Mossel. In the present paper we prove a quantitative version of the Gibbard--Satterthwaite theorem for general social choice functions for any number $k \geq 3$ of alternatives. In particular we show that for a social choice function $f$ on $k \geq 3$ alternatives and $n$ voters, which is $\epsilon$-far from the family of nonmanipulable functions, a uniformly chosen voter profile is manipulable with probability at least inverse polynomial in $n$, $k$, and $\epsilon^{-1}$. Removing the neutrality assumption of previous theorems is important for multiple reasons. For one, it is known that there is a conflict between anonymity and neutrality, and since most common voting rules are anonymous, they cannot always be neutral. Second, virtual elections are used in many applications in artificial intelligence, where there are often restrictions on the outcome of the election, and so neutrality is not a natural assumption in these situations. Ours is a unified proof which in particular covers all previous cases established before. The proof crucially uses reverse hypercontractivity in addition to several ideas from the two previous proofs. Much of the work is devoted to understanding functions of a single voter, and in particular we also prove a quantitative Gibbard--Satterthwaite theorem for one voter.", acknowledgement = ack-nhfb, } @InProceedings{Bourgain:2012:ME, author = "Jean Bourgain and Amir Yehudayoff", title = "Monotone expansion", crossref = "ACM:2012:SPA", pages = "1061--1078", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214073", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "This work presents an explicit construction of a family of monotone expanders, which are bi-partite expander graphs whose edge-set is defined by (partial) monotone functions. The family is essentially defined by the M{\"o}bius action of ${\rm SL}_2(R)$, the group of $2 \times 2$ matrices with determinant one, on the interval $[0, 1]$. No other proof-of-existence for monotone expanders is known, not even using the probabilistic method. The proof extends recent results on finite/compact groups to the non-compact scenario. Specifically, we show a product-growth theorem for ${\rm SL}_2(R)$; roughly, that for every $A \subset {\rm SL}_2(R)$ with certain properties, the size of $AAA$ is much larger than that of $A$. We mention two applications of this construction: Dvir and Shpilka showed that it yields a construction of explicit dimension expanders, which are a generalization of standard expander graphs. Dvir and Wigderson proved that it yields the existence of explicit pushdown expanders, which are graphs that arise in Turing machine simulations.", acknowledgement = ack-nhfb, } @InProceedings{Alon:2012:NCG, author = "Noga Alon and Ankur Moitra and Benny Sudakov", title = "Nearly complete graphs decomposable into large induced matchings and their applications", crossref = "ACM:2012:SPA", pages = "1079--1090", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214074", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We describe two constructions of (very) dense graphs which are edge disjoint unions of large {\em induced\/} matchings. The first construction exhibits graphs on $N$ vertices with ${N \choose 2} - o(N^2)$ edges, which can be decomposed into pairwise disjoint induced matchings, each of size $N^{1 - o(1)}$. The second construction provides a covering of all edges of the complete graph $K_N$ by two graphs, each being the edge disjoint union of at most $N^{2 - \delta}$ induced matchings, where $\delta > 0.076$. This disproves (in a strong form) a conjecture of Meshulam, substantially improves a result of Birk, Linial and Meshulam on communicating over a shared channel, and (slightly) extends the analysis of Hastad and Wigderson of the graph test of Samorodnitsky and Trevisan for linearity. Additionally, our constructions settle a combinatorial question of Vempala regarding a candidate rounding scheme for the directed Steiner tree problem.", acknowledgement = ack-nhfb, } @InProceedings{Kuperberg:2012:PER, author = "Greg Kuperberg and Shachar Lovett and Ron Peled", title = "Probabilistic existence of rigid combinatorial structures", crossref = "ACM:2012:SPA", pages = "1091--1106", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214075", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show the existence of rigid combinatorial objects which previously were not known to exist. Specifically, for a wide range of the underlying parameters, we show the existence of non-trivial orthogonal arrays, $t$-designs, and $t$-wise permutations. In all cases, the sizes of the objects are optimal up to polynomial overhead. The proof of existence is probabilistic. We show that a randomly chosen such object has the required properties with positive yet tiny probability. The main technical ingredient is a special local central limit theorem for suitable lattice random walks with finitely many steps.", acknowledgement = ack-nhfb, } @InProceedings{Dobzinski:2012:QCC, author = "Shahar Dobzinski and Jan Vondrak", title = "From query complexity to computational complexity", crossref = "ACM:2012:SPA", pages = "1107--1116", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214076", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider submodular optimization problems, and provide a general way of translating oracle inapproximability results arising from the symmetry gap technique to computational complexity inapproximability results, where the submodular function is given explicitly (under the assumption that NP $\not=$ RP). Applications of our technique include an optimal computational hardness of $(1/2 + \epsilon)$-approximation for maximizing a symmetric nonnegative submodular function, an optimal hardness of $(1 - (1 - 1 / k)^k + \epsilon)$-approximation for welfare maximization in combinatorial auctions with $k$ submodular bidders (for constant $k$), super-constant hardness for maximizing a nonnegative submodular function over matroid bases, and tighter bounds for maximizing a monotone submodular function subject to a cardinality constraint. Unlike the vast majority of computational inapproximability results, our approach does not use the PCP machinery or the Unique Games Conjecture, but relies instead on a direct reduction from Unique-SAT using list-decodable codes.", acknowledgement = ack-nhfb, } @InProceedings{Lee:2012:MWS, author = "James R. Lee and Shayan Oveis Gharan and Luca Trevisan", title = "Multi-way spectral partitioning and higher-order {Cheeger} inequalities", crossref = "ACM:2012:SPA", pages = "1117--1130", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214078", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A basic fact in spectral graph theory is that the number of connected components in an undirected graph is equal to the multiplicity of the eigenvalue zero in the Laplacian matrix of the graph. In particular, the graph is disconnected if and only if there are at least two eigenvalues equal to zero. Cheeger's inequality and its variants provide an approximate version of the latter fact; they state that a graph has a sparse cut if and only if there are at least two eigenvalues that are close to zero. It has been conjectured that an analogous characterization holds for higher multiplicities, i.e., there are $k$ eigenvalues close to zero if and only if the vertex set can be partitioned into $k$ subsets, each defining a sparse cut. We resolve this conjecture. Our result provides a theoretical justification for clustering algorithms that use the bottom $k$ eigenvectors to embed the vertices into R$^k$, and then apply geometric considerations to the embedding. We also show that these techniques yield a nearly optimal quantitative connection between the expansion of sets of size $\approx n/k$ and $\lambda_k$, the $k$th smallest eigenvalue of the normalized Laplacian, where $n$ is the number of vertices. In particular, we show that in every graph there are at least $k / 2$ disjoint sets (one of which will have size at most $2 n / k$), each having expansion at most $O(\sqrt{\lambda_k \log k})$. Louis, Raghavendra, Tetali, and Vempala have independently proved a slightly weaker version of this last result [LRTV12]. The $\sqrt{\log k}$ bound is tight, up to constant factors, for the ``noisy hypercube'' graphs.", acknowledgement = ack-nhfb, } @InProceedings{Louis:2012:MSC, author = "Anand Louis and Prasad Raghavendra and Prasad Tetali and Santosh Vempala", title = "Many sparse cuts via higher eigenvalues", crossref = "ACM:2012:SPA", pages = "1131--1140", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214079", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Cheeger's fundamental inequality states that any edge-weighted graph has a vertex subset $S$ such that its expansion (a.k.a. conductance) is bounded as follows:\par $$\phi(S) {\hbox{\tiny \rm def} \atop =} (w(S, \bar{S})) / \min \{w(S), w(\bar{S})\} \leq \sqrt{2 \lambda_2}$$\par where $w$ is the total edge weight of a subset or a cut and $\lambda_2$ is the second smallest eigenvalue of the normalized Laplacian of the graph. Here we prove the following natural generalization: for any integer $k \in [n]$, there exist $c k$ disjoint subsets $S_1$, \ldots{}, $S_{c k}$, such that\par $$\max_i \phi(S_i) \leq C \sqrt{\lambda_k \log k}$$\par where $\lambda_k$ is the $k$th smallest eigenvalue of the normalized Laplacian and $c < 1$, $C > 0$ are suitable absolute constants. Our proof is via a polynomial-time algorithm to find such subsets, consisting of a spectral projection and a randomized rounding. As a consequence, we get the same upper bound for the small set expansion problem, namely for any $k$, there is a subset $S$ whose weight is at most a $O(1/k)$ fraction of the total weight and $\phi(S) \leq C \sqrt{\lambda_k \log k}$. Both results are the best possible up to constant factors.\par The underlying algorithmic problem, namely finding $k$ subsets such that the maximum expansion is minimized, besides extending sparse cuts to more than one subset, appears to be a natural clustering problem in its own right.", acknowledgement = ack-nhfb, } @InProceedings{Orecchia:2012:AEL, author = "Lorenzo Orecchia and Sushant Sachdeva and Nisheeth K. Vishnoi", title = "Approximating the exponential, the {Lanczos} method and an {$\tilde{O}(m)$}-time spectral algorithm for balanced separator", crossref = "ACM:2012:SPA", pages = "1141--1160", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214080", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give a novel spectral approximation algorithm for the balanced (edge-)separator problem that, given a graph $G$, a constant balance $b \in (0,1/2]$, and a parameter $\gamma$, either finds an $\Omega(b)$-balanced cut of conductance $O(\sqrt \gamma)$ in $G$, or outputs a certificate that all $b$-balanced cuts in $G$ have conductance at least $\gamma$, and runs in time $\tilde{O}(m)$. This settles the question of designing asymptotically optimal spectral algorithms for balanced separator. Our algorithm relies on a variant of the heat kernel random walk and requires, as a subroutine, an algorithm to compute $\exp(-L) v$ where $L$ is the Laplacian of a graph related to $G$ and $v$ is a vector. Algorithms for computing the matrix-exponential-vector product efficiently comprise our next set of results. Our main result here is a new algorithm which computes a good approximation to $\exp(-A) v$ for a class of symmetric positive semidefinite (PSD) matrices $A$ and a given vector $v$, in time roughly $\tilde{O}(m_A)$, independent of the norm of $A$, where $m_A$ is the number of non-zero entries of $A$. This uses, in a non-trivial way, the result of Spielman and Teng on inverting symmetric and diagonally-dominant matrices in $\tilde{O}(m_A)$ time. Finally, using old and new uniform approximations to $e^{-x}$ we show how to obtain, via the Lanczos method, a simple algorithm to compute $\exp(-A) v$ for symmetric PSD matrices that runs in time roughly $O(t_A \cdot \sqrt{\norm(A)})$, where $t_A$ is the time required for the computation of the vector $A w$ for given vector w. As an application, we obtain a simple and practical algorithm, with output conductance $O(\sqrt \gamma)$, for balanced separator that runs in time $O(m / \sqrt \gamma)$. This latter algorithm matches the running time, but improves on the approximation guarantee of the Evolving-Sets-based algorithm by Andersen and Peres for balanced separator.", acknowledgement = ack-nhfb, } @InProceedings{Goemans:2012:MIG, author = "Michel X. Goemans and Neil Olver and Thomas Rothvo{\ss} and Rico Zenklusen", title = "Matroids and integrality gaps for hypergraphic {Steiner} tree relaxations", crossref = "ACM:2012:SPA", pages = "1161--1176", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214081", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Until recently, LP relaxations have only played a very limited role in the design of approximation algorithms for the Steiner tree problem. In particular, no (efficiently solvable) Steiner tree relaxation was known to have an integrality gap bounded away from $2$, before Byrka et al. [3] showed an upper bound of $\approx 1.55$ of a hypergraphic LP relaxation and presented a $\ln(4) + \epsilon \approx 1.39$ approximation based on this relaxation. Interestingly, even though their approach is LP based, they do not compare the solution produced against the LP value.\par We take a fresh look at hypergraphic LP relaxations for the Steiner tree problem --- one that heavily exploits methods and results from the theory of matroids and submodular functions --- which leads to stronger integrality gaps, faster algorithms, and a variety of structural insights of independent interest. More precisely, along the lines of the algorithm of Byrka et al.[3], we present a deterministic $\ln(4) + \epsilon$ approximation that compares against the LP value and therefore proves a matching $\ln(4)$ upper bound on the integrality gap of hypergraphic relaxations.\par Similarly to [3], we iteratively fix one component and update the LP solution. However, whereas in [3] the LP is solved at every iteration after contracting a component, we show how feasibility can be maintained by a greedy procedure on a well-chosen matroid. Apart from avoiding the expensive step of solving a hypergraphic LP at each iteration, our algorithm can be analyzed using a simple potential function. This potential function gives an easy means to determine stronger approximation guarantees and integrality gaps when considering restricted graph topologies. In particular, this readily leads to a $73/60 \approx 1.217$ upper bound on the integrality gap of hypergraphic relaxations for quasi-bipartite graphs.\par Additionally, for the case of quasi-bipartite graphs, we present a simple algorithm to transform an optimal solution to the bidirected cut relaxation to an optimal solution of the hypergraphic relaxation, leading to a fast $73/60$ approximation for quasi-bipartite graphs. Furthermore, we show how the separation problem of the hypergraphic relaxation can be solved by computing maximum flows, which provides a way to obtain a fast independence oracle for the matroids that we use in our approach.", acknowledgement = ack-nhfb, } @InProceedings{Brodal:2012:SFH, author = "Gerth St{\o}lting Brodal and George Lagogiannis and Robert E. Tarjan", title = "Strict {Fibonacci} heaps", crossref = "ACM:2012:SPA", pages = "1177--1184", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214082", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present the first pointer-based heap implementation with time bounds matching those of Fibonacci heaps in the worst case. We support make-heap, insert, find-min, meld and decrease-key in worst-case $O(1)$ time, and delete and delete-min in worst-case $O(\lg n)$ time, where $n$ is the size of the heap. The data structure uses linear space. A previous, very complicated, solution achieving the same time bounds in the RAM model made essential use of arrays and extensive use of redundant counter schemes to maintain balance. Our solution uses neither. Our key simplification is to discard the structure of the smaller heap when doing a meld. We use the pigeonhole principle in place of the redundant counter mechanism.", acknowledgement = ack-nhfb, } @InProceedings{Bulnek:2012:TLB, author = "Jan Bul{\'a}nek and Michal Kouck{\'y} and Michael Saks", title = "Tight lower bounds for the online labeling problem", crossref = "ACM:2012:SPA", pages = "1185--1198", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214083", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider the file maintenance problem (also called the online labeling problem) in which $n$ integer items from the set $\{1, \ldots{}, r\}$ are to be stored in an array of size $m \geq n$. The items are presented sequentially in an arbitrary order, and must be stored in the array in sorted order (but not necessarily in consecutive locations in the array). Each new item must be stored in the array before the next item is received. If $r \leq m$ then we can simply store item $j$ in location $j$ but if $r > m$ then we may have to shift the location of stored items to make space for a newly arrived item. The algorithm is charged each time an item is stored in the array, or moved to a new location. The goal is to minimize the total number of such moves the algorithm has to do. This problem is non-trivial when $n \leq m < r$. In the case that $m = Cn$ for some $C > 1$, algorithms for this problem with cost $O(\log(n)^2)$ per item have been given [Itai et al. (1981), Willard (1992), Bender et al. (2002)]. When $m = n$, algorithms with cost $O(\log(n)^3)$ per item were given [Zhang (1993),Bird and Sadnicki (2007)]. In this paper we prove lower bounds that show that these algorithms are optimal, up to constant factors. Previously, the only lower bound known for this range of parameters was a lower bound of $\Omega(\log(n)^2)$ for the restricted class of smooth algorithms [Dietz et al. (2005), Zhang (1993)]. We also provide an algorithm for the sparse case: If the number of items is polylogarithmic in the array size then the problem can be solved in amortized constant time per item.", acknowledgement = ack-nhfb, } @InProceedings{Abraham:2012:FDA, author = "Ittai Abraham and Shiri Chechik and Cyril Gavoille", title = "Fully dynamic approximate distance oracles for planar graphs via forbidden-set distance labels", crossref = "ACM:2012:SPA", pages = "1199--1218", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214084", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "This paper considers fully dynamic $(1 + \epsilon)$ distance oracles and $(1 + \epsilon)$ forbidden-set labeling schemes for planar graphs. For a given $n$-vertex planar graph $G$ with edge weights drawn from $[1,M]$ and parameter $\epsilon > 0$, our forbidden-set labeling scheme uses labels of length $\lambda = O(\epsilon^{-1} \log^2 n \log (n M) \cdot \log n)$. Given the labels of two vertices $s$ and $t$ and of a set $F$ of faulty vertices\slash edges, our scheme approximates the distance between $s$ and $t$ in $G \backslash F$ with stretch $(1 + \epsilon)$, in $O(|F|^2 \lambda)$ time.\par We then present a general method to transform $(1 + \epsilon)$ forbidden-set labeling schemas into a fully dynamic $(1 + \epsilon)$ distance oracle. Our fully dynamic $(1 + \epsilon)$ distance oracle is of size $O(n \log n \cdot (\epsilon^{-1} + \log n))$ and has $\tilde{O}(n^{1/2})$ query and update time, both the query and the update time are worst case. This improves on the best previously known $(1 + \epsilon)$ dynamic distance oracle for planar graphs, which has worst case query time $\tilde{O}(n^{2/3})$ and amortized update time of $\tilde{O}(n^{2/3})$.\par Our $(1 + \epsilon)$ forbidden-set labeling scheme can also be extended into a forbidden-set labeled routing scheme with stretch $(1 + \epsilon)$.", acknowledgement = ack-nhfb, } @InProceedings{Lopez-Alt:2012:FMC, author = "Adriana L{\'o}pez-Alt and Eran Tromer and Vinod Vaikuntanathan", title = "On-the-fly multiparty computation on the cloud via multikey fully homomorphic encryption", crossref = "ACM:2012:SPA", pages = "1219--1234", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214086", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We propose a new notion of secure multiparty computation aided by a computationally-powerful but untrusted ``cloud'' server. In this notion that we call on-the-fly multiparty computation (MPC), the cloud can non-interactively perform arbitrary, dynamically chosen computations on data belonging to arbitrary sets of users chosen on-the-fly. All user's input data and intermediate results are protected from snooping by the cloud as well as other users. This extends the standard notion of fully homomorphic encryption (FHE), where users can only enlist the cloud's help in evaluating functions on their own encrypted data. In on-the-fly MPC, each user is involved only when initially uploading his (encrypted) data to the cloud, and in a final output decryption phase when outputs are revealed; the complexity of both is independent of the function being computed and the total number of users in the system. When users upload their data, they need not decide in advance which function will be computed, nor who they will compute with; they need only retroactively approve the eventually-chosen functions and on whose data the functions were evaluated. This notion is qualitatively the best possible in minimizing interaction, since the users' interaction in the decryption stage is inevitable: we show that removing it would imply generic program obfuscation and is thus impossible. Our contributions are two-fold:- We show how on-the-fly MPC can be achieved using a new type of encryption scheme that we call multikey FHE, which is capable of operating on inputs encrypted under multiple, unrelated keys. A ciphertext resulting from a multikey evaluation can be jointly decrypted using the secret keys of all the users involved in the computation. --- We construct a multikey FHE scheme based on NTRU, a very efficient public-key encryption scheme proposed in the 1990s. It was previously not known how to make NTRU fully homomorphic even for a single party. We view the construction of (multikey) FHE from NTRU encryption as a main contribution of independent interest. Although the transformation to a fully homomorphic system deteriorates the efficiency of NTRU somewhat, we believe that this system is a leading candidate for a practical FHE scheme.", acknowledgement = ack-nhfb, } @InProceedings{Boyle:2012:MCS, author = "Elette Boyle and Shafi Goldwasser and Abhishek Jain and Yael Tauman Kalai", title = "Multiparty computation secure against continual memory leakage", crossref = "ACM:2012:SPA", pages = "1235--1254", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214087", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We construct a multiparty computation (MPC) protocol that is secure even if a malicious adversary, in addition to corrupting $1 - \epsilon$ fraction of all parties for an arbitrarily small constant $\epsilon > 0$, can leak information about the secret state of each honest party. This leakage can be continuous for an unbounded number of executions of the MPC protocol, computing different functions on the same or different set of inputs. We assume a (necessary) ``leak-free'' preprocessing stage. We emphasize that we achieve leakage resilience without weakening the security guarantee of classical MPC. Namely, an adversary who is given leakage on honest parties' states, is guaranteed to learn nothing beyond the input and output values of corrupted parties. This is in contrast with previous works on leakage in the multi-party protocol setting, which weaken the security notion, and only guarantee that a protocol which leaks $l$ bits about the parties' secret states, yields at most $l$ bits of leakage on the parties' private inputs. For some functions, such as voting, such leakage can be detrimental. Our result relies on standard cryptographic assumptions, and our security parameter is polynomially related to the number of parties.", acknowledgement = ack-nhfb, } @InProceedings{Hardt:2012:BRR, author = "Moritz Hardt and Aaron Roth", title = "Beating randomized response on incoherent matrices", crossref = "ACM:2012:SPA", pages = "1255--1268", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214088", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Computing accurate low rank approximations of large matrices is a fundamental data mining task. In many applications however the matrix contains sensitive information about individuals. In such case we would like to release a low rank approximation that satisfies a strong privacy guarantee such as differential privacy. Unfortunately, to date the best known algorithm for this task that satisfies differential privacy is based on naive input perturbation or randomized response: Each entry of the matrix is perturbed independently by a sufficiently large random noise variable, a low rank approximation is then computed on the resulting matrix. We give (the first) significant improvements in accuracy over randomized response under the natural and necessary assumption that the matrix has low coherence. Our algorithm is also very efficient and finds a constant rank approximation of an $m \times n$ matrix in time $O(m n)$. Note that even generating the noise matrix required for randomized response already requires time $O(mn)$.", acknowledgement = ack-nhfb, } @InProceedings{Bhaskara:2012:UDP, author = "Aditya Bhaskara and Daniel Dadush and Ravishankar Krishnaswamy and Kunal Talwar", title = "Unconditional differentially private mechanisms for linear queries", crossref = "ACM:2012:SPA", pages = "1269--1284", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214089", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We investigate the problem of designing differentially private mechanisms for a set of $d$ linear queries over a database, while adding as little error as possible. Hardt and Talwar [HT10] related this problem to geometric properties of a convex body defined by the set of queries and gave a $O(\log^3 d)$-approximation to the minimum $l_2^2$ error, assuming a conjecture from convex geometry called the Slicing or Hyperplane conjecture. In this work we give a mechanism that works unconditionally, and also gives an improved $O(\log^2 d)$ approximation to the expected $l_2^2$ error. We remove the dependence on the Slicing conjecture by using a result of Klartag [Kla06] that shows that any convex body is close to one for which the conjecture holds; our main contribution is in making this result constructive by using recent techniques of Dadush, Peikert and Vempala [DPV10]. The improvement in approximation ratio relies on a stronger lower bound we derive on the optimum. This new lower bound goes beyond the packing argument that has traditionally been used in Differential Privacy and allows us to add the packing lower bounds obtained from orthogonal subspaces. We are able to achieve this via a symmetrization argument which argues that there always exists a near optimal differentially private mechanism which adds noise that is independent of the input database! We believe this result should be of independent interest, and also discuss some interesting consequences.", acknowledgement = ack-nhfb, } @InProceedings{Muthukrishnan:2012:OPH, author = "S. Muthukrishnan and Aleksandar Nikolov", title = "Optimal private halfspace counting via discrepancy", crossref = "ACM:2012:SPA", pages = "1285--1292", year = "2012", DOI = "http://dx.doi.org/10.1145/2213977.2214090", bibdate = "Thu Nov 8 19:11:58 MST 2012", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A {\em range counting\/} problem is specified by a set $P$ of size $|P| = n$ of points in $\mathbb{R}^d$, an integer {\em weight\/} $x_p$ associated to each point $p \in P$, and a {\em range space\/} ${\cal R} \subseteq 2^P$. Given a query range $R \in {\cal R}$, the output is $R(x) = \sum_{p \in R} x_p$. The {\em average squared error\/} of an algorithm ${\cal A}$ is $1/|R| \sum_{R \in {\cal R}} ({\cal A}(R, x) - R(x))^2$. Range counting for different range spaces is a central problem in Computational Geometry.\par We study $(\epsilon, \delta)$-differentially private algorithms for range counting. Our main results are for the range space given by hyperplanes, that is, the halfspace counting problem. We present an $(\epsilon, \delta)$-differentially private algorithm for halfspace counting in $d$ dimensions which is $O(n^{1 - 1/d})$ approximate for average squared error. This contrasts with the $\Omega(n)$ lower bound established by the classical result of Dinur and Nissim on approximation for arbitrary subset counting queries. We also show a matching lower bound of $\Omega(n^{1 - 1 /d})$ approximation for any $(\epsilon, \delta)$-differentially private algorithm for halfspace counting.\par Both bounds are obtained using discrepancy theory. For the lower bound, we use a modified discrepancy measure and bound approximation of $(\epsilon, \delta)$-differentially private algorithms for range counting queries in terms of this discrepancy. We also relate the modified discrepancy measure to classical combinatorial discrepancy, which allows us to exploit known discrepancy lower bounds. This approach also yields a lower bound of $\Omega((\log n)^{d - 1})$ for $(\epsilon, \delta)$-differentially private {\em orthogonal\/} range counting in $d$ dimensions, the first known superconstant lower bound for this problem. For the upper bound, we use an approach inspired by partial coloring methods for proving discrepancy upper bounds, and obtain $(\epsilon, \delta)$-differentially private algorithms for range counting with polynomially bounded shatter function range spaces.", acknowledgement = ack-nhfb, } @InProceedings{Kane:2013:PLF, author = "Daniel M. Kane and Raghu Meka", title = "A {PRG} for {Lipschitz} functions of polynomials with applications to sparsest cut", crossref = "ACM:2013:SPF", pages = "1--10", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488610", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/prng.bib; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give improved pseudorandom generators (PRGs) for Lipschitz functions of low-degree polynomials over the hypercube. These are functions of the form $ \psi (P(x)) $, where $ P : {1, - 1}^n \to R $ is a low-degree polynomial and $ \psi : R \to R $ is a function with small Lipschitz constant. PRGs for smooth functions of low-degree polynomials have received a lot of attention recently and play an important role in constructing PRGs for the natural class of polynomial threshold functions [12,13,24,16,15]. In spite of the recent progress, no nontrivial PRGs were known for fooling Lipschitz functions of degree $ O(\log n) $ polynomials even for constant error rate. In this work, we give the first such generator obtaining a seed-length of $ (\log n) O(l^2 / \epsilon^2) $ for fooling degree $l$ polynomials with error $ \epsilon $. Previous generators had an exponential dependence on the degree $l$. We use our PRG to get better integrality gap instances for sparsest cut, a fundamental problem in graph theory with many applications in graph optimization. We give an instance of uniform sparsest cut for which a powerful semi-definite relaxation (SDP) first introduced by Goemans and Linial and studied in the seminal work of Arora, Rao and Vazirani [3] has an integrality gap of $ \exp (\Omega ((\log \log n)^{1 / 2})) $. Understanding the performance of the Goemans--Linial SDP for uniform sparsest cut is an important open problem in approximation algorithms and metric embeddings. Our work gives a near-exponential improvement over previous lower bounds which achieved a gap of $ \Omega (\log \log n) $ [11,21]. Our gap instance builds on the recent short code gadgets of Barak et al. [5].", acknowledgement = ack-nhfb, } @InProceedings{Kwok:2013:ICI, author = "Tsz Chiu Kwok and Lap Chi Lau and Yin Tat Lee and Shayan Oveis Gharan and Luca Trevisan", title = "Improved {Cheeger}'s inequality: analysis of spectral partitioning algorithms through higher order spectral gap", crossref = "ACM:2013:SPF", pages = "11--20", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488611", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Let $ \phi (G) $ be the minimum conductance of an undirected graph $G$, and let $ 0 = \lambda_1 \leq \lambda_2 \leq \ldots {} \leq \lambda_n \leq 2 $ be the eigenvalues of the normalized Laplacian matrix of $G$. We prove that for any graph $G$ and any $ k \geq 2 $, $ [\phi (G) = O(k) l_2 / \sqrt l_k] $, and this performance guarantee is achieved by the spectral partitioning algorithm. This improves Cheeger's inequality, and the bound is optimal up to a constant factor for any $k$. Our result shows that the spectral partitioning algorithm is a constant factor approximation algorithm for finding a sparse cut if $ l_k $ is a constant for some constant $k$. This provides some theoretical justification to its empirical performance in image segmentation and clustering problems. We extend the analysis to spectral algorithms for other graph partitioning problems, including multi-way partition, balanced separator, and maximum cut.", acknowledgement = ack-nhfb, } @InProceedings{Williams:2013:NPV, author = "Ryan Williams", title = "Natural proofs versus derandomization", crossref = "ACM:2013:SPF", pages = "21--30", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488612", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study connections between Natural Proofs, derandomization, and the problem of proving ``weak'' circuit lower bounds such as $ {\rm NEXP} \not \subset {\rm TC}^0 $, which are still wide open. Natural Proofs have three properties: they are constructive (an efficient algorithm $A$ is embedded in them), have largeness ($A$ accepts a large fraction of strings), and are useful ($A$ rejects all strings which are truth tables of small circuits). Strong circuit lower bounds that are ``naturalizing'' would contradict present cryptographic understanding, yet the vast majority of known circuit lower bound proofs are naturalizing. So it is imperative to understand how to pursue un-Natural Proofs. Some heuristic arguments say constructivity should be circumventable. Largeness is inherent in many proof techniques, and it is probably our presently weak techniques that yield constructivity. We prove: Constructivity is unavoidable, even for NEXP lower bounds. Informally, we prove for all ``typical'' non-uniform circuit classes $C$, $ {\rm NEXP} \not \subset C $ if and only if there is a polynomial-time algorithm distinguishing some function from all functions computable by $C$-circuits. Hence $ {\rm NEXP} \not \subset C $ is equivalent to exhibiting a constructive property useful against $C$. There are no P-natural properties useful against $C$ if and only if randomized exponential time can be ``derandomized'' using truth tables of circuits from $C$ as random seeds. Therefore the task of proving there are no $P$-natural properties is inherently a derandomization problem, weaker than but implied by the existence of strong pseudorandom functions. These characterizations are applied to yield several new results. The two main applications are that $ {\rm NEXP} \cap {\rm coNEXP} $ does not have $ n^{\log n} $ size ACC circuits, and a mild derandomization result for RP.", acknowledgement = ack-nhfb, } @InProceedings{Bei:2013:CTE, author = "Xiaohui Bei and Ning Chen and Shengyu Zhang", title = "On the complexity of trial and error", crossref = "ACM:2013:SPF", pages = "31--40", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488613", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Motivated by certain applications from physics, biochemistry, economics, and computer science in which the objects under investigation are unknown or not directly accessible because of various limitations, we propose a trial-and-error model to examine search problems in which inputs are unknown. More specifically, we consider constraint satisfaction problems $ \wedge_i C_i $, where the constraints $ C_i $ are hidden, and the goal is to find a solution satisfying all constraints. We can adaptively propose a candidate solution (i.e., trial), and there is a verification oracle that either confirms that it is a valid solution, or returns the index i of a violated constraint (i.e., error), with the exact content of $ C_i $ still hidden. We studied the time and trial complexities of a number of natural CSPs, summarized as follows. On one hand, despite the seemingly very little information provided by the oracle, efficient algorithms do exist for Nash, Core, Stable Matching, and SAT problems, whose unknown-input versions are shown to be as hard as the corresponding known-input versions up to a factor of polynomial. The techniques employed vary considerably, including, e.g., order theory and the ellipsoid method with a strong separation oracle. On the other hand, there are problems whose complexities are substantially increased in the unknown-input model. In particular, no time-efficient algorithms exist for Graph Isomorphism and Group Isomorphism (unless PH collapses or P = NP). The proofs use quite nonstandard reductions, in which an efficient simulator is carefully designed to simulate a desirable but computationally unaffordable oracle. Our model investigates the value of input information, and our results demonstrate that the lack of input information can introduce various levels of extra difficulty. The model accommodates a wide range of combinatorial and algebraic structures, and exhibits intimate connections with (and hopefully can also serve as a useful supplement to) certain existing learning and complexity theories.", acknowledgement = ack-nhfb, } @InProceedings{Bhawalkar:2013:COF, author = "Kshipra Bhawalkar and Sreenivas Gollapudi and Kamesh Munagala", title = "Coevolutionary opinion formation games", crossref = "ACM:2013:SPF", pages = "41--50", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488615", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present game-theoretic models of opinion formation in social networks where opinions themselves co-evolve with friendships. In these models, nodes form their opinions by maximizing agreements with friends weighted by the strength of the relationships, which in turn depend on difference in opinion with the respective friends. We define a social cost of this process by generalizing recent work of Bindel et al., FOCS 2011. We tightly bound the price of anarchy of the resulting dynamics via local smoothness arguments, and characterize it as a function of how much nodes value their own (intrinsic) opinion, as well as how strongly they weigh links to friends with whom they agree more.", acknowledgement = ack-nhfb, } @InProceedings{Chawla:2013:PIM, author = "Shuchi Chawla and Jason D. Hartline and David Malec and Balasubramanian Sivan", title = "Prior-independent mechanisms for scheduling", crossref = "ACM:2013:SPF", pages = "51--60", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488616", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the makespan minimization problem with unrelated selfish machines under the assumption that job sizes are stochastic. We design simple truthful mechanisms that under different distributional assumptions provide constant and sublogarithmic approximations to expected makespan. Our mechanisms are prior-independent in that they do not rely on knowledge of the job size distributions. Prior-independent approximations were previously known only for the revenue maximization objective [13, 11, 26]. In contrast to our results, in prior-free settings no truthful anonymous deterministic mechanism for the makespan objective can provide a sublinear approximation [3].", acknowledgement = ack-nhfb, } @InProceedings{Feldman:2013:CWE, author = "Michal Feldman and Nick Gravin and Brendan Lucier", title = "Combinatorial {Walrasian Equilibrium}", crossref = "ACM:2013:SPF", pages = "61--70", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488617", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study a combinatorial market design problem, where a collection of indivisible objects is to be priced and sold to potential buyers subject to equilibrium constraints. The classic solution concept for such problems is Walrasian Equilibrium (WE), which provides a simple and transparent pricing structure that achieves optimal social welfare. The main weakness of the WE notion is that it exists only in very restrictive cases. To overcome this limitation, we introduce the notion of a Combinatorial Walrasian equilibium (CWE), a natural relaxation of WE. The difference between a CWE and a (non-combinatorial) WE is that the seller can package the items into indivisible bundles prior to sale, and the market does not necessarily clear. We show that every valuation profile admits a CWE that obtains at least half of the optimal (unconstrained) social welfare. Moreover, we devise a poly-time algorithm that, given an arbitrary allocation X, computes a CWE that achieves at least half of the welfare of X. Thus, the economic problem of finding a CWE with high social welfare reduces to the algorithmic problem of social-welfare approximation. In addition, we show that every valuation profile admits a CWE that extracts a logarithmic fraction of the optimal welfare as revenue. Finally, these results are complemented by strong lower bounds when the seller is restricted to using item prices only, which motivates the use of bundles. The strength of our results derives partly from their generality --- our results hold for arbitrary valuations that may exhibit complex combinations of substitutes and complements.", acknowledgement = ack-nhfb, } @InProceedings{Naor:2013:ERN, author = "Assaf Naor and Oded Regev and Thomas Vidick", title = "Efficient rounding for the noncommutative {Grothendieck} inequality", crossref = "ACM:2013:SPF", pages = "71--80", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488618", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The classical Grothendieck inequality has applications to the design of approximation algorithms for NP-hard optimization problems. We show that an algorithmic interpretation may also be given for a noncommutative generalization of the Grothendieck inequality due to Pisier and Haagerup. Our main result, an efficient rounding procedure for this inequality, leads to a constant-factor polynomial time approximation algorithm for an optimization problem which generalizes the Cut Norm problem of Frieze and Kannan, and is shown here to have additional applications to robust principle component analysis and the orthogonal Procrustes problem.", acknowledgement = ack-nhfb, } @InProceedings{Clarkson:2013:LRA, author = "Kenneth L. Clarkson and David P. Woodruff", title = "Low rank approximation and regression in input sparsity time", crossref = "ACM:2013:SPF", pages = "81--90", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488620", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We design a new distribution over $ \poly (r \epsilon^{-1}) \times n $ matrices $S$ so that for any fixed $ n \times d $ matrix $A$ of rank $r$, with probability at least $ 9 / 10 $, $ S A x_2 = (1 \pm \epsilon) A x_2 $ simultaneously for all $ x \in R^d $. Such a matrix $S$ is called a subspace embedding. Furthermore, $ S A $ can be computed in $ O({\rm nnz}(A)) + \tilde O (r^2 \epsilon^{-2}) $ time, where $ {\rm nnz}(A) $ is the number of non-zero entries of $A$. This improves over all previous subspace embeddings, which required at least $ \Omega (n d \log d) $ time to achieve this property. We call our matrices $S$ sparse embedding matrices. Using our sparse embedding matrices, we obtain the fastest known algorithms for overconstrained least-squares regression, low-rank approximation, approximating all leverage scores, and $ l_p $ -regression: to output an $ x' $ for which $ A x' - b_2 \leq (1 + \epsilon) \min_x A x - b_2 $ for an $ n \times d $ matrix $A$ and an $ n \times 1 $ column vector $b$, we obtain an algorithm running in $ O({\rm nnz}(A)) + \tilde O(d^3 \epsilon^{-2}) $ time, and another in $ O({\rm nnz}(A) \log (1 / \epsilon)) + \tilde O(d^3 \log (1 / \epsilon)) $ time. (Here $ \tilde O(f) = f \cdot l o g^{O(1)} (f) $.) to obtain a decomposition of an $ n \times n $ matrix $A$ into a product of an $ n \times k $ matrix $L$, a $ k \times k $ diagonal matrix $D$, and a $ n \times k $ matrix $W$, for which $ \{ F A - L D W \} \leq (1 + \epsilon) F \{ A - A_k \} $, where $ A_k $ is the best rank-$k$ approximation, our algorithm runs in $ O({\rm nnz}(A)) + \tilde O(n k^2 \epsilon^{-4} \log n + k^3 \epsilon^{-5} \log^2 n) $ time. to output an approximation to all leverage scores of an $ n \times d $ input matrix $A$ simultaneously, with constant relative error, our algorithms run in $ O({\rm nnz}(A) \log n) + \tilde O(r^3) $ time. to output an $ x' $ for which $ A x' - b_p \leq (1 + \epsilon) \min_x A x - b_p $ for an $ n \times d $ matrix $A$ and an $ n \times 1 $ column vector $b$, we obtain an algorithm running in $ O({\rm nnz}(A) \log n) + \poly (r \epsilon^{-1}) $ time, for any constant $ 1 \leq p < \infty $. We optimize the polynomial factors in the above stated running times, and show various tradeoffs. Finally, we provide preliminary experimental results which suggest that our algorithms are of interest in practice.", acknowledgement = ack-nhfb, } @InProceedings{Meng:2013:LDS, author = "Xiangrui Meng and Michael W. Mahoney", title = "Low-distortion subspace embeddings in input-sparsity time and applications to robust linear regression", crossref = "ACM:2013:SPF", pages = "91--100", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488621", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Low-distortion embeddings are critical building blocks for developing random sampling and random projection algorithms for common linear algebra problems. We show that, given a matrix $ A \in R^{n x d} $ with $ n \gg d $ and $ a p \in [1, 2) $, with a constant probability, we can construct a low-distortion embedding matrix $ \Pi \in R^{O({\rm poly}(d)) x n} $ that embeds $ A_p $, the $ l_p $ subspace spanned by $A$'s columns, into $ (R^{O({\rm poly}(d))}, | \cdot |_p) $; the distortion of our embeddings is only $ O({\rm poly}(d)) $, and we can compute $ \Pi A $ in $ O(n n z(A)) $ time, i.e., input-sparsity time. Our result generalizes the input-sparsity time $ l_2 $ subspace embedding by Clarkson and Woodruff [STOC'13]; and for completeness, we present a simpler and improved analysis of their construction for $ l_2 $. These input-sparsity time $ l_p $ embeddings are optimal, up to constants, in terms of their running time; and the improved running time propagates to applications such as $ (1 p m \epsilon) $-distortion $ l_p $ subspace embedding and relative-error $ l_p $ regression. For $ l_2 $, we show that a $ (1 + \epsilon) $-approximate solution to the $ l_2 $ regression problem specified by the matrix $A$ and a vector $ b \in R^n $ can be computed in $ O(n n z(A) + d^3 \log (d / \epsilon) / \epsilon^2) $ time; and for $ l_p $, via a subspace-preserving sampling procedure, we show that a $ (1 p m \epsilon) $-distortion embedding of $ A_p $ into $ R^{O({\rm poly}(d))} $ can be computed in $ O(n n z(A) \cdot \log n) $ time, and we also show that a $ (1 + \epsilon) $-approximate solution to the $ l_p $ regression problem $ \min_{x \in R^d} |A x - b|_p $ can be computed in $ O(n n z(A) \cdot \log n + {\rm poly}(d) \log (1 / \epsilon) / \epsilon^2) $ time. Moreover, we can also improve the embedding dimension or equivalently the sample size to $ O(d^{3 + p / 2} \log (1 / \epsilon) / \epsilon^2) $ without increasing the complexity.", acknowledgement = ack-nhfb, } @InProceedings{Nelson:2013:SLB, author = "Jelani Nelson and Huy L. Nguyen", title = "Sparsity lower bounds for dimensionality reducing maps", crossref = "ACM:2013:SPF", pages = "101--110", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488622", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give near-tight lower bounds for the sparsity required in several dimensionality reducing linear maps. First, consider the Johnson--Lindenstrauss (JL) lemma which states that for any set of $n$ vectors in $ R^d $ there is an $ A \in R^{m \times d} $ with $ m = O(\epsilon^{-2} \log n) $ such that mapping by $A$ preserves the pairwise Euclidean distances up to a $ 1 \pm \epsilon $ factor. We show there exists a set of $n$ vectors such that any such $A$ with at most $s$ non-zero entries per column must have $ s = \Omega (\epsilon^{-1} \log n / \log (1 / \epsilon)) $ if $ m < O(n / \log (1 / \epsilon)) $. This improves the lower bound of $ \Omega (\min \{ \epsilon^{-2}, \epsilon^{-1} \sqrt (\log_m d) \}) $ by [Dasgupta-Kumar-Sarlos, STOC 2010], which only held against the stronger property of distributional JL, and only against a certain restricted class of distributions. Meanwhile our lower bound is against the JL lemma itself, with no restrictions. Our lower bound matches the sparse JL upper bound of [Kane-Nelson, SODA 2012] up to an $ O(\log (1 / \epsilon)) $ factor. Next, we show that any $ m \times n $ matrix with the $k$-restricted isometry property (RIP) with constant distortion must have $ \Omega (k \log (n / k)) $ non-zeroes per column if $ m = O(k \log (n / k)) $, the optimal number of rows for RIP, and $ k < n / \polylog n $. This improves the previous lower bound of $ \Omega (\min \{ k, n / m \}) $ by [Chandar, 2010] and shows that for most $k$ it is impossible to have a sparse RIP matrix with an optimal number of rows. Both lower bounds above also offer a tradeoff between sparsity and the number of rows. Lastly, we show that any oblivious distribution over subspace embedding matrices with 1 non-zero per column and preserving distances in a $d$ dimensional-subspace up to a constant factor must have at least $ \Omega (d^2) $ rows. This matches an upper bound in [Nelson-Nguy{\^e}n, arXiv abs/1211.1002] and shows the impossibility of obtaining the best of both of constructions in that work, namely 1 non-zero per column and $ d \cdot \polylog d $ rows.", acknowledgement = ack-nhfb, } @InProceedings{Bitansky:2013:RCB, author = "Nir Bitansky and Ran Canetti and Alessandro Chiesa and Eran Tromer", title = "Recursive composition and bootstrapping for {SNARKS} and proof-carrying data", crossref = "ACM:2013:SPF", pages = "111--120", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488623", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Succinct non-interactive arguments of knowledge (SNARKs) enable verifying NP statements with complexity that is essentially independent of that required for classical NP verification. In particular, they provide strong solutions to the problem of verifiably delegating computation. We construct the first fully-succinct publicly-verifiable SNARK. To do that, we first show how to ``bootstrap'' any SNARK that requires expensive preprocessing to obtain a SNARK that does not, while preserving public verifiability. We then apply this transformation to known SNARKs with preprocessing. Moreover, the SNARK we construct only requires of the prover time and space that are essentially the same as that required for classical NP verification. Our transformation assumes only collision-resistant hashing; curiously, it does not rely on PCPs. We also show an analogous transformation for privately-verifiable SNARKs, assuming fully-homomorphic encryption. At the heart of our transformations is a technique for recursive composition of SNARKs. This technique uses in an essential way the proof-carrying data (PCD) framework, which extends SNARKs to the setting of distributed networks of provers and verifiers. Concretely, to bootstrap a given SNARK, we recursively compose the SNARK to obtain a ``weak'' PCD system for shallow distributed computations, and then use the PCD framework to attain stronger notions of SNARKs and PCD systems.", acknowledgement = ack-nhfb, } @InProceedings{Hardt:2013:HRL, author = "Moritz Hardt and David P. Woodruff", title = "How robust are linear sketches to adaptive inputs?", crossref = "ACM:2013:SPF", pages = "121--130", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488624", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Linear sketches are powerful algorithmic tools that turn an $n$-dimensional input into a concise lower-dimensional representation via a linear transformation. Such sketches have seen a wide range of applications including norm estimation over data streams, compressed sensing, and distributed computing. In almost any realistic setting, however, a linear sketch faces the possibility that its inputs are correlated with previous evaluations of the sketch. Known techniques no longer guarantee the correctness of the output in the presence of such correlations. We therefore ask: Are linear sketches inherently non-robust to adaptively chosen inputs? We give a strong affirmative answer to this question. Specifically, we show that no linear sketch approximates the Euclidean norm of its input to within an arbitrary multiplicative approximation factor on a polynomial number of adaptively chosen inputs. The result remains true even if the dimension of the sketch is d=n-o(n) and the sketch is given unbounded computation time. Our result is based on an algorithm with running time polynomial in d that adaptively finds a distribution over inputs on which the sketch is incorrect with constant probability. Our result implies several corollaries for related problems including l$_p$ -norm estimation and compressed sensing. Notably, we resolve an open problem in compressed sensing regarding the feasibility of l$_2$ /l$_2$ -recovery guarantees in presence of computationally bounded adversaries.", acknowledgement = ack-nhfb, } @InProceedings{Bohm:2013:EDO, author = "Stanislav B{\"o}hm and Stefan G{\"o}ller and Petr Jancar", title = "Equivalence of deterministic one-counter automata is {NL}-complete", crossref = "ACM:2013:SPF", pages = "131--140", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488626", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove that language equivalence of deterministic one-counter automata is NL-complete. This improves the superpolynomial time complexity upper bound shown by Valiant and Paterson in 1975. Our main contribution is to prove that two deterministic one-counter automata are inequivalent if and only if they can be distinguished by a word of length polynomial in the size of the two input automata.", acknowledgement = ack-nhfb, } @InProceedings{Burgisser:2013:ELB, author = "Peter B{\"u}rgisser and Christian Ikenmeyer", title = "Explicit lower bounds via geometric complexity theory", crossref = "ACM:2013:SPF", pages = "141--150", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488627", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove the lower bound $ R (M_m) \geq 3 / 2 m^2 - 2 $ on the border rank of $ m \times m $ matrix multiplication by exhibiting explicit representation theoretic (occurrence) obstructions in the sense of Mulmuley and Sohoni's geometric complexity theory (GCT) program. While this bound is weaker than the one recently obtained by Landsberg and Ottaviani, these are the first significant lower bounds obtained within the GCT program. Behind the proof is an explicit description of the highest weight vectors in Sym$^d \otimes^3 (C^n)*$ in terms of combinatorial objects, called obstruction designs. This description results from analyzing the process of polarization and Schur--Weyl duality.", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2013:IEC, author = "Mark Braverman and Ankit Garg and Denis Pankratov and Omri Weinstein", title = "From information to exact communication", crossref = "ACM:2013:SPF", pages = "151--160", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488628", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We develop a new local characterization of the zero-error information complexity function for two-party communication problems, and use it to compute the exact internal and external information complexity of the 2-bit AND function: $ {\rm IC}({\rm AND}, 0) = C_{\wedge } \cong 1.4923 $ bits, and $ {\rm IC}^{ext} ({\rm AND}, 0) = \log_2 3 \cong 1.5839 $ bits. This leads to a tight (upper and lower bound) characterization of the communication complexity of the set intersection problem on subsets of $ \{ 1, \ldots {}, n \} $ (the player are required to compute the intersection of their sets), whose randomized communication complexity tends to $ C_{\wedge } \cdot n \pm o(n) $ as the error tends to zero. The information-optimal protocol we present has an infinite number of rounds. We show this is necessary by proving that the rate of convergence of the r-round information cost of AND to $ {\rm IC}({\rm AND}, 0) = C_{\wedge } $ behaves like $ \Theta (1 / r^2) $, i.e. that the $r$-round information complexity of AND is $ C_\wedge + \Theta (1 / r^2) $. We leverage the tight analysis obtained for the information complexity of AND to calculate and prove the exact communication complexity of the set disjointness function $ {\rm Disj}_n (X, Y) = - v_{i = 1}^n {\rm AND}(x_i, y_i) $ with error tending to $0$, which turns out to be $ = C_{\rm DISJ} \cdot n \pm o(n) $, where $ C_{\rm DISJ} \cong 0.4827 $. Our rate of convergence results imply that an asymptotically optimal protocol for set disjointness will have to use $ \omega (1) $ rounds of communication, since every $r$-round protocol will be sub-optimal by at least $ \Omega (n / r^2) $ bits of communication. We also obtain the tight bound of $ 2 / \ln 2 k \pm o(k) $ on the communication complexity of disjointness of sets of size $ \leq k $. An asymptotic bound of $ \Theta (k) $ was previously shown by Hastad and Wigderson.", acknowledgement = ack-nhfb, } @InProceedings{Braverman:2013:ICA, author = "Mark Braverman and Ankur Moitra", title = "An information complexity approach to extended formulations", crossref = "ACM:2013:SPF", pages = "161--170", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488629", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove an unconditional lower bound that any linear program that achieves an $ O(n^{1 - \epsilon }) $ approximation for clique has size $ 2^{ \Omega (n \epsilon)} $. There has been considerable recent interest in proving unconditional lower bounds against any linear program. Fiorini et al. proved that there is no polynomial sized linear program for traveling salesman. Braun et al. proved that there is no polynomial sized $ O(n^{1 / 2 - \epsilon }) $-approximate linear program for clique. Here we prove an optimal and unconditional lower bound against linear programs for clique that matches Hastad's celebrated hardness result. Interestingly, the techniques used to prove such lower bounds have closely followed the progression of techniques used in communication complexity. Here we develop an information theoretic framework to approach these questions, and we use it to prove our main result. Also we resolve a related question: How many bits of communication are needed to get $ \epsilon $ advantage over random guessing for disjointness? Kalyanasundaram and Schnitger proved that a protocol that gets constant advantage requires $ \Omega (n) $ bits of communication. This result in conjunction with amplification implies that any protocol that gets $ \epsilon $-advantage requires $ \Omega (\epsilon^2 n) $ bits of communication. Here we improve this bound to $ \Omega (\epsilon n) $, which is optimal for any $ \epsilon > 0 $.", acknowledgement = ack-nhfb, } @InProceedings{Komargodski:2013:ACL, author = "Ilan Komargodski and Ran Raz", title = "Average-case lower bounds for formula size", crossref = "ACM:2013:SPF", pages = "171--180", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488630", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give an explicit function $ h : \{ 0, 1 \}^n \to \{ 0, 1 \} $ such that any deMorgan formula of size $ O(n^{2.499}) $ agrees with $h$ on at most $ 1 / 2 + \epsilon $ fraction of the inputs, where $ \epsilon $ is exponentially small (i.e. $ \epsilon = 2^{-n \Omega (1)} $ ). We also show, using the same technique, that any boolean formula of size $ O(n^{1.999}) $ over the complete basis, agrees with $h$ on at most $ 1 / 2 + \epsilon $ fraction of the inputs, where $ \epsilon $ is exponentially small (i.e. $ \epsilon = 2^{-vn \Omega (1)} $ ). Our construction is based on Andreev's $ \Omega (n^{2.5 - o(1)}) $ formula size lower bound that was proved for the case of exact computation.", acknowledgement = ack-nhfb, } @InProceedings{Chen:2013:CNM, author = "Xi Chen and Dimitris Paparas and Mihalis Yannakakis", title = "The complexity of non-monotone markets", crossref = "ACM:2013:SPF", pages = "181--190", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488632", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We introduce the notion of non-monotone utilities, which covers a wide variety of utility functions in economic theory. We show that it is PPAD-hard to compute an approximate Arrow--Debreu market equilibrium in markets with linear and non-monotone utilities. Building on this result, we settle the long-standing open problem regarding the computation of an approximate Arrow--Debreu market equilibrium in markets with CES utilities, by proving that it is PPAD-complete when the Constant Elasticity of Substitution parameter, $ \rho $, is any constant less than $ - 1 $.", acknowledgement = ack-nhfb, } @InProceedings{Cheung:2013:TBG, author = "Yun Kuen Cheung and Richard Cole and Nikhil Devanur", title = "Tatonnement beyond gross substitutes?: gradient descent to the rescue", crossref = "ACM:2013:SPF", pages = "191--200", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488633", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Tatonnement is a simple and natural rule for updating prices in Exchange (Arrow--Debreu) markets. In this paper we define a class of markets for which tatonnement is equivalent to gradient descent. This is the class of markets for which there is a convex potential function whose gradient is always equal to the negative of the excess demand and we call it Convex Potential Function (CPF) markets. We show the following results. CPF markets contain the class of Eisenberg Gale (EG) markets, defined previously by Jain and Vazirani. The subclass of CPF markets for which the demand is a differentiable function contains exactly those markets whose demand function has a symmetric negative semi-definite Jacobian. We define a family of continuous versions of tatonnement based on gradient descent using a Bregman divergence. As we show, all processes in this family converge to an equilibrium for any CPF market. This is analogous to the classic result for markets satisfying the Weak Gross Substitutes property. A discrete version of tatonnement converges toward the equilibrium for the following markets of complementary goods; its convergence rate for these settings is analyzed using a common potential function. Fisher markets in which all buyers have Leontief utilities. The tatonnement process reduces the distance to the equilibrium, as measured by the potential function, to an $ \epsilon $ fraction of its initial value in $ O(1 / \epsilon) $ rounds of price updates. Fisher markets in which all buyers have complementary CES utilities. Here, the distance to the equilibrium is reduced to an $ \epsilon $ fraction of its initial value in $ O(\log (1 / \epsilon)) $ rounds of price updates. This shows that tatonnement converges for the entire range of Fisher markets when buyers have complementary CES utilities, in contrast to prior work, which could analyze only the substitutes range, together with a small portion of the complementary range.", acknowledgement = ack-nhfb, } @InProceedings{Feldman:2013:SAA, author = "Michal Feldman and Hu Fu and Nick Gravin and Brendan Lucier", title = "Simultaneous auctions are (almost) efficient", crossref = "ACM:2013:SPF", pages = "201--210", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488634", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Simultaneous item auctions are simple and practical procedures for allocating items to bidders with potentially complex preferences. In a simultaneous auction, every bidder submits independent bids on all items simultaneously. The allocation and prices are then resolved for each item separately, based solely on the bids submitted on that item. We study the efficiency of Bayes-Nash equilibrium (BNE) outcomes of simultaneous first- and second-price auctions when bidders have complement-free (a.k.a. subadditive) valuations. While it is known that the social welfare of every pure Nash equilibrium (NE) constitutes a constant fraction of the optimal social welfare, a pure NE rarely exists, and moreover, the full information assumption is often unrealistic. Therefore, quantifying the welfare loss in Bayes-Nash equilibria is of particular interest. Previous work established a logarithmic bound on the ratio between the social welfare of a BNE and the expected optimal social welfare in both first-price auctions (Hassidim et al., 2011) and second-price auctions (Bhawalkar and Roughgarden, 2011), leaving a large gap between a constant and a logarithmic ratio. We introduce a new proof technique and use it to resolve both of these gaps in a unified way. Specifically, we show that the expected social welfare of any BNE is at least 1/2 of the optimal social welfare in the case of first-price auctions, and at least 1/4 in the case of second-price auctions.", acknowledgement = ack-nhfb, } @InProceedings{Syrgkanis:2013:CEM, author = "Vasilis Syrgkanis and Eva Tardos", title = "Composable and efficient mechanisms", crossref = "ACM:2013:SPF", pages = "211--220", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488635", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We initiate the study of efficient mechanism design with guaranteed good properties even when players participate in multiple mechanisms simultaneously or sequentially. We define the class of smooth mechanisms, related to smooth games defined by Roughgarden, that can be thought of as mechanisms that generate approximately market clearing prices. We show that smooth mechanisms result in high quality outcome both in equilibrium and in learning outcomes in the full information setting, as well as in Bayesian equilibrium with uncertainty about participants. Our main result is to show that smooth mechanisms compose well: smoothness locally at each mechanism implies global efficiency. For mechanisms where good performance requires that bidders do not bid above their value, we identify the notion of a weakly smooth mechanism. Weakly smooth mechanisms, such as the Vickrey auction, are approximately efficient under the no-overbidding assumption, and the weak smoothness property is also maintained by composition. In most of the paper we assume participants have quasi-linear valuations. We also extend some of our results to settings where participants have budget constraints.", acknowledgement = ack-nhfb, } @InProceedings{Goyal:2013:NBB, author = "Vipul Goyal", title = "Non-black-box simulation in the fully concurrent setting", crossref = "ACM:2013:SPF", pages = "221--230", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488637", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present a new zero-knowledge argument protocol by relying on the non-black-box simulation technique of Barak (FOCS'01). Similar to the protocol of Barak, ours is public-coin, is based on the existence of collision-resistant hash functions, and, is not based on ``rewinding techniques'' but rather uses non-black-box simulation. However in contrast to the protocol of Barak, our protocol is secure even if there are any unbounded (polynomial) number of concurrent sessions. This gives us the first construction of public-coin concurrent zero-knowledge. Prior to our work, Pass, Tseng and Wikstrom (SIAM J. Comp. 2011) had shown that using black-box simulation, getting a construction for even public-coin parallel zero-knowledge is impossible. A public-coin concurrent zero-knowledge protocol directly implies the existence of a concurrent resettably-sound zero-knowledge protocol. This is an improvement over the corresponding construction of Deng, Goyal and Sahai (FOCS'09) which was based on stronger assumptions. Furthermore, this also directly leads to an alternative (and arguable cleaner) construction of a simultaneous resettable zero-knowledge argument system. An important feature of our protocol is the existence of a ``straight-line'' simulator. This gives a fundamentally different tool for constructing concurrently secure computation protocols (for functionalities even beyond zero-knowledge). The round complexity of our protocol is $ n^\epsilon $ (for any constant $ \epsilon > 0 $ ), and, the simulator runs in strict polynomial time. The main technique behind our construction is purely combinatorial in nature.", acknowledgement = ack-nhfb, } @InProceedings{Chung:2013:NBB, author = "Kai-Min Chung and Rafael Pass and Karn Seth", title = "Non-black-box simulation from one-way functions and applications to resettable security", crossref = "ACM:2013:SPF", pages = "231--240", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488638", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The simulation paradigm, introduced by Goldwasser, Micali and Rackoff, is of fundamental importance to modern cryptography. In a breakthrough work from 2001, Barak (FOCS'01) introduced a novel non-black-box simulation technique. This technique enabled the construction of new cryptographic primitives, such as resettably-sound zero-knowledge arguments, that cannot be proven secure using just black-box simulation techniques. The work of Barak and its follow-ups, however, all require stronger cryptographic hardness assumptions than the minimal assumption of one-way functions. In this work, we show how to perform non-black-box simulation assuming just the existence of one-way functions. In particular, we demonstrate the existence of a constant-round resettably-sound zero-knowledge argument based only on the existence of one-way functions. Using this technique, we determine necessary and sufficient assumptions for several other notions of resettable security of zero-knowledge proofs. An additional benefit of our approach is that it seemingly makes practical implementations of non-black-box zero-knowledge viable.", acknowledgement = ack-nhfb, } @InProceedings{Bitansky:2013:IAO, author = "Nir Bitansky and Omer Paneth", title = "On the impossibility of approximate obfuscation and applications to resettable cryptography", crossref = "ACM:2013:SPF", pages = "241--250", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488639", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The traditional notion of program obfuscation requires that an obfuscation \char`{\~P}rog of a program Prog computes the exact same function as Prog, but beyond that, the code of \char`{\~P}rog should not leak any information about Prog. This strong notion of virtual black-box security was shown by Barak et al. (CRYPTO 2001) to be impossible to achieve, for certain unobfuscatable function families. The same work raised the question of approximate obfuscation, where the obfuscated \char`{\~P}rog is only required to approximate Prog; that is, \char`{\~P}rog only agrees with Prog with high enough probability on some input distribution. We show that, assuming trapdoor permutations, there exist families of robust unobfuscatable functions for which even approximate obfuscation is impossible. Specifically, obfuscation is impossible even if the obfuscated \char`{\~P}rog is only required to agree with Prog with probability slightly more than 1/2, on a uniformly sampled input (below 1/2-agreement, the function obfuscated by \char`{\~P}rog is not uniquely defined). Additionally, assuming only one-way functions, we rule out approximate obfuscation where \char`{\~P}rog may output bot with probability close to $1$, but otherwise must agree with Prog. We demonstrate the power of robust unobfuscatable functions by exhibiting new implications to resettable protocols. Concretely, we reduce the assumptions required for resettably-sound zero-knowledge to one-way functions, as well as reduce round-complexity. We also present a new simplified construction of a simultaneously-resettable zero-knowledge protocol. Finally, we construct a three-message simultaneously-resettable witness-indistinguishable argument of knowledge (with a non-black-box knowledge extractor). Our constructions use a new non-black-box simulation technique that is based on a special kind of ``resettable slots''. These slots are useful for a non-black-box simulator, but not for a resetting prover.", acknowledgement = ack-nhfb, } @InProceedings{Miles:2013:SCG, author = "Eric Miles and Emanuele Viola", title = "Shielding circuits with groups", crossref = "ACM:2013:SPF", pages = "251--260", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488640", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show how to efficiently compile any given circuit C into a leakage-resistant circuit C' such that any function on the wires of C' that leaks information during a computation C'(x) yields advantage in computing the product of |C'|$^{ \Omega (1)}$ elements of the alternating group A$_u$. In combination with new compression bounds for A$_u$ products, also obtained here, C' withstands leakage from virtually any class of functions against which average-case lower bounds are known. This includes communication protocols, and A$ C^0 $ circuits augmented with few arbitrary symmetric gates. If N$ C^1 $ ' T$ C^0 $ then the construction resists T$ C^0 $ leakage as well. We also conjecture that our construction resists N$ C^1 $ leakage. In addition, we extend the construction to the multi-query setting by relying on a simple secure hardware component. We build on Barrington's theorem [JCSS '89] and on the previous leakage-resistant constructions by Ishai et al. [Crypto '03] and Faust et al. [Eurocrypt '10]. Our construction exploits properties of A$_u$ beyond what is sufficient for Barrington's theorem.", acknowledgement = ack-nhfb, } @InProceedings{Babai:2013:QTC, author = "Laszlo Babai and John Wilmes", title = "Quasipolynomial-time canonical form for {Steiner} designs", crossref = "ACM:2013:SPF", pages = "261--270", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488642", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A Steiner 2-design is a finite geometry consisting of a set of ``points'' together with a set of ``lines'' (subsets of points of uniform cardinality) such that each pair of points belongs to exactly one line. In this paper we analyse the individualization/refinement heuristic and conclude that after individualizing $ O(\log n) $ points (assigning individual colors to them), the refinement process gives each point an individual color. The following consequences are immediate: (a) isomorphism of Steiner 2-designs can be tested in $ n^{O(\log n)} $ time, where $n$ is the number of lines; (b) a canonical form of Steiner 2-designs can be computed within the same time bound; (c) all isomorphisms between two Steiner 2-designs can be listed within the same time bound; (d) the number of automorphisms of a Steiner 2-design is at most n$^{O(\log n)}$ (a fact of interest to finite geometry and group theory.) The best previous bound in each of these four statements was moderately exponential, $ \exp (\tilde O(n^{1 / 4})) $ (Spielman, STOC'96). Our result removes an exponential bottleneck from Spielman's analysis of the Graph Isomorphism problem for strongly regular graphs. The results extend to Steiner $t$-designs for all $ t \geq 2 $. Strongly regular (s.r.) graphs have been known as hard cases for graph isomorphism testing; the best previously known bound for this case is moderately exponential, $ \exp (\tilde O(n^{1 / 3})) $ where $n$ is the number of vertices (Spielman, STOC'96). Line graphs of Steiner $2$-designs enter as a critical subclass via Neumaier's 1979 classification of s.r. graphs. Previously, $ n^{O(\log n)} $ isomorphism testing and canonical forms for Steiner 2-designs was known for the case when the lines of the Steiner 2-design have bounded length (Babai and Luks, STOC'83). That paper relied on Luks's group-theoretic divide-and-conquer algorithms and did not yield a subexponential bound on the number of automorphisms. To analyse the individualization/refinement heuristic, we develop a new structure theory of Steiner 2-designs based on the analysis of controlled growth and on an addressing scheme that produces a hierarchy of increasing sets of pairwise independent, uniformly distributed points. This scheme represents a new expression of the structural homogeneity of Steiner 2-designs that allows applications of the second moment method. We also address the problem of reconstruction of Steiner 2-designs from their line-graphs beyond the point of unique reconstructability, in a manner analogous to list-decoding, and as a consequence achieve an $ \exp (\tilde O(n^{1 / 6})) $ bound for isomorphism testing for this class of s.r. graphs. Results, essentially identical to our main results, were obtained simultaneously by Xi Chen, Xiaorui Sun, and Shang-Hua Teng, building on a different philosophy and combinatorial structure theory than the present paper. They do not claim an analysis of the individualization/refinement algorithm but of a more complex combinatorial algorithm. We comment on how this paper fits into the overall project of improved isomorphism testing for strongly regular graphs (the ultimate goal being subexponential $ \exp (n^{o(1)}) $ time). In the remaining cases we need to deal with s.r. graphs satisfying ``Neumaier's claw bound,'' permitting the use of a separate set of asymptotic structural tools. In joint work (in progress) with Chen, Sun, and Teng, we address that case and have already pushed the overall bound below $ \exp (\tilde O(n^{1 / 4})) $ The present paper is a methodologically distinct and stand-alone part of the overall project.", acknowledgement = ack-nhfb, } @InProceedings{Chen:2013:MSD, author = "Xi Chen and Xiaorui Sun and Shang-Hua Teng", title = "Multi-stage design for quasipolynomial-time isomorphism testing of {Steiner} $2$-systems", crossref = "ACM:2013:SPF", pages = "271--280", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488643", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A standard heuristic for testing graph isomorphism is to first assign distinct labels to a small set of vertices of an input graph, and then propagate to create new vertex labels across the graph, aiming to assign distinct and isomorphism-invariant labels to all vertices in the graph. This is usually referred to as the individualization/refinement method for canonical labeling of graphs. We present a quasipolynomial-time algorithm for isomorphism testing of Steiner 2-systems. A Steiner 2-system consists of points and lines, where each line passes the same number of points and each pair of points uniquely determines a line. Each Steiner 2-system induces a Steiner graph, in which vertices represent lines and edges represent intersections of lines. Steiner graphs are an important subfamily of strongly regular graphs whose isomorphism testing has challenged researchers for years. Inspired by both the individualization/refinement method and the previous analyses of Babai and Spielman, we consider an extended framework for isomorphism testing of Steiner 2-systems, in which we use a small set of randomly chosen points and lines to build isomorphism-invariant multi-stage combinatorial structures that are sufficient to distinguish all pairs of points of a Steiner 2-system. Applying this framework, we show that isomorphism of Steiner 2-systems with $n$ lines can be tested in time $\smash n^{O(\log n)}$, improving the previous best bound of $ \smash \exp (\tilde O(n^{1 / 4})) $ by Spielman. Before our result, quasipolynomial-time isomorphism testing was only known for the case when the line size is polylogarithmic, as shown by Babai and Luks. A result essentially identical to ours was obtained simultaneously by Laszlo Babai and John Wilmes. They performed a direct analysis of the individualization/refinement method, building on a different philosophy and combinatorial structure theory. We comment on how this paper fits into the overall project of improved isomorphism testing for strongly regular graphs (the ultimate goal being subexponential $ \exp (n^{o(1)}) $ time). In the remaining cases, we only need to deal with strongly regular graphs satisfying ``Neumaier's claw bound,'' permitting the use of a separate set of asymptotic structural tools. In joint work (in progress) with Babai and Wilmes, we address that case and have already pushed the overall bound below $ \smash \exp (\tilde O(n^{1 / 4})) $. The present paper is a methodologically distinct and stand-alone part of the overall project.", acknowledgement = ack-nhfb, } @InProceedings{Gupta:2013:SCB, author = "Anupam Gupta and Kunal Talwar and David Witmer", title = "Sparsest cut on bounded treewidth graphs: algorithms and hardness results", crossref = "ACM:2013:SPF", pages = "281--290", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488644", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give a 2-approximation algorithm for the non-uniform Sparsest Cut problem that runs in time $ n^{O(k)} $, where $k$ is the treewidth of the graph. This improves on the previous $ 2^{2 k} $ approximation in time $ \poly (n) 2^{O(k)} $ due to Chlamtac et al. [18]. To complement this algorithm, we show the following hardness results: If the non-uniform Sparsest Cut has a $ \rho $-approximation for series-parallel graphs (where $ \rho \geq 1 $ ), then the MaxCut problem has an algorithm with approximation factor arbitrarily close to $ 1 / \rho $. Hence, even for such restricted graphs (which have treewidth 2), the Sparsest Cut problem is NP-hard to approximate better than $ 17 / 16 - \epsilon $ for $ \epsilon > 0 $; assuming the Unique Games Conjecture the hardness becomes $ 1 / \alpha_{GW} - \epsilon $. For graphs with large (but constant) treewidth, we show a hardness result of $ 2 - \epsilon $ assuming the Unique Games Conjecture. Our algorithm rounds a linear program based on (a subset of) the Sherali--Adams lift of the standard Sparsest Cut LP. We show that even for treewidth-2 graphs, the LP has an integrality gap close to 2 even after polynomially many rounds of Sherali--Adams. Hence our approach cannot be improved even on such restricted graphs without using a stronger relaxation.", acknowledgement = ack-nhfb, } @InProceedings{Chekuri:2013:LTG, author = "Chandra Chekuri and Julia Chuzhoy", title = "Large-treewidth graph decompositions and applications", crossref = "ACM:2013:SPF", pages = "291--300", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488645", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Treewidth is a graph parameter that plays a fundamental role in several structural and algorithmic results. We study the problem of decomposing a given graph $G$ into node-disjoint subgraphs, where each subgraph has sufficiently large treewidth. We prove two theorems on the tradeoff between the number of the desired subgraphs h, and the desired lower bound r on the treewidth of each subgraph. The theorems assert that, given a graph $G$ with treewidth $k$, a decomposition with parameters $ h, r $ is feasible whenever $ h r^2 \leq k / \polylog (k) $, or $ h^3 r \leq k / \polylog (k) $ holds. We then show a framework for using these theorems to bypass the well-known Grid-Minor Theorem of Robertson and Seymour in some applications. In particular, this leads to substantially improved parameters in some Erd{\H{o}}s--Posa-type results, and faster algorithms for some fixed-parameter tractable problems.", acknowledgement = ack-nhfb, } @InProceedings{Cygan:2013:FHC, author = "Marek Cygan and Stefan Kratsch and Jesper Nederlof", title = "Fast {Hamiltonicity} checking via bases of perfect matchings", crossref = "ACM:2013:SPF", pages = "301--310", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488646", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "For an even integer $ t \geq 2 $, the Matching Connectivity matrix $ H_t $ is a matrix that has rows and columns both labeled by all perfect matchings of the complete graph $ K_t $ on $t$ vertices; an entry $ H_t [M_1, M_2] $ is $1$ if $ M_1 \cup M_2 $ is a Hamiltonian cycle and $0$ otherwise. Motivated by the computational study of the Hamiltonicity problem, we present three results on the structure of $ H_t $: We first show that $ H_t $ has rank exactly $ 2^{t / 2 - 1} $ over GF(2) via an appropriate factorization that explicitly provides families of matchings $ X_t $ forming bases for $ H_t $. Second, we show how to quickly change representation between such bases. Third, we notice that the sets of matchings $ X_t $ induce permutation matrices within $ H_t $. We use the factorization to derive an $ 1.888^n n^{O(1)} $ time Monte Carlo algorithm that solves the Hamiltonicity problem in directed bipartite graphs. Our algorithm as well counts the number of Hamiltonian cycles modulo two in directed bipartite or undirected graphs in the same time bound. Moreover, we use the fast basis change algorithm from the second result to present a Monte Carlo algorithm that given an undirected graph on $n$ vertices along with a path decomposition of width at most pw decides Hamiltonicity in $ (2 + \sqrt 2)^{pw} n^{O(1)} $ time. Finally, we use the third result to show that for every $ \epsilon > 0 $ this cannot be improved to $ (2 + \sqrt 2 - \epsilon)^{pw} n^{O(1)} $ time unless the Strong Exponential Time Hypothesis fails, i.e., a faster algorithm for this problem would imply the breakthrough result of an $ O((2 - \epsilon ')^n) $ time algorithm for CNF-Sat.", acknowledgement = ack-nhfb, } @InProceedings{Keevash:2013:PTP, author = "Peter Keevash and Fiachra Knox and Richard Mycroft", title = "Polynomial-time perfect matchings in dense hypergraphs", crossref = "ACM:2013:SPF", pages = "311--320", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488647", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Let H be a $k$-graph on $n$ vertices, with minimum codegree at least n/k + cn for some fixed c > 0. In this paper we construct a polynomial-time algorithm which finds either a perfect matching in H or a certificate that none exists. This essentially solves a problem of Karpinski, Rucinski and Szymanska, who previously showed that this problem is NP-hard for a minimum codegree of n/k --- cn. Our algorithm relies on a theoretical result of independent interest, in which we characterise any such hypergraph with no perfect matching using a family of lattice-based constructions.", acknowledgement = ack-nhfb, } @InProceedings{Agrawal:2013:QPH, author = "Manindra Agrawal and Chandan Saha and Nitin Saxena", title = "Quasi-polynomial hitting-set for set-depth-{$ \Delta $} formulas", crossref = "ACM:2013:SPF", pages = "321--330", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488649", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We call a depth-4 formula $C$ set-depth-4 if there exists a (unknown) partition $ X_1 \sqcup \cdots \sqcup X_d $ of the variable indices $ [n] $ that the top product layer respects, i.e. $ C({{\rm term} x}) = \Sigma_{i = 1}^k \prod_{j = 1}^d f_{i, j} ({{\rm term} x}_{X j}) $, where $ f_{i, j} $ is a sparse polynomial in $ F[{{\rm term} x}_{X j}] $. Extending this definition to any depth --- we call a depth-$D$ formula $C$ (consisting of alternating layers of $ \Sigma $ and $ \Pi $ gates, with a $ \Sigma $-gate on top) a set-depth-$D$ formula if every $ \Pi $-layer in $C$ respects a (unknown) partition on the variables; if $D$ is even then the product gates of the bottom-most $ \Pi $-layer are allowed to compute arbitrary monomials. In this work, we give a hitting-set generator for set-depth-$D$ formulas (over any field) with running time polynomial in $ \exp ((D^2 \log s)^{\Delta - 1}) $, where $s$ is the size bound on the input set-depth-$D$ formula. In other words, we give a quasi-polynomial time blackbox polynomial identity test for such constant-depth formulas. Previously, the very special case of $ D = 3 $ (also known as set-multilinear depth-3 circuits) had no known sub-exponential time hitting-set generator. This was declared as an open problem by Shpilka {\&} Yehudayoff (FnT-TCS 2010); the model being first studied by Nisan {\&} Wigderson (FOCS 1995) and recently by Forbes {\&} Shpilka (STOC 2012 {\&} ECCC TR12-115). Our work settles this question, not only for depth-3 but, up to depth $ \epsilon \log s / \log \log s $, for a fixed constant $ \epsilon < 1 $. The technique is to investigate depth-$D$ formulas via depth-$ (D - 1) $ formulas over a Hadamard algebra, after applying a shift' on the variables. We propose a new algebraic conjecture about the low-support rank-concentration in the latter formulas, and manage to prove it in the case of set-depth-$D$ formulas.", acknowledgement = ack-nhfb, } @InProceedings{Hardt:2013:BWC, author = "Moritz Hardt and Aaron Roth", title = "Beyond worst-case analysis in private singular vector computation", crossref = "ACM:2013:SPF", pages = "331--340", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488650", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider differentially private approximate singular vector computation. Known worst-case lower bounds show that the error of any differentially private algorithm must scale polynomially with the dimension of the singular vector. We are able to replace this dependence on the dimension by a natural parameter known as the coherence of the matrix that is often observed to be significantly smaller than the dimension both theoretically and empirically. We also prove a matching lower bound showing that our guarantee is nearly optimal for every setting of the coherence parameter. Notably, we achieve our bounds by giving a robust analysis of the well-known power iteration algorithm, which may be of independent interest. Our algorithm also leads to improvements in worst-case settings and to better low-rank approximations in the spectral norm.", acknowledgement = ack-nhfb, } @InProceedings{Hsu:2013:DPA, author = "Justin Hsu and Aaron Roth and Jonathan Ullman", title = "Differential privacy for the analyst via private equilibrium computation", crossref = "ACM:2013:SPF", pages = "341--350", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488651", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give new mechanisms for answering exponentially many queries from multiple analysts on a private database, while protecting dif- ferential privacy both for the individuals in the database and for the analysts. That is, our mechanism's answer to each query is nearly insensitive to changes in the queries asked by other analysts. Our mechanism is the first to offer differential privacy on the joint distribution over analysts' answers, providing privacy for data an- alysts even if the other data analysts collude or register multiple accounts. In some settings, we are able to achieve nearly optimal error rates (even compared to mechanisms which do not offer an- alyst privacy), and we are able to extend our techniques to handle non-linear queries. Our analysis is based on a novel view of the private query-release problem as a two-player zero-sum game, which may be of independent interest.", acknowledgement = ack-nhfb, } @InProceedings{Nikolov:2013:GDP, author = "Aleksandar Nikolov and Kunal Talwar and Li Zhang", title = "The geometry of differential privacy: the sparse and approximate cases", crossref = "ACM:2013:SPF", pages = "351--360", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488652", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study trade-offs between accuracy and privacy in the context of linear queries over histograms. This is a rich class of queries that includes contingency tables and range queries and has been the focus of a long line of work. For a given set of $d$ linear queries over a database $ x \in R^N $, we seek to find the differentially private mechanism that has the minimum mean squared error. For pure differential privacy, [5, 32] give an $ O(\log^2 d) $ approximation to the optimal mechanism. Our first contribution is to give an efficient $ O(\log^2 d) $ approximation guarantee for the case of $ (\epsilon, \delta) $-differential privacy. Our mechanism adds carefully chosen correlated Gaussian noise to the answers. We prove its approximation guarantee relative to the hereditary discrepancy lower bound of [44], using tools from convex geometry. We next consider the sparse case when the number of queries exceeds the number of individuals in the database, i.e. when $ d > n \Delta |x|_1 $. The lower bounds used in the previous approximation algorithm no longer apply --- in fact better mechanisms are known in this setting [7, 27, 28, 31, 49]. Our second main contribution is to give an efficient $ (\epsilon, \delta) $-differentially private mechanism that, for any given query set $A$ and an upper bound $n$ on $ |x|_1 $, has mean squared error within $ \polylog (d, N) $ of the optimal for $A$ and $n$. This approximation is achieved by coupling the Gaussian noise addition approach with linear regression over the $ l_1 $ ball. Additionally, we show a similar polylogarithmic approximation guarantee for the optimal $ \epsilon $-differentially private mechanism in this sparse setting. Our work also shows that for arbitrary counting queries, i.e. $A$ with entries in $ \{ 0, 1 \} $, there is an $ \epsilon $-differentially private mechanism with expected error $ \tilde O(\sqrt n) $ per query, improving on the $ \tilde O(n^{2 / 3}) $ bound of [7] and matching the lower bound implied by [15] up to logarithmic factors. The connection between the hereditary discrepancy and the privacy mechanism enables us to derive the first polylogarithmic approximation to the hereditary discrepancy of a matrix $A$.", acknowledgement = ack-nhfb, } @InProceedings{Ullman:2013:ACQ, author = "Jonathan Ullman", title = "Answering $ n_{2 + o(1)} $ counting queries with differential privacy is hard", crossref = "ACM:2013:SPF", pages = "361--370", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488653", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A central problem in differentially private data analysis is how to design efficient algorithms capable of answering large numbers of counting queries on a sensitive database. Counting queries are of the form ``What fraction of individual records in the database satisfy the property $q$ ?'' We prove that if one-way functions exist, then there is no algorithm that takes as input a database $ {\rm db} \in {\rm dbset} $, and $ k = \tilde \Theta (n^2) $ arbitrary efficiently computable counting queries, runs in time $ \poly (d, n) $, and returns an approximate answer to each query, while satisfying differential privacy. We also consider the complexity of answering ``simple'' counting queries, and make some progress in this direction by showing that the above result holds even when we require that the queries are computable by constant-depth ($ {\rm AC}^0 $) circuits. Our result is almost tight because it is known that $ \tilde \Omega (n^2) $ counting queries can be answered efficiently while satisfying differential privacy. Moreover, many more than $ n^2 $ queries (even exponential in $n$) can be answered in exponential time. We prove our results by extending the connection between differentially private query release and cryptographic traitor-tracing schemes to the setting where the queries are given to the sanitizer as input, and by constructing a traitor-tracing scheme that is secure in this setting.", acknowledgement = ack-nhfb, } @InProceedings{Thorup:2013:BPS, author = "Mikkel Thorup", title = "Bottom-$k$ and priority sampling, set similarity and subset sums with minimal independence", crossref = "ACM:2013:SPF", pages = "371--380", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488655", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider bottom-$k$ sampling for a set X, picking a sample S$_k$ (X) consisting of the k elements that are smallest according to a given hash function h. With this sample we can estimate the relative size f=|Y|/|X| of any subset Y as |S$_k$ (X) intersect Y|/k. A standard application is the estimation of the Jaccard similarity f=|A intersect B|/|A union B| between sets A and B. Given the bottom-$k$ samples from A and B, we construct the bottom-$k$ sample of their union as S$_k$ (A union B)=S$_k$ (S$_k$ (A) union S$_k$ (B)), and then the similarity is estimated as |S$_k$ (A union B) intersect S$_k$ (A) intersect S$_k$ (B)|/k. We show here that even if the hash function is only 2-independent, the expected relative error is $ O(1 \sqrt (f_k)) $. For $ f_k = \Omega(1) $ this is within a constant factor of the expected relative error with truly random hashing. For comparison, consider the classic approach of kxmin-wise where we use $k$ hash independent functions $ h_1, \ldots {}, h_k $, storing the smallest element with each hash function. For kxmin-wise there is an at least constant bias with constant independence, and it is not reduced with larger $k$. Recently Feigenblat et al. showed that bottom-$k$ circumvents the bias if the hash function is 8-independent and $k$ is sufficiently large. We get down to 2-independence for any $k$. Our result is based on a simply union bound, transferring generic concentration bounds for the hashing scheme to the bottom-$k$ sample, e.g., getting stronger probability error bounds with higher independence. For weighted sets, we consider priority sampling which adapts efficiently to the concrete input weights, e.g., benefiting strongly from heavy-tailed input. This time, the analysis is much more involved, but again we show that generic concentration bounds can be applied.", acknowledgement = ack-nhfb, } @InProceedings{Lenzen:2013:FRT, author = "Christoph Lenzen and Boaz Patt-Shamir", title = "Fast routing table construction using small messages: extended abstract", crossref = "ACM:2013:SPF", pages = "381--390", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488656", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We describe a distributed randomized algorithm to construct routing tables. Given $ 0 < \epsilon \leq 1 / 2 $, the algorithm runs in time $ \tilde O(n^{1 / 2 + \epsilon } + {\rm HD}) $, where $n$ is the number of nodes and HD denotes the diameter of the network in hops (i.e., as if the network is unweighted). The weighted length of the produced routes is at most $ O(\epsilon^{-1} \log \epsilon^{-1}) $ times the optimal weighted length. This is the first algorithm to break the $ \Omega (n) $ complexity barrier for computing weighted shortest paths even for a single source. Moreover, the algorithm nearly meets the $ \tilde \Omega (n^{1 / 2} + {\rm HD}) $ lower bound for distributed computation of routing tables and approximate distances (with optimality, up to $ \polylog $ factors, for $ \epsilon = 1 / \log n $). The presented techniques have many applications, including improved distributed approximation algorithms for Generalized Steiner Forest, all-pairs distance estimation, and estimation of the weighted diameter.", acknowledgement = ack-nhfb, } @InProceedings{Mendes:2013:MAA, author = "Hammurabi Mendes and Maurice Herlihy", title = "Multidimensional approximate agreement in {Byzantine} asynchronous systems", crossref = "ACM:2013:SPF", pages = "391--400", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488657", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The problem of $ \epsilon $-approximate agreement in Byzantine asynchronous systems is well-understood when all values lie on the real line. In this paper, we generalize the problem to consider values that lie in $ R^m $, for $ m \geq 1 $, and present an optimal protocol in regard to fault tolerance. Our scenario is the following. Processes start with values in $ R^m $, for $ m \geq 1 $, and communicate via message-passing. The system is asynchronous: there is no upper bound on processes' relative speeds or on message delay. Some faulty processes can display arbitrarily malicious (i.e. Byzantine) behavior. Non-faulty processes must decide on values that are: (1) in $ R^m $; (2) within distance $ \epsilon $ of each other; and (3) in the convex hull of the non-faulty processes' inputs. We give an algorithm with a matching lower bound on fault tolerance: we require $ n > t(m + 2) $, where $n$ is the number of processes, $t$ is the number of Byzantine processes, and input and output values reside in $ R^m $. Non-faulty processes send $ O(n^2 d \log (m / \epsilon \max \{ \delta (d) : 1 \leq d \leq m \})) $ messages in total, where $ \delta (d) $ is the range of non-faulty inputs projected at coordinate $d$. The Byzantine processes do not affect the algorithm's running time.", acknowledgement = ack-nhfb, } @InProceedings{King:2013:BAP, author = "Valerie King and Jared Saia", title = "{Byzantine} agreement in polynomial expected time: [extended abstract]", crossref = "ACM:2013:SPF", pages = "401--410", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488658", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In the classic asynchronous Byzantine agreement problem, communication is via asynchronous message-passing and the adversary is adaptive with full information. In particular, the adversary can adaptively determine which processors to corrupt and what strategy these processors should use as the algorithm proceeds; the scheduling of the delivery of messages is set by the adversary, so that the delays are unpredictable to the algorithm; and the adversary knows the states of all processors at any time, and is assumed to be computationally unbounded. Such an adversary is also known as ``strong''. We present a polynomial expected time algorithm to solve asynchronous Byzantine Agreement with a strong adversary that controls up to a constant fraction of the processors. This is the first improvement in running time for this problem since Ben-Or's exponential expected time solution in 1983. Our algorithm tolerates an adversary that controls up to a $ 1 / 500 $ fraction of the processors.", acknowledgement = ack-nhfb, } @InProceedings{Chakrabarty:2013:MTB, author = "Deeparnab Chakrabarty and C. Seshadhri", title = "A $ o(n) $ monotonicity tester for {Boolean} functions over the hypercube", crossref = "ACM:2013:SPF", pages = "411--418", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488660", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Given oracle access to a Boolean function $ f : \{ 0, 1 \}^n \to \{ 0, 1 \} $, we design a randomized tester that takes as input a parameter $ \epsilon > 0 $, and outputs Yes if the function is monotonically non-increasing, and outputs No with probability $ > 2 / 3 $, if the function is $ \epsilon $-far from being monotone, that is, $f$ needs to be modified at $ \epsilon $-fraction of the points to make it monotone. Our non-adaptive, one-sided tester makes $ \tilde O(n^{5 / 6} \epsilon^{-5 / 3}) $ queries to the oracle.", acknowledgement = ack-nhfb, } @InProceedings{Chakrabarty:2013:OBM, author = "Deeparnab Chakrabarty and C. Seshadhri", title = "Optimal bounds for monotonicity and {Lipschitz} testing over hypercubes and hypergrids", crossref = "ACM:2013:SPF", pages = "419--428", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488661", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The problem of monotonicity testing over the hypergrid and its special case, the hypercube, is a classic question in property testing. We are given query access to $ f : [k]^n \to R $ (for some ordered range $R$). The hypergrid/cube has a natural partial order given by coordinate-wise ordering, denoted by prec. A function is monotone if for all pairs $ x \prec y $, $ f(x) \leq f(y) $. The distance to monotonicity, $ \epsilon_f $, is the minimum fraction of values of $f$ that need to be changed to make f monotone. For $ k = 2 $ (the boolean hypercube), the usual tester is the edge tester, which checks monotonicity on adjacent pairs of domain points. It is known that the edge tester using $ O(\epsilon^{-1} n \log |R|) $ samples can distinguish a monotone function from one where $ \epsilon_f > \epsilon $. On the other hand, the best lower bound for monotonicity testing over general $R$ is $ \Omega (n) $. We resolve this long standing open problem and prove that $ O(n / \epsilon) $ samples suffice for the edge tester. For hypergrids, known testers require $ O(\epsilon^{-1} n \log k \log |R|) $ samples, while the best known (non-adaptive) lower bound is $ \Omega (\epsilon^{-1} n \log k) $. We give a (non-adaptive) monotonicity tester for hypergrids running in $ O(\epsilon^{{-1} n \log k}) $ time. Our techniques lead to optimal property testers (with the same running time) for the natural Lipschitz property on hypercubes and hypergrids. (A $c$-Lipschitz function is one where $ |f(x) - f(y)| \leq c || x - y ||_1 $.) In fact, we give a general unified proof for $ O(\epsilon^{-1} n \log k) $-query testers for a class of ``bounded-derivative'' properties, a class containing both monotonicity and Lipschitz.", acknowledgement = ack-nhfb, } @InProceedings{Bhattacharyya:2013:ELC, author = "Arnab Bhattacharyya and Eldar Fischer and Hamed Hatami and Pooya Hatami and Shachar Lovett", title = "Every locally characterized affine-invariant property is testable", crossref = "ACM:2013:SPF", pages = "429--436", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488662", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Set $ F = F_p $ for any fixed prime $ p \geq 2 $. An affine-invariant property is a property of functions over $ F^n $ that is closed under taking affine transformations of the domain. We prove that all affine-invariant properties having local characterizations are testable. In fact, we show a proximity-oblivious test for any such property cP, meaning that given an input function $f$, we make a constant number of queries to $f$, always accept if $f$ satisfies cP, and otherwise reject with probability larger than a positive number that depends only on the distance between $f$ and cP. More generally, we show that any affine-invariant property that is closed under taking restrictions to subspaces and has bounded complexity is testable. We also prove that any property that can be described as the property of decomposing into a known structure of low-degree polynomials is locally characterized and is, hence, testable. For example, whether a function is a product of two degree-$d$ polynomials, whether a function splits into a product of d linear polynomials, and whether a function has low rank are all examples of degree-structural properties and are therefore locally characterized. Our results depend on a new Gowers inverse theorem by Tao and Ziegler for low characteristic fields that decomposes any polynomial with large Gowers norm into a function of a small number of low-degree non-classical polynomials. We establish a new equidistribution result for high rank non-classical polynomials that drives the proofs of both the testability results and the local characterization of degree-structural properties.", acknowledgement = ack-nhfb, } @InProceedings{Kawarabayashi:2013:TSF, author = "Ken-ichi Kawarabayashi and Yuichi Yoshida", title = "Testing subdivision-freeness: property testing meets structural graph theory", crossref = "ACM:2013:SPF", pages = "437--446", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488663", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Testing a property $P$ of graphs in the bounded-degree model deals with the following problem: given a graph $G$ of bounded degree $d$, we should distinguish (with probability 2/3, say) between the case that $G$ satisfies P and the case that one should add/remove at least $ \epsilon d n $ edges of $G$ to make it satisfy $P$. In sharp contrast to property testing of dense graphs, which is relatively well understood, only few properties are known to be testable with a constant number of queries in the bounded-degree model. In particular, no global monotone (i.e., closed under edge deletions) property that expander graphs can satisfy has been shown to be testable in constant time so far. In this paper, we identify for the first time a natural family of global monotone property that expander graphs can satisfy and can be efficiently tested in the bounded degree model. Specifically, we show that, for any integer $ t \geq 1 $, $ K_t $-subdivision-freeness is testable with a constant number of queries in the bounded-degree model. This property was not previously known to be testable even with o(n) queries. Note that an expander graph with all degree less than $ t - 1 $ does not have a $ K_t $-subdivision. The proof is based on a novel combination of some results that develop the framework of partitioning oracles, together with structural graph theory results that develop the seminal graph minor theory by Robertson and Seymour. As far as we aware, this is the first result that bridges property testing and structural graph theory. Although we know a rough structure for graphs without $H$-minors from the famous graph minor theory by Robertson and Seymour, there is no corresponding structure theorem for graphs without $H$-subdivisions so far, even $ K_5 $-subdivision-free graphs. Therefore, subdivisions and minors are very different in a graph structural sense.", acknowledgement = ack-nhfb, } @InProceedings{Chan:2013:ARP, author = "Siu On Chan", title = "Approximation resistance from pairwise independent subgroups", crossref = "ACM:2013:SPF", pages = "447--456", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488665", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show optimal (up to constant factor) NP-hardness for Max-$k$-CSP over any domain, whenever k is larger than the domain size. This follows from our main result concerning predicates over abelian groups. We show that a predicate is approximation resistant if it contains a subgroup that is balanced pairwise independent. This gives an unconditional analogue of Austrin--Mossel hardness result, bypassing the Unique-Games Conjecture for predicates with an abelian subgroup structure. Our main ingredient is a new gap-amplification technique inspired by XOR-lemmas. Using this technique, we also improve the NP-hardness of approximating Independent-Set on bounded-degree graphs, Almost-Coloring, Two-Prover-One-Round-Game, and various other problems.", acknowledgement = ack-nhfb, } @InProceedings{Huang:2013:ARS, author = "Sangxia Huang", title = "Approximation resistance on satisfiable instances for predicates with few accepting inputs", crossref = "ACM:2013:SPF", pages = "457--466", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488666", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove that for all integer $ k \geq 3 $, there is a predicate P on k Boolean variables with 2$^{\tilde O(k 1 / 3)}$ accepting assignments that is approximation resistant even on satisfiable instances. That is, given a satisfiable CSP instance with constraint P, we cannot achieve better approximation ratio than simply picking random assignments. This improves the best previously known result by Hastad and Khot where the predicate has 2$^{O(k 1 / 2)}$ accepting assignments. Our construction is inspired by several recent developments. One is the idea of using direct sums to improve soundness of PCPs, developed by Chan [5]. We also use techniques from Wenner [32] to construct PCPs with perfect completeness without relying on the d-to-1 Conjecture.", acknowledgement = ack-nhfb, } @InProceedings{Garg:2013:WEA, author = "Sanjam Garg and Craig Gentry and Amit Sahai and Brent Waters", title = "Witness encryption and its applications", crossref = "ACM:2013:SPF", pages = "467--476", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488667", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We put forth the concept of witness encryption. A witness encryption scheme is defined for an NP language L (with corresponding witness relation R). In such a scheme, a user can encrypt a message M to a particular problem instance x to produce a ciphertext. A recipient of a ciphertext is able to decrypt the message if x is in the language and the recipient knows a witness w where R(x,w) holds. However, if x is not in the language, then no polynomial-time attacker can distinguish between encryptions of any two equal length messages. We emphasize that the encrypter himself may have no idea whether $x$ is actually in the language. Our contributions in this paper are threefold. First, we introduce and formally define witness encryption. Second, we show how to build several cryptographic primitives from witness encryption. Finally, we give a candidate construction based on the NP-complete Exact Cover problem and Garg, Gentry, and Halevi's recent construction of ``approximate'' multilinear maps. Our method for witness encryption also yields the first candidate construction for an open problem posed by Rudich in 1989: constructing computational secret sharing schemes for an NP-complete access structure.", acknowledgement = ack-nhfb, } @InProceedings{De:2013:MSD, author = "Anindya De and Elchanan Mossel and Joe Neeman", title = "Majority is stablest: discrete and {SoS}", crossref = "ACM:2013:SPF", pages = "477--486", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488668", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Majority is Stablest Theorem has numerous applications in hardness of approximation and social choice theory. We give a new proof of the Majority is Stablest Theorem by induction on the dimension of the discrete cube. Unlike the previous proof, it uses neither the ``invariance principle'' nor Borell's result in Gaussian space. The new proof is general enough to include all previous variants of majority is stablest such as ``it ain't over until it's over'' and ``Majority is most predictable''. Moreover, the new proof allows us to derive a proof of Majority is Stablest in a constant level of the Sum of Squares hierarchy. This implies in particular that Khot-Vishnoi instance of Max-Cut does not provide a gap instance for the Lasserre hierarchy.", acknowledgement = ack-nhfb, } @InProceedings{Beck:2013:SEH, author = "Christopher Beck and Russell Impagliazzo", title = "Strong {ETH} holds for regular resolution", crossref = "ACM:2013:SPF", pages = "487--494", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488669", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We obtain asymptotically sharper lower bounds on resolution complexity for $k$-CNF's than was known previously. We show that for any large enough $k$ there are $k$-CNF's which require resolution width $ (1 - \tilde O(k^{-1 / 4}))n $, regular resolution size 2$^{(1 - \tilde O(k^{-1 / 4}))n}$, and general resolution size (3/2)$^{(1 - \tilde O(k - 1 / 4))n}$.", acknowledgement = ack-nhfb, } @InProceedings{Lee:2013:NCO, author = "James R. Lee and Manor Mendel and Mohammad Moharrami", title = "A node-capacitated {Okamura--Seymour} theorem", crossref = "ACM:2013:SPF", pages = "495--504", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488671", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The classical Okamura-Seymour theorem states that for an edge-capacitated, multi-commodity flow instance in which all terminals lie on a single face of a planar graph, there exists a feasible concurrent flow if and only if the cut conditions are satisfied. Simple examples show that a similar theorem is impossible in the node-capacitated setting. Nevertheless, we prove that an approximate flow/cut theorem does hold: For some universal $ \epsilon > 0 $, if the node cut conditions are satisfied, then one can simultaneously route an $ \epsilon $-fraction of all the demands. This answers an open question of Chekuri and Kawarabayashi. More generally, we show that this holds in the setting of multi-commodity polymatroid networks introduced by Chekuri, et. al. Our approach employs a new type of random metric embedding in order to round the convex programs corresponding to these more general flow problems.", acknowledgement = ack-nhfb, } @InProceedings{Klein:2013:SRS, author = "Philip N. Klein and Shay Mozes and Christian Sommer", title = "Structured recursive separator decompositions for planar graphs in linear time", crossref = "ACM:2013:SPF", pages = "505--514", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488672", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Given a triangulated planar graph $G$ on $n$ vertices and an integer $r$ $r$-division of $G$ with few holes is a decomposition of $G$ into $ O(n / r) $ regions of size at most r such that each region contains at most a constant number of faces that are not faces of $G$ (also called holes), and such that, for each region, the total number of vertices on these faces is $ O(\sqrt r) $. We provide an algorithm for computing $r$-divisions with few holes in linear time. In fact, our algorithm computes a structure, called decomposition tree, which represents a recursive decomposition of $G$ that includes $r$-divisions for essentially all values of $r$. In particular, given an exponentially increasing sequence $ \{ \vec r \} = (r_1, r_2, \ldots {}) $, our algorithm can produce a recursive $ \{ \vec r \} $-division with few holes in linear time. $r$-divisions with few holes have been used in efficient algorithms to compute shortest paths, minimum cuts, and maximum flows. Our linear-time algorithm improves upon the decomposition algorithm used in the state-of-the-art algorithm for minimum st--cut (Italiano, Nussbaum, Sankowski, and Wulff-Nilsen, STOC 2011), removing one of the bottlenecks in the overall running time of their algorithm (analogously for minimum cut in planar and bounded-genus graphs).", acknowledgement = ack-nhfb, } @InProceedings{Roditty:2013:FAA, author = "Liam Roditty and Virginia Vassilevska Williams", title = "Fast approximation algorithms for the diameter and radius of sparse graphs", crossref = "ACM:2013:SPF", pages = "515--524", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488673", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The diameter and the radius of a graph are fundamental topological parameters that have many important practical applications in real world networks. The fastest combinatorial algorithm for both parameters works by solving the all-pairs shortest paths problem (APSP) and has a running time of $ \tilde O(m n) $ in $m$-edge, $n$-node graphs. In a seminal paper, Aingworth, Chekuri, Indyk and Motwani [SODA'96 and SICOMP'99] presented an algorithm that computes in $ \tilde O(m \sqrt n + n^2) $ time an estimate $D$ for the diameter $D$, such that $ \lfloor 2 / 3 D \rfloor \leq^D \leq D $. Their paper spawned a long line of research on approximate APSP. For the specific problem of diameter approximation, however, no improvement has been achieved in over 15 years. Our paper presents the first improvement over the diameter approximation algorithm of Aingworth et. al, producing an algorithm with the same estimate but with an expected running time of $ \tilde O(m \sqrt n) $. We thus show that for all sparse enough graphs, the diameter can be 3/2-approximated in $ o(n^2) $ time. Our algorithm is obtained using a surprisingly simple method of neighborhood depth estimation that is strong enough to also approximate, in the same running time, the radius and more generally, all of the eccentricities, i.e. for every node the distance to its furthest node. We also provide strong evidence that our diameter approximation result may be hard to improve. We show that if for some constant $ \epsilon > 0 $ there is an $ O(m^{2 - \epsilon }) $ time $ (3 / 2 - \epsilon) $-approximation algorithm for the diameter of undirected unweighted graphs, then there is an $ O*((2 - \delta)^n) $ time algorithm for CNF-SAT on $n$ variables for constant $ \delta > 0 $, and the strong exponential time hypothesis of [Impagliazzo, Paturi, Zane JCSS'01] is false. Motivated by this negative result, we give several improved diameter approximation algorithms for special cases. We show for instance that for unweighted graphs of constant diameter $D$ not divisible by $3$, there is an $ O(m^{2 - \epsilon }) $ time algorithm that gives a $ (3 / 2 - \epsilon) $ approximation for constant $ \epsilon > 0 $. This is interesting since the diameter approximation problem is hardest to solve for small $D$.", acknowledgement = ack-nhfb, } @InProceedings{Gu:2013:PDM, author = "Albert Gu and Anupam Gupta and Amit Kumar", title = "The power of deferral: maintaining a constant-competitive {Steiner} tree online", crossref = "ACM:2013:SPF", pages = "525--534", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488674", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In the online Steiner tree problem, a sequence of points is revealed one-by-one: when a point arrives, we only have time to add a single edge connecting this point to the previous ones, and we want to minimize the total length of edges added. Here, a tight bound has been known for two decades: the greedy algorithm maintains a tree whose cost is $ O(\log n) $ times the Steiner tree cost, and this is best possible. But suppose, in addition to the new edge we add, we have time to change a single edge from the previous set of edges: can we do much better? Can we, e.g., maintain a tree that is constant-competitive? We answer this question in the affirmative. We give a primal-dual algorithm that makes only a single swap per step (in addition to adding the edge connecting the new point to the previous ones), and such that the tree's cost is only a constant times the optimal cost. Our dual-based analysis is quite different from previous primal-only analyses. In particular, we give a correspondence between radii of dual balls and lengths of tree edges; since dual balls are associated with points and hence do not move around (in contrast to edges), we can closely monitor the edge lengths based on the dual radii. Showing that these dual radii cannot change too rapidly is the technical heart of the paper, and allows us to give a hard bound on the number of swaps per arrival, while maintaining a constant-competitive tree at all times. Previous results for this problem gave an algorithm that performed an amortized constant number of swaps: for each n, the number of swaps in the first $n$ steps was O(n). We also give a simpler tight analysis for this amortized case.", acknowledgement = ack-nhfb, } @InProceedings{Buchbinder:2013:SPE, author = "Niv Buchbinder and Joseph (Seffi) Naor and Roy Schwartz", title = "Simplex partitioning via exponential clocks and the multiway cut problem", crossref = "ACM:2013:SPF", pages = "535--544", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488675", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The Multiway-Cut problem is a fundamental graph partitioning problem in which the objective is to find a minimum weight set of edges disconnecting a given set of special vertices called terminals. This problem is NP-hard and there is a well known geometric relaxation in which the graph is embedded into a high dimensional simplex. Rounding a solution to the geometric relaxation is equivalent to partitioning the simplex. We present a novel simplex partitioning algorithm which is based on em competing exponential clocks and distortion. Unlike previous methods, it utilizes cuts that are not parallel to the faces of the simplex. Applying this partitioning algorithm to the multiway cut problem, we obtain a simple (4/3)-approximation algorithm, thus, improving upon the current best known result. This bound is further pushed to obtain an approximation factor of 1.32388. It is known that under the assumption of the unique games conjecture, the best possible approximation for the Multiway-Cut problem can be attained via the geometric relaxation.", acknowledgement = ack-nhfb, } @InProceedings{Gorbunov:2013:ABE, author = "Sergey Gorbunov and Vinod Vaikuntanathan and Hoeteck Wee", title = "Attribute-based encryption for circuits", crossref = "ACM:2013:SPF", pages = "545--554", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488677", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In an attribute-based encryption (ABE) scheme, a ciphertext is associated with an $l$-bit public index pind and a message $m$, and a secret key is associated with a Boolean predicate $P$. The secret key allows to decrypt the ciphertext and learn $m$ iff $P({\rm pind}) = 1$. Moreover, the scheme should be secure against collusions of users, namely, given secret keys for polynomially many predicates, an adversary learns nothing about the message if none of the secret keys can individually decrypt the ciphertext. We present attribute-based encryption schemes for circuits of any arbitrary polynomial size, where the public parameters and the ciphertext grow linearly with the depth of the circuit. Our construction is secure under the standard learning with errors (LWE) assumption. Previous constructions of attribute-based encryption were for Boolean formulas, captured by the complexity class ${\rm NC}^1 $. In the course of our construction, we present a new framework for constructing ABE schemes. As a by-product of our framework, we obtain ABE schemes for polynomial-size branching programs, corresponding to the complexity class LOGSPACE, under quantitatively better assumptions.", acknowledgement = ack-nhfb, } @InProceedings{Goldwasser:2013:RGC, author = "Shafi Goldwasser and Yael Kalai and Raluca Ada Popa and Vinod Vaikuntanathan and Nickolai Zeldovich", title = "Reusable garbled circuits and succinct functional encryption", crossref = "ACM:2013:SPF", pages = "555--564", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488678", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Garbled circuits, introduced by Yao in the mid 80s, allow computing a function f on an input x without leaking anything about f or x besides f(x). Garbled circuits found numerous applications, but every known construction suffers from one limitation: it offers no security if used on multiple inputs x. In this paper, we construct for the first time reusable garbled circuits. The key building block is a new succinct single-key functional encryption scheme. Functional encryption is an ambitious primitive: given an encryption $ {\rm Enc}(x) $ of a value $x$, and a secret key $ {\rm sk}_f $ for a function $f$, anyone can compute $ f(x) $ without learning any other information about $x$. We construct, for the first time, a succinct functional encryption scheme for {\em any} polynomial-time function f where succinctness means that the ciphertext size does not grow with the size of the circuit for $f$, but only with its depth. The security of our construction is based on the intractability of the Learning with Errors (LWE) problem and holds as long as an adversary has access to a single key $ {\rm sk}_f $ (or even an a priori bounded number of keys for different functions). Building on our succinct single-key functional encryption scheme, we show several new applications in addition to reusable garbled circuits, such as a paradigm for general function obfuscation which we call token-based obfuscation, homomorphic encryption for a class of Turing machines where the evaluation runs in input-specific time rather than worst-case time, and a scheme for delegating computation which is publicly verifiable and maintains the privacy of the computation.", acknowledgement = ack-nhfb, } @InProceedings{Kalai:2013:DBS, author = "Yael Tauman Kalai and Ran Raz and Ron D. Rothblum", title = "Delegation for bounded space", crossref = "ACM:2013:SPF", pages = "565--574", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488679", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We construct a 1-round delegation scheme for every language computable in time t=t(n) and space s=s(n), where the running time of the prover is poly(t) and the running time of the verifier is $ \tilde O(n + p o l y(s)) $ (where $ \tilde O $ hides $ \polylog (t) $ factors). The proof exploits a curious connection between the problem of computation delegation and the model of multi-prover interactive proofs that are sound against no-signaling (cheating) strategies, a model that was studied in the context of multi-prover interactive proofs with provers that share quantum entanglement, and is motivated by the physical principle that information cannot travel faster than light. For any language computable in time $ t = t(n) $ and space $ s = s(n) $, we construct MIPs that are sound against no-signaling strategies, where the running time of the provers is $ \poly (t) $, the number of provers is $ \tilde O(s) $, and the running time of the verifier is $ \tilde O(s + n) $. We then show how to use the method suggested by Aiello et-al (ICALP, 2000) to convert our MIP into a 1-round delegation scheme, by using a computational private information retrieval (PIR) scheme. Thus, assuming the existence of a sub-exponentially secure PIR scheme, we get our 1-round delegation scheme.", acknowledgement = ack-nhfb, } @InProceedings{Brakerski:2013:CHL, author = "Zvika Brakerski and Adeline Langlois and Chris Peikert and Oded Regev and Damien Stehl{\'e}", title = "Classical hardness of learning with errors", crossref = "ACM:2013:SPF", pages = "575--584", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488680", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show that the Learning with Errors (LWE) problem is classically at least as hard as standard worst-case lattice problems. Previously this was only known under quantum reductions. Our techniques capture the tradeoff between the dimension and the modulus of LWE instances, leading to a much better understanding of the landscape of the problem. The proof is inspired by techniques from several recent cryptographic constructions, most notably fully homomorphic encryption schemes.", acknowledgement = ack-nhfb, } @InProceedings{Ben-Sasson:2013:CEP, author = "Eli Ben-Sasson and Alessandro Chiesa and Daniel Genkin and Eran Tromer", title = "On the concrete efficiency of probabilistically-checkable proofs", crossref = "ACM:2013:SPF", pages = "585--594", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488681", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Probabilistically-Checkable Proofs (PCPs) form the algorithmic core that enables fast verification of long computations in many cryptographic constructions. Yet, despite the wonderful asymptotic savings they bring, PCPs are also the infamous computational bottleneck preventing these powerful cryptographic constructions from being used in practice. To address this problem, we present several results about the computational efficiency of PCPs. We construct the first PCP where the prover and verifier time complexities are quasi-optimal (i.e., optimal up to poly-logarithmic factors). The prover and verifier are also highly-parallelizable, and these computational guarantees hold even when proving and verifying the correctness of random-access machine computations. Our construction is explicit and has the requisite properties for being used in the cryptographic applications mentioned above. Next, to better understand the efficiency of our PCP, we propose a new efficiency measure for PCPs (and their major components, locally-testable codes and PCPs of proximity). We define a concrete-efficiency threshold that indicates the smallest problem size beyond which the PCP becomes ``useful'', in the sense that using it is cheaper than performing naive verification (i.e., rerunning the computation); our definition accounts for both the prover and verifier complexity. We then show that our PCP has a finite concrete-efficiency threshold. That such a PCP exists does not follow from existing works on PCPs with polylogarithmic-time verifiers. As in [Ben-Sasson and Sudan, STOC '05], PCPs of proximity for Reed--Solomon (RS) codes are the main component of our PCP. We construct a PCP of proximity that reduces the concrete-efficiency threshold for testing proximity to RS codes from 2$^{683}$ in their work to 2$^{43}$, which is tantalizingly close to practicality.", acknowledgement = ack-nhfb, } @InProceedings{Cadek:2013:ECM, author = "Martin Cadek and Marek Krcal and Jiri Matousek and Lukas Vokrinek and Uli Wagner", title = "Extending continuous maps: polynomiality and undecidability", crossref = "ACM:2013:SPF", pages = "595--604", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488683", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider several basic problems of algebraic topology, with connections to combinatorial and geometric questions, from the point of view of computational complexity. The extension problem asks, given topological spaces $X$, $Y$, a subspace $ A \subseteq X $, and a (continuous) map $ f : A \to Y $, whether $f$ can be extended to a map $ X \to Y $. For computational purposes, we assume that X and Y are represented as finite simplicial complexes, $A$ is a subcomplex of $X$, and $f$ is given as a simplicial map. In this generality the problem is undecidable, as follows from Novikov's result from the 1950s on uncomputability of the fundamental group $ \pi_1 (Y) $. We thus study the problem under the assumption that, for some $ k \geq 2 $, $Y$ is $ (k - 1) $-connected; informally, this means that $Y$ has ``no holes up to dimension $ k - 1 $'' i.e., the first $ k - 1 $ homotopy groups of $Y$ vanish (a basic example of such a $Y$ is the sphere $ S^k $). We prove that, on the one hand, this problem is still undecidable for $ \dim X = 2 k $. On the other hand, for every fixed $ k \geq 2 $, we obtain an algorithm that solves the extension problem in polynomial time assuming $Y$ $ (k - 1) $-connected and $ \dim X \leq 2 k - 1 $. For $ \dim X \leq 2 k - 2 $, the algorithm also provides a classification of all extensions up to homotopy (continuous deformation). This relies on results of our SODA 2012 paper, and the main new ingredient is a machinery of objects with polynomial-time homology, which is a polynomial-time analog of objects with effective homology developed earlier by Sergeraert et al. We also consider the computation of the higher homotopy groups $ \pi_k (Y) $, $ k \geq 2 $, for a 1-connected $Y$. Their computability was established by Brown in 1957; we show that $ \pi_k(Y) $ can be computed in polynomial time for every fixed $ k \geq 2 $. On the other hand, Anick proved in 1989 that computing $ \pi_k(Y) $ is \#P-hard if $k$ is a part of input, where $Y$ is a cell complex with certain rather compact encoding. We strengthen his result to \#P-hardness for $Y$ given as a simplicial complex.", acknowledgement = ack-nhfb, } @InProceedings{Har-Peled:2013:NPL, author = "Sariel Har-Peled and Benjamin Adam Raichel", title = "Net and prune: a linear time algorithm for {Euclidean} distance problems", crossref = "ACM:2013:SPF", pages = "605--614", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488684", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We provide a general framework for getting linear time constant factor approximations (and in many cases FPTAS's) to a copious amount of well known and well studied problems in Computational Geometry, such as $k$-center clustering and furthest nearest neighbor. The new approach is robust to variations in the input problem, and yet it is simple, elegant and practical. In particular, many of these well studied problems which fit easily into our framework, either previously had no linear time approximation algorithm, or required rather involved algorithms and analysis. A short list of the problems we consider include furthest nearest neighbor, $k$-center clustering, smallest disk enclosing k points, $k$-th largest distance, $k$-th smallest $m$-nearest neighbor distance, $k$-th heaviest edge in the MST and other spanning forest type problems, problems involving upward closed set systems, and more. Finally, we show how to extend our framework such that the linear running time bound holds with high probability.", acknowledgement = ack-nhfb, } @InProceedings{Caputo:2013:RLT, author = "Pietro Caputo and Fabio Martinelli and Alistair Sinclair and Alexandre Stauffer", title = "Random lattice triangulations: structure and algorithms", crossref = "ACM:2013:SPF", pages = "615--624", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488685", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The paper concerns lattice triangulations, i.e., triangulations of the integer points in a polygon in $ R^2 $ whose vertices are also integer points. Lattice triangulations have been studied extensively both as geometric objects in their own right and by virtue of applications in algebraic geometry. Our focus is on random triangulations in which a triangulation $ \sigma $ has weight $ \lambda^{| \sigma |} $, where $ \lambda $ is a positive real parameter and $ | \sigma | $ is the total length of the edges in $ \sigma $. Empirically, this model exhibits a ``phase transition'' at $ \lambda = 1 $ (corresponding to the uniform distribution): for $ \lambda < 1 $ distant edges behave essentially independently, while for $ \lambda > 1 $ very large regions of aligned edges appear. We substantiate this picture as follows. For $ \lambda < 1 $ sufficiently small, we show that correlations between edges decay exponentially with distance (suitably defined), and also that the Glauber dynamics (a local Markov chain based on flipping edges) is rapidly mixing (in time polynomial in the number of edges). This dynamics has been proposed by several authors as an algorithm for generating random triangulations. By contrast, for $ \lambda > 1 $ we show that the mixing time is exponential. These are apparently the first rigorous quantitative results on spatial mixing properties and dynamics of random lattice triangulations.", acknowledgement = ack-nhfb, } @InProceedings{Sinclair:2013:LYT, author = "Alistair Sinclair and Piyush Srivastava", title = "{Lee--Yang} theorems and the complexity of computing averages", crossref = "ACM:2013:SPF", pages = "625--634", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488686", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the complexity of computing average quantities related to spin systems, such as the mean magnetization and susceptibility in the ferromagnetic Ising model, and the average dimer count (or average size of a matching) in the monomer-dimer model. By establishing connections between the complexity of computing these averages and the location of the complex zeros of the partition function, we show that these averages are \#P-hard to compute. In case of the Ising model, our approach requires us to prove an extension of the famous Lee--Yang Theorem from the 1950s.", acknowledgement = ack-nhfb, } @InProceedings{Cai:2013:CDR, author = "Jin-Yi Cai and Heng Guo and Tyson Williams", title = "A complete dichotomy rises from the capture of vanishing signatures: extended abstract", crossref = "ACM:2013:SPF", pages = "635--644", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488687", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We prove a complexity dichotomy theorem for Holant problems over an arbitrary set of complex-valued symmetric constraint functions {F} on Boolean variables. This extends and unifies all previous dichotomies for Holant problems on symmetric constraint functions (taking values without a finite modulus). We define and characterize all symmetric vanishing signatures. They turned out to be essential to the complete classification of Holant problems. The dichotomy theorem has an explicit tractability criterion. A Holant problem defined by a set of constraint functions {F} is solvable in polynomial time if it satisfies this tractability criterion, and is \#P-hard otherwise. The tractability criterion can be intuitively stated as follows: A set {F} is tractable if (1) every function in {F} has arity at most two, or (2) {F} is transformable to an affine type, or (3) {F} is transformable to a product type, or (4) {F} is vanishing, combined with the right type of binary functions, or (5) {F} belongs to a special category of vanishing type Fibonacci gates. The proof of this theorem utilizes many previous dichotomy theorems on Holant problems and Boolean \#CSP. Holographic transformations play an indispensable role, not only as a proof technique, but also in the statement of the dichotomy criterion.", acknowledgement = ack-nhfb, } @InProceedings{Miller:2013:SLO, author = "Gary L. Miller", title = "Solving large optimization problems using spectral graph theory", crossref = "ACM:2013:SPF", pages = "981--981", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488689", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Spectral Graph Theory is the interplay between linear algebra and combinatorial graph theory. One application of this interplay is a nearly linear time solver for Symmetric Diagonally Dominate systems (SDD). This seemingly restrictive class of systems has received much interest in the last 15 years. Both algorithm design theory and practical implementations have made substantial progress. There is also a growing number of problems that can be efficiently solved using SDD solvers including: image segmentation, image denoising, finding solutions to elliptic equations, computing maximum flow in a graph, graph sparsification, and graphics. All these examples can be viewed as special case of convex optimization problems.", acknowledgement = ack-nhfb, } @InProceedings{Elkin:2013:OES, author = "Michael Elkin and Shay Solomon", title = "Optimal {Euclidean} spanners: really short, thin and lanky", crossref = "ACM:2013:SPF", pages = "645--654", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488691", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The degree, the (hop-)diameter, and the weight are the most basic and well-studied parameters of geometric spanners. In a seminal STOC'95 paper, titled ``Euclidean spanners: short, thin and lanky'', Arya et al. [2] devised a construction of Euclidean $ (1 + \epsilon) $-spanners that achieves constant degree, diameter $ O(\log n) $, weight $ O(\log^2 n) \cdot \omega ({\rm MST}) $, and has running time $ O(n \cdot \log n) $. This construction applies to $n$-point constant-dimensional Euclidean spaces. Moreover, Arya et al. conjectured that the weight bound can be improved by a logarithmic factor, without increasing the degree and the diameter of the spanner, and within the same running time. This conjecture of Arya et al. became one of the most central open problems in the area of Euclidean spanners. Nevertheless, the only progress since 1995 towards its resolution was achieved in the lower bounds front: Any spanner with diameter $ O(\log n) $ must incur weight $ \Omega (\log n) \cdot \omega ({\rm MST}) $, and this lower bound holds regardless of the stretch or the degree of the spanner [12, 1]. In this paper we resolve the long-standing conjecture of Arya et al. in the affirmative. We present a spanner construction with the same stretch, degree, diameter, and running time, as in Arya et al.'s result, but with optimal weight $ O(\log n) \cdot \omega ({\rm MST}) $. So our spanners are as thin and lanky as those of Arya et al., but they are really short! Moreover, our result is more general in three ways. First, we demonstrate that the conjecture holds true not only in constant-dimensional Euclidean spaces, but also in doubling metrics. Second, we provide a general tradeoff between the three involved parameters, which is tight in the entire range. Third, we devise a transformation that decreases the lightness of spanners in general metrics, while keeping all their other parameters in check. Our main result is obtained as a corollary of this transformation.", acknowledgement = ack-nhfb, } @InProceedings{Feldman:2013:SAL, author = "Vitaly Feldman and Elena Grigorescu and Lev Reyzin and Santosh Vempala and Ying Xiao", title = "Statistical algorithms and a lower bound for detecting planted cliques", crossref = "ACM:2013:SPF", pages = "655--664", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488692", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We introduce a framework for proving lower bounds on computational problems over distributions, based on a class of algorithms called statistical algorithms. For such algorithms, access to the input distribution is limited to obtaining an estimate of the expectation of any given function on a sample drawn randomly from the input distribution, rather than directly accessing samples. Most natural algorithms of interest in theory and in practice, e.g., moments-based methods, local search, standard iterative methods for convex optimization, MCMC and simulated annealing, are statistical algorithms or have statistical counterparts. Our framework is inspired by and generalize the statistical query model in learning theory [34]. Our main application is a nearly optimal lower bound on the complexity of any statistical algorithm for detecting planted bipartite clique distributions (or planted dense subgraph distributions) when the planted clique has size $ O(n^{1 / 2 - \delta}) $ for any constant $ \delta > 0 $. Variants of these problems have been assumed to be hard to prove hardness for other problems and for cryptographic applications. Our lower bounds provide concrete evidence of hardness, thus supporting these assumptions.", acknowledgement = ack-nhfb, } @InProceedings{Jain:2013:LRM, author = "Prateek Jain and Praneeth Netrapalli and Sujay Sanghavi", title = "Low-rank matrix completion using alternating minimization", crossref = "ACM:2013:SPF", pages = "665--674", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488693", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Alternating minimization represents a widely applicable and empirically successful approach for finding low-rank matrices that best fit the given data. For example, for the problem of low-rank matrix completion, this method is believed to be one of the most accurate and efficient, and formed a major component of the winning entry in the Netflix Challenge [17]. In the alternating minimization approach, the low-rank target matrix is written in a bi-linear form, i.e. X = UV$^+$; the algorithm then alternates between finding the best U and the best V. Typically, each alternating step in isolation is convex and tractable. However the overall problem becomes non-convex and is prone to local minima. In fact, there has been almost no theoretical understanding of when this approach yields a good result. In this paper we present one of the first theoretical analyses of the performance of alternating minimization for matrix completion, and the related problem of matrix sensing. For both these problems, celebrated recent results have shown that they become well-posed and tractable once certain (now standard) conditions are imposed on the problem. We show that alternating minimization also succeeds under similar conditions. Moreover, compared to existing results, our paper shows that alternating minimization guarantees faster (in particular, geometric) convergence to the true matrix, while allowing a significantly simpler analysis.", acknowledgement = ack-nhfb, } @InProceedings{Alon:2013:ARM, author = "Noga Alon and Troy Lee and Adi Shraibman and Santosh Vempala", title = "The approximate rank of a matrix and its algorithmic applications: approximate rank", crossref = "ACM:2013:SPF", pages = "675--684", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488694", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the $ \epsilon $-rank of a real matrix A, defined for any $ \epsilon > 0 $ as the minimum rank over matrices that approximate every entry of $A$ to within an additive $ \epsilon $. This parameter is connected to other notions of approximate rank and is motivated by problems from various topics including communication complexity, combinatorial optimization, game theory, computational geometry and learning theory. Here we give bounds on the $ \epsilon $-rank and use them for algorithmic applications. Our main algorithmic results are (a) polynomial-time additive approximation schemes for Nash equilibria for 2-player games when the payoff matrices are positive semidefinite or have logarithmic rank and (b) an additive PTAS for the densest subgraph problem for similar classes of weighted graphs. We use combinatorial, geometric and spectral techniques; our main new tool is an algorithm for efficiently covering a convex body with translates of another convex body.", acknowledgement = ack-nhfb, } @InProceedings{Harris:2013:CSP, author = "David G. Harris and Aravind Srinivasan", title = "Constraint satisfaction, packet routing, and the {Lovasz} local lemma", crossref = "ACM:2013:SPF", pages = "685--694", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488696", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Constraint-satisfaction problems (CSPs) form a basic family of NP-hard optimization problems that includes satisfiability. Motivated by the sufficient condition for the satisfiability of SAT formulae that is offered by the Lovasz Local Lemma, we seek such sufficient conditions for arbitrary CSPs. To this end, we identify a variable-covering radius--type parameter for the infeasible configurations of a given CSP, and also develop an extension of the Lovasz Local Lemma in which many of the events to be avoided have probabilities arbitrarily close to one; these lead to a general sufficient condition for the satisfiability of arbitrary CSPs. One primary application is to packet-routing in the classical Leighton-Maggs-Rao setting, where we introduce several additional ideas in order to prove the existence of near-optimal schedules; further applications in combinatorial optimization are also shown.", acknowledgement = ack-nhfb, } @InProceedings{Thapper:2013:CFV, author = "Johan Thapper and Stanislav Zivny", title = "The complexity of finite-valued {CSPs}", crossref = "ACM:2013:SPF", pages = "695--704", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488697", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Let $ \Gamma $ be a set of rational-valued functions on a fixed finite domain; such a set is called a finite-valued constraint language. The valued constraint satisfaction problem, VCSP( $ \Gamma $), is the problem of minimising a function given as a sum of functions from $ \Gamma $ . We establish a dichotomy theorem with respect to exact solvability for all finite-valued languages defined on domains of arbitrary finite size. We show that every core language $ \Gamma $ either admits a binary idempotent and symmetric fractional polymorphism in which case the basic linear programming relaxation solves any instance of VCSP( $ \Gamma $) exactly, or $ \Gamma $ satisfies a simple hardness condition that allows for a polynomial-time reduction from Max-Cut to VCSP( $ \Gamma $). In other words, there is a single algorithm for all tractable cases and a single reason for intractability. Our results show that for exact solvability of VCSPs the basic linear programming relaxation suffices and semidefinite relaxations do not add any power. Our results generalise all previous partial classifications of finite-valued languages: the classification of {0,1}-valued languages containing all unary functions obtained by Deineko et al. [JACM'06]; the classifications of {0,1}-valued languages on two-element, three-element, and four-element domains obtained by Creignou [JCSS'95], Jonsson et al. [SICOMP'06], and Jonsson et al.[CP'11], respectively; the classifications of finite-valued languages on two-element and three-element domains obtained by Cohen et al. [AIJ'06] and Huber et al. [SODA'13], respectively; the classification of finite-valued languages containing all {0,1}-valued unary functions obtained by Kolmogorov and Zivny [JACM'13]; and the classification of Min-0-Ext problems obtained by Hirai [SODA'13].", acknowledgement = ack-nhfb, } @InProceedings{Coja-Oghlan:2013:GAK, author = "Amin Coja-Oghlan and Konstantinos Panagiotou", title = "Going after the {k-SAT} threshold", crossref = "ACM:2013:SPF", pages = "705--714", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488698", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Random $k$-SAT is the single most intensely studied example of a random constraint satisfaction problem. But despite substantial progress over the past decade, the threshold for the existence of satisfying assignments is not known precisely for any $ k \geq 3 $. The best current results, based on the second moment method, yield upper and lower bounds that differ by an additive $ k \cdot {\ln 2} / 2 $, a term that is unbounded in $k$ (Achlioptas, Peres: STOC 2003). The basic reason for this gap is the inherent asymmetry of the Boolean values 'true' and 'false' in contrast to the perfect symmetry, e.g., among the various colors in a graph coloring problem. Here we develop a new asymmetric second moment method that allows us to tackle this issue head on for the first time in the theory of random CSPs. This technique enables us to compute the $k$-SAT threshold up to an additive $ \ln 2 - 1 / 2 + O(1 / k) \sim 0.19 $. Independently of the rigorous work, physicists have developed a sophisticated but non-rigorous technique called the ``cavity method'' for the study of random CSPs (Mezard, Parisi, Zecchina: Science~2002). Our result matches the best bound that can be obtained from the so-called ``replica symmetric'' version of the cavity method, and indeed our proof directly harnesses parts of the physics calculations.", acknowledgement = ack-nhfb, } @InProceedings{Kol:2013:ICC, author = "Gillat Kol and Ran Raz", title = "Interactive channel capacity", crossref = "ACM:2013:SPF", pages = "715--724", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488699", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the interactive channel capacity of an $ \epsilon $-noisy channel. The interactive channel capacity $ C(\epsilon) $ is defined as the minimal ratio between the communication complexity of a problem (over a non-noisy channel), and the communication complexity of the same problem over the binary symmetric channel with noise rate $ \epsilon $, where the communication complexity tends to infinity. Our main result is the upper bound $ C(\epsilon) \leq 1 - \Omega (\sqrt H(\epsilon)) $. This compares with Shannon's non-interactive channel capacity of $ 1 - H(\epsilon) $. In particular, for a small enough $ \epsilon $, our result gives the first separation between interactive and non-interactive channel capacity, answering an open problem by Schulman [Schulman1]. We complement this result by the lower bound $ C(\epsilon) \geq 1 - O(\sqrt H(\epsilon)) $, proved for the case where the players take alternating turns.", acknowledgement = ack-nhfb, } @InProceedings{Bernstein:2013:MSP, author = "Aaron Bernstein", title = "Maintaining shortest paths under deletions in weighted directed graphs: [extended abstract]", crossref = "ACM:2013:SPF", pages = "725--734", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488701", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present an improved algorithm for maintaining all-pairs $ 1 + \epsilon $ approximate shortest paths under deletions and weight-increases. The previous state of the art for this problem was total update time $ \tilde O (n^2 \sqrt m / \epsilon) $ for directed, unweighted graphs [2], and $ \tilde O(m n / \epsilon) $ for undirected, unweighted graphs [12]. Both algorithms were randomized and had constant query time. Note that $ \tilde O(m n) $ is a natural barrier because even with a $ (1 + \epsilon) $ approximation, there is no $ o(m n) $ combinatorial algorithm for the static all-pairs shortest path problem. Our algorithm works on directed, weighted graphs and has total (randomized) update time $ \tilde O (m n \log (R) / \epsilon) $ where $R$ is the ratio of the largest edge weight ever seen in the graph, to the smallest such weight (our query time is constant). Note that $ \log (R) = O(\log (n)) $ as long as weights are polynomial in $n$. Although $ \tilde O(m n \log (R) / \epsilon) $ is the total time over all updates, our algorithm also requires a clearly unavoidable constant time per update. Thus, we effectively expand the $ \tilde O(m n) $ total update time bound from undirected, unweighted graphs to directed graphs with polynomial weights. This is in fact the first non-trivial algorithm for decremental all-pairs shortest paths that works on weighted graphs (previous algorithms could only handle small integer weights). By a well known reduction from decremental algorithms to fully dynamic ones [9], our improved decremental algorithm leads to improved query-update tradeoffs for fully dynamic $ (1 + \epsilon) $ approximate APSP algorithm in directed graphs.", acknowledgement = ack-nhfb, } @InProceedings{Eisenstat:2013:LTA, author = "David Eisenstat and Philip N. Klein", title = "Linear-time algorithms for max flow and multiple-source shortest paths in unit-weight planar graphs", crossref = "ACM:2013:SPF", pages = "735--744", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488702", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give simple linear-time algorithms for two problems in planar graphs: max st-flow in directed graphs with unit capacities, and multiple-source shortest paths in undirected graphs with unit lengths.", acknowledgement = ack-nhfb, } @InProceedings{Neiman:2013:SDA, author = "Ofer Neiman and Shay Solomon", title = "Simple deterministic algorithms for fully dynamic maximal matching", crossref = "ACM:2013:SPF", pages = "745--754", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488703", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A maximal matching can be maintained in fully dynamic (supporting both addition and deletion of edges) $n$-vertex graphs using a trivial deterministic algorithm with a worst-case update time of $ O(n) $. No deterministic algorithm that outperforms the naive $ O(n) $ one was reported up to this date. The only progress in this direction is due to Ivkovic and Lloyd [14], who in 1993 devised a deterministic algorithm with an amortized update time of O((n+m)$^{ \sqrt 2 / 2}$), where m is the number of edges. In this paper we show the first deterministic fully dynamic algorithm that outperforms the trivial one. Specifically, we provide a deterministic worst-case update time of $ O(\sqrt m) $. Moreover, our algorithm maintains a matching which is in fact a 3/2-approximate maximum cardinality matching (MCM). We remark that no fully dynamic algorithm for maintaining $ (2 - \epsilon) $-approximate MCM improving upon the naive $ O(n) $ was known prior to this work, even allowing amortized time bounds and randomization. For low arboricity graphs (e.g., planar graphs and graphs excluding fixed minors), we devise another simple deterministic algorithm with sub-logarithmic update time. Specifically, it maintains a fully dynamic maximal matching with amortized update time of $ O(\log n / \log \log n) $. This result addresses an open question of Onak and Rubinfeld [19]. We also show a deterministic algorithm with optimal space usage of $ O(n + m) $, that for arbitrary graphs maintains a maximal matching with amortized update time of $ O(\sqrt m) $.", acknowledgement = ack-nhfb, } @InProceedings{Lee:2013:NAC, author = "Yin Tat Lee and Satish Rao and Nikhil Srivastava", title = "A new approach to computing maximum flows using electrical flows", crossref = "ACM:2013:SPF", pages = "755--764", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488704", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We give an algorithm which computes a $ (1 - \epsilon) $-approximately maximum st-flow in an undirected uncapacitated graph in time $ O(1 / \epsilon \sqrt m / F \cdot m \log^2 n) $ where $F$ is the flow value. By trading this off against the Karger-Levine algorithm for undirected graphs which takes $ \tilde O(m + n F) $ time, we obtain a running time of $ \tilde O(m n^{1 / 3} / \epsilon^{2 / 3}) $ for uncapacitated graphs, improving the previous best dependence on $ \epsilon $ by a factor of $ O(1 / \epsilon^3) $. Like the algorithm of Christiano, Kelner, Madry, Spielman and Teng, our algorithm reduces the problem to electrical flow computations which are carried out in linear time using fast Laplacian solvers. However, in contrast to previous work, our algorithm does not reweight the edges of the graph in any way, and instead uses local (i.e., non s-t) electrical flows to reroute the flow on congested edges. The algorithm is simple and may be viewed as trying to find a point at the intersection of two convex sets (the affine subspace of st-flows of value $F$ and the $ l_\infty $ ball) by an accelerated version of the method of alternating projections due to Nesterov. By combining this with Ford and Fulkerson's augmenting paths algorithm, we obtain an exact algorithm with running time $ \tilde O(m^{5 / 4} F^{1 / 4}) $ for uncapacitated undirected graphs, improving the previous best running time of $ \tilde O(m + \min (n F, m^{3 / 2})) $. We give a related algorithm with the same running time for approximate minimum cut, based on minimizing a smoothed version of the $ l_1 $ norm inside the cut space of the input graph. We show that the minimizer of this norm is related to an approximate blocking flow and use this to give an algorithm for computing a length $k$ approximately blocking flow in time $ \tilde O(m \sqrt k) $.", acknowledgement = ack-nhfb, } @InProceedings{Orlin:2013:MFN, author = "James B. Orlin", title = "{Max} flows in {$ O(n m) $} time, or better", crossref = "ACM:2013:SPF", pages = "765--774", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488705", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In this paper, we present improved polynomial time algorithms for the max flow problem defined on sparse networks with $n$ nodes and $m$ arcs. We show how to solve the max flow problem in $ O(n m + m^{31 / 16} \log^2 n) $ time. In the case that $ m = O(n^{1.06}) $, this improves upon the best previous algorithm due to King, Rao, and Tarjan, who solved the max flow problem in $ O(n m \log_{m / (n \log n)} n) $ time. This establishes that the max flow problem is solvable in $ O(n m) $ time for all values of $n$ and $m$. In the case that $ m = O(n) $, we improve the running time to $ O(n^2 / \log n) $.", acknowledgement = ack-nhfb, } @InProceedings{Bringmann:2013:SSD, author = "Karl Bringmann and Kasper Green Larsen", title = "Succinct sampling from discrete distributions", crossref = "ACM:2013:SPF", pages = "775--782", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488707", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We revisit the classic problem of sampling from a discrete distribution: Given n non-negative w-bit integers x$_1$ ,..,x$_n$, the task is to build a data structure that allows sampling i with probability proportional to x$_i$. The classic solution is Walker's alias method that takes, when implemented on a Word RAM, $ O(n) $ preprocessing time, $ O(1) $ expected query time for one sample, and $ n(w + 2 \lg (n) + o(1)) $ bits of space. Using the terminology of succinct data structures, this solution has redundancy $ 2 n \lg (n) + o(n) $ bits, i.e., it uses $ 2 n \lg (n) + o(n) $ bits in addition to the information theoretic minimum required for storing the input. In this paper, we study whether this space usage can be improved. In the systematic case, in which the input is read-only, we present a novel data structure using $ r + O(w) $ redundant bits, $ O(n / r) $ expected query time and $ O(n) $ preprocessing time for any $r$. This is an improvement in redundancy by a factor of $ \Omega (\log n) $ over the alias method for $ r = n $, even though the alias method is not systematic. Moreover, we complement our data structure with a lower bound showing that this trade-off is tight for systematic data structures. In the non-systematic case, in which the input numbers may be represented in more clever ways than just storing them one-by-one, we demonstrate a very surprising separation from the systematic case: With only 1 redundant bit, it is possible to support optimal $ O(1) $ expected query time and $ O(n) $ preprocessing time! On the one hand, our results improve upon the space requirement of the classic solution for a fundamental sampling problem, on the other hand, they provide the strongest known separation between the systematic and non-systematic case for any data structure problem. Finally, we also believe our upper bounds are practically efficient and simpler than Walker's alias method.", acknowledgement = ack-nhfb, } @InProceedings{Li:2013:NIS, author = "Xin Li", title = "New independent source extractors with exponential improvement", crossref = "ACM:2013:SPF", pages = "783--792", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488708", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the problem of constructing explicit extractors for independent general weak random sources. For weak sources on $n$ bits with min-entropy $k$, previously the best known extractor needs to use at least $ \log n / \log k $ independent sources [22, 3]. In this paper we give a new extractor that only uses $ O(\log (\log n / \log k)) + O(1) $ independent sources. Thus, our result improves the previous best result exponentially. We then use our new extractor to give improved network extractor protocols, as defined in [14]. The network extractor protocols also give new results in distributed computing with general weak random sources, which dramatically improve previous results. For example, we can tolerate a nearly optimal fraction of faulty players in synchronous Byzantine agreement and leader election, even if the players only have access to independent $n$-bit weak random sources with min-entropy as small as $ k = \polylog (n) $. Our extractor for independent sources is based on a new condenser for somewhere random sources with a special structure. We believe our techniques are interesting in their own right and are promising for further improvement.", acknowledgement = ack-nhfb, } @InProceedings{Rothblum:2013:IPP, author = "Guy N. Rothblum and Salil Vadhan and Avi Wigderson", title = "Interactive proofs of proximity: delegating computation in sublinear time", crossref = "ACM:2013:SPF", pages = "793--802", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488709", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study interactive proofs with sublinear-time verifiers. These proof systems can be used to ensure approximate correctness for the results of computations delegated to an untrusted server. Following the literature on property testing, we seek proof systems where with high probability the verifier accepts every input in the language, and rejects every input that is far from the language. The verifier's query complexity (and computation complexity), as well as the communication, should all be sublinear. We call such a proof system an Interactive Proof of Proximity (IPP). On the positive side, our main result is that all languages in NC have Interactive Proofs of Proximity with roughly $ \sqrt n $ query and communication and complexities, and $ \polylog (n) $ communication rounds. This is achieved by identifying a natural language, membership in an affine subspace (for a structured class of subspaces), that is complete for constructing interactive proofs of proximity, and providing efficient protocols for it. In building an IPP for this complete language, we show a tradeoff between the query and communication complexity and the number of rounds. For example, we give a 2-round protocol with roughly $ n^{3 / 4} $ queries and communication. On the negative side, we show that there exist natural languages in N$ C^1 $, for which the sum of queries and communication in any constant-round interactive proof of proximity must be polynomially related to $n$. In particular, for any 2-round protocol, the sum of queries and communication must be at least $ \tilde \Omega (\sqrt n) $. Finally, we construct much better IPPs for specific functions, such as bipartiteness on random or well-mixing graphs, and the majority function. The query complexities of these protocols are provably better (by exponential or polynomial factors) than what is possible in the standard property testing model, i.e. without a prover.", acknowledgement = ack-nhfb, } @InProceedings{Ajtai:2013:LBR, author = "Miklos Ajtai", title = "Lower bounds for {RAMs} and quantifier elimination", crossref = "ACM:2013:SPF", pages = "803--812", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488710", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "For each natural number $d$ we consider a finite structure $ M_d $ whose universe is the set of all $ 0, 1 $-sequence of length $ n = 2^d $, each representing a natural number in the set $ \{ 0, 1, \ldots {}, 2^n - 1 \} $ in binary form. The operations included in the structure are the four constants $ 0, 1, 2^n - 1, n $, multiplication and addition modulo $ 2^n $, the unary function $ \{ \min 2^x, 2^n - 1 \} $, the binary functions $ \lfloor x / y \rfloor $ (with $ \lfloor x / 0 \rfloor = 0 $), $ \max (x, y) $, $ \min (x, y) $, and the boolean vector operations, $ {\rm vee} $, $-$ defined on $ 0, 1 $ sequences of length $n$, by performing the operations on all components simultaneously. These are essentially the arithmetic operations that can be performed on a RAM, with wordlength $n$, by a single instruction. We show that there exists an $ \epsilon > 0 $ and a term (that is, an algebraic expression) $ F(x, y) $ built up from the mentioned operations, with the only free variables $x$, $y$, such that if $ G_d(y), d = 0, 1, 2, \ldots {} $, is a sequence of terms, and for all $ d = 0, 1, 2, \ldots {}, M_d $ models $ \forall x, [G_d (x) \to \exists y, F(x, y) = 0] $, then for infinitely many integers $d$, the depth of the term $ G_d $, that is, the maximal number of nestings of the operations in it, is at least $ \epsilon (\log d)^{1 / 2} = \epsilon (\log \log n)^{1 / 2} $. The following is a consequence. We are considering RAMs $ N_n $, with wordlength $ n = 2^d $, whose arithmetic instructions are the arithmetic operations listed above, and also have the usual other RAM instructions. The size of the memory is restricted only by the address space, that is, it is $ 2^n $ words. The RAMs has a finite instruction set, each instruction is encoded by a fixed natural number independently of $n$. Therefore a program $P$ can run on each machine $ N_n $, if $ n = 2^d $ is sufficiently large. We show that there exists an $ \epsilon > 0 $ and a program $P$, such that it satisfies the following two conditions. (i) For all sufficiently large $ n = 2^d $, if $P$ running on N$_n$ gets an input consisting of two words $a$ and $b$, then, in constant time, it gives a $ 0, 1 $ output $ P_n(a, b) $. (ii) Suppose that $Q$ is a program such that for each sufficiently large $ n = 2^d $, if $Q$, running on $ N_n $, gets a word $a$ of length $n$ as an input, then it decides whether there exists a word $b$ of length $n$ such that $ P_n(a, b) = 0. $ Then, for infinitely many positive integers $d$, there exists a word $a$ of length $ n = 2^d $, such that the running time of $Q$ on $ N_n $ at input $a$ is at least $ \epsilon (\log d)^{1 / 2} (\log \log d)^{-1} \geq (\log d)^{1 / 2 - \epsilon } = (\log \log n)^{1 / 2 - \epsilon } $.", acknowledgement = ack-nhfb, } @InProceedings{Beck:2013:STR, author = "Chris Beck and Jakob Nordstrom and Bangsheng Tang", title = "Some trade-off results for polynomial calculus: extended abstract", crossref = "ACM:2013:SPF", pages = "813--822", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488711", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present size-space trade-offs for the polynomial calculus (PC) and polynomial calculus resolution (PCR) proof systems. These are the first true size-space trade-offs in any algebraic proof system, showing that size and space cannot be simultaneously optimized in these models. We achieve this by extending essentially all known size-space trade-offs for resolution to PC and PCR. As such, our results cover space complexity from constant all the way up to exponential and yield mostly superpolynomial or even exponential size blow-ups. Since the upper bounds in our trade-offs hold for resolution, our work shows that there are formulas for which adding algebraic reasoning on top of resolution does not improve the trade-off properties in any significant way. As byproducts of our analysis, we also obtain trade-offs between space and degree in PC and PCR exactly matching analogous results for space versus width in resolution, and strengthen the resolution trade-offs in [Beame, Beck, and Impagliazzo '12] to apply also to $k$-CNF formulas.", acknowledgement = ack-nhfb, } @InProceedings{Bhowmick:2013:NBM, author = "Abhishek Bhowmick and Zeev Dvir and Shachar Lovett", title = "New bounds for matching vector families", crossref = "ACM:2013:SPF", pages = "823--832", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488713", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A Matching Vector (MV) family modulo m is a pair of ordered lists $ U = (u_1, \ldots {}, u_t) $ and $ V = (v_1, \ldots {}, v_t) $ where $ u_i, v_j \in Z_m^n $ with the following inner product pattern: for any $ i, \{ u_i, v_i \} = 0 $, and for any $ i \neq j $, $ \{ u_i, v_j \} \neq 0 $. A MV family is called $q$-restricted if inner products $ \{ u_i, v_j \} $ take at most $q$ different values. Our interest in MV families stems from their recent application in the construction of sub-exponential locally decodable codes (LDCs). There, $q$-restricted MV families are used to construct LDCs with $q$ queries, and there is special interest in the regime where $q$ is constant. When $m$ is a prime it is known that such constructions yield codes with exponential block length. However, for composite $m$ the behaviour is dramatically different. A recent work by Efremenko [8] (based on an approach initiated by Yekhanin [24]) gives the first sub-exponential LDC with constant queries. It is based on a construction of a MV family of super-polynomial size by Grolmusz [10] modulo composite $m$. In this work, we prove two lower bounds on the block length of LDCs which are based on black box construction using MV families. When $q$ is constant (or sufficiently small), we prove that such LDCs must have a quadratic block length. When the modulus $m$ is constant (as it is in the construction of Efremenko [8]) we prove a super-polynomial lower bound on the block-length of the LDCs, assuming a well-known conjecture in additive combinatorics, the polynomial Freiman-Ruzsa conjecture over Z$_m$.", acknowledgement = ack-nhfb, } @InProceedings{Ben-Sasson:2013:NFL, author = "Eli Ben-Sasson and Ariel Gabizon and Yohay Kaplan and Swastik Kopparty and Shubangi Saraf", title = "A new family of locally correctable codes based on degree-lifted algebraic geometry codes", crossref = "ACM:2013:SPF", pages = "833--842", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488714", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We describe new constructions of error correcting codes, obtained by ``degree-lifting'' a short algebraic geometry base-code of block-length $q$ to a lifted-code of block-length $ q^m $, for arbitrary integer $m$. The construction generalizes the way degree-d, univariate polynomials evaluated over the $q$-element field (also known as Reed--Solomon codes) are ``lifted'' to degree-d, $m$-variate polynomials (Reed--Muller codes). A number of properties are established: The rate of the degree-lifted code is approximately a 1/m!-fraction of the rate of the base-code. The relative distance of the degree-lifted code is at least as large as that of the base-code. This is proved using a generalization of the Schwartz-Zippel Lemma to degree-lifted Algebraic-Geometry codes. [Local correction] If the base code is invariant under a group that is ``close'' to being doubly-transitive (in a precise manner defined later) then the degree-lifted code is locally correctable with query complexity at most $ q^2 $. The automorphisms of the base-code are crucially used to generate query-sets, abstracting the use of affine-lines in the local correction procedure of Reed--Muller codes. Taking a concrete illustrating example, we show that degree-lifted Hermitian codes form a family of locally correctable codes over an alphabet that is significantly smaller than that obtained by Reed--Muller codes of similar constant rate, message length, and distance.", acknowledgement = ack-nhfb, } @InProceedings{Guruswami:2013:LDR, author = "Venkatesan Guruswami and Chaoping Xing", title = "List decoding {Reed--Solomon}, algebraic-geometric, and {Gabidulin} subcodes up to the singleton bound", crossref = "ACM:2013:SPF", pages = "843--852", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488715", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider Reed--Solomon (RS) codes whose evaluation points belong to a subfield, and give a linear-algebraic list decoding algorithm that can correct a fraction of errors approaching the code distance, while pinning down the candidate messages to a well-structured affine space of dimension a constant factor smaller than the code dimension. By pre-coding the message polynomials into a subspace-evasive set, we get a Monte Carlo construction of a subcode of Reed--Solomon codes that can be list decoded from a fraction $ (1 - R - \epsilon) $ of errors in polynomial time (for any fixed $ \epsilon > 0 $) with a list size of $ O(1 / \epsilon) $. Our methods extend to algebraic-geometric (AG) codes, leading to a similar claim over constant-sized alphabets. This matches parameters of recent results based on folded variants of RS and AG codes. but our construction here gives subcodes of Reed--Solomon and AG codes themselves (albeit with restrictions on the evaluation points). Further, the underlying algebraic idea also extends nicely to Gabidulin's construction of rank-metric codes based on linearized polynomials. This gives the first construction of positive rate rank-metric codes list decodable beyond half the distance, and in fact gives codes of rate $R$ list decodable up to the optimal $ (1 - R - \epsilon) $ fraction of rank errors. A similar claim holds for the closely related subspace codes studied by Koetter and Kschischang. We introduce a new notion called subspace designs as another way to pre-code messages and prune the subspace of candidate solutions. Using these, we also get a deterministic construction of a polynomial time list decodable subcode of RS codes. By using a cascade of several subspace designs, we extend our approach to AG codes, which gives the first deterministic construction of an algebraic code family of rate $R$ with efficient list decoding from $ 1 - R - \epsilon $ fraction of errors over an alphabet of constant size (that depends only on $ \epsilon $). The list size bound is almost a constant (governed by $ \log * $ (block length)), and the code can be constructed in quasi-polynomial time.", acknowledgement = ack-nhfb, } @InProceedings{Wootters:2013:LDR, author = "Mary Wootters", title = "On the list decodability of random linear codes with large error rates", crossref = "ACM:2013:SPF", pages = "853--860", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488716", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "It is well known that a random $q$-ary code of rate $ \Omega (\epsilon^2) $ is list decodable up to radius $ (1 - 1 / q - \epsilon) $ with list sizes on the order of $ 1 / \epsilon^2 $, with probability $ 1 - o(1) $. However, until recently, a similar statement about random linear codes has until remained elusive. In a recent paper, Cheraghchi, Guruswami, and Velingker show a connection between list decodability of random linear codes and the Restricted Isometry Property from compressed sensing, and use this connection to prove that a random linear code of rate $ \Omega (\epsilon^2 / \log^3 (1 / \epsilon)) $ achieves the list decoding properties above, with constant probability. We improve on their result to show that in fact we may take the rate to be $ \Omega (\epsilon^2) $, which is optimal, and further that the success probability is $ 1 - o(1) $, rather than constant. As an added benefit, our proof is relatively simple. Finally, we extend our methods to more general ensembles of linear codes. As an example, we show that randomly punctured Reed--Muller codes have the same list decoding properties as the original codes, even when the rate is improved to a constant.", acknowledgement = ack-nhfb, } @InProceedings{Brandao:2013:QFT, author = "Fernando G. S. L. Brandao and Aram W. Harrow", title = "Quantum {de Finetti} theorems under local measurements with applications", crossref = "ACM:2013:SPF", pages = "861--870", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488718", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Quantum de Finetti theorems are a useful tool in the study of correlations in quantum multipartite states. In this paper we prove two new quantum de Finetti theorems, both showing that under tests formed by local measurements in each of the subsystems one can get a much improved error dependence on the dimension of the subsystems. We also obtain similar results for non-signaling probability distributions. We give the following applications of the results to quantum complexity theory, polynomial optimization, and quantum information theory: We prove the optimality of the Chen-Drucker protocol for 3-SAT, under the assumption there is no subexponential-time algorithm for SAT. In the protocol a prover sends to a verifier $ \sqrt n \polylog (n) $ unentangled quantum states, each composed of $ O(\log (n)) $ qubits, as a proof of the satisfiability of a 3-SAT instance with $n$ variables and $ O(n) $ clauses. The quantum verifier checks the validity of the proof by performing local measurements on each of the proofs and classically processing the outcomes. We show that any similar protocol with $ O(n^{1 / 2 - \epsilon }) $ qubits would imply a $ \exp (n^{1 - 2 \epsilon } \polylog (n)) $-time algorithm for 3-SAT. We show that the maximum winning probability of free games (in which the questions to each prover are chosen independently) can be estimated by linear programming in time $ \exp (O(\log |Q| + \log^2 |A| / \epsilon^2)) $, with $ |Q| $ and $ |A| $ the question and answer alphabet sizes, respectively, matching the performance of a previously known algorithm due to Aaronson, Impagliazzo, Moshkovitz, and Shor. This result follows from a new monogamy relation for non-locality, showing that $k$-extendible non-signaling distributions give at most a $ O(k^{-1 / 2}) $ advantage over classical strategies for free games. We also show that 3-SAT with $n$ variables can be reduced to obtaining a constant error approximation of the maximum winning probability under entangled strategies of $ O(\sqrt n) $-player one-round non-local games, in which only two players are selected to send $ O(\sqrt n) $-bit messages. We show that the optimization of certain polynomials over the complex hypersphere can be performed in quasipolynomial time in the number of variables $n$ by considering $ O(\log (n)) $ rounds of the Sum-of-Squares (Parrilo/Lasserre) hierarchy of semidefinite programs. This can be considered an analogue to the hypersphere of a similar known results for the simplex. As an application to entanglement theory, we find a quasipolynomial-time algorithm for deciding multipartite separability. We consider a quantum tomography result due to Aaronson --- showing that given an unknown $n$-qubit state one can perform tomography that works well for most observables by measuring only $ O(n) $ independent and identically distributed (i.i.d.) copies of the state --- and relax the assumption of having i.i.d copies of the state to merely the ability to select subsystems at random from a quantum multipartite state. The proofs of the new quantum de Finetti theorems are based on information theory, in particular on the chain rule of mutual information. The results constitute improvements and generalizations of a recent de Finetti theorem due to Brandao, Christandl and Yard.", acknowledgement = ack-nhfb, } @InProceedings{Brandao:2013:PSA, author = "Fernando G. S. L. Brandao and Aram W. Harrow", title = "Product-state approximations to quantum ground states", crossref = "ACM:2013:SPF", pages = "871--880", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488719", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "The local Hamiltonian problem consists of estimating the ground-state energy (given by the minimum eigenvalue) of a local quantum Hamiltonian. It can be considered as a quantum generalization of constraint satisfaction problems (CSPs) and has a key role in quantum complexity theory, being the first and most natural QMA -complete problem known. An interesting regime for the local Hamiltonian problem is that of extensive error, where one is interested in estimating the mean ground-state energy to constant accuracy. The problem is NP -hard by the PCP theorem, but whether it is QMA -hard is an important open question in quantum complexity theory. A positive solution would represent a quantum analogue of the PCP theorem. A key feature that distinguishes quantum Hamiltonians from classical CSPs is that the solutions may involve complicated entangled states. In this paper, we demonstrate several large classes of Hamiltonians for which product (i.e. unentangled) states can approximate the ground state energy to within a small extensive error. First, we show the mere existence of a good product-state approximation for the ground-state energy of 2-local Hamiltonians with one of more of the following properties: (1) super-constant degree, (2) small expansion, or (3) a ground state with sublinear entanglement with respect to some partition into small pieces. The approximation based on degree is a new and surprising difference between quantum Hamiltonians and classical CSPs, since in the classical setting, higher degree is usually associated with harder CSPs. The approximation based on expansion is not new, but the approximation based on low entanglement was previously known only in the regime where the entanglement was close to zero. Since the existence of a low-energy product state can be checked in NP, this implies that any Hamiltonian used for a quantum PCP theorem should have: (1) constant degree, (2) constant expansion, (3) a ``volume law'' for entanglement with respect to any partition into small parts. Second, we show that in several cases, good product-state approximations not only exist, but can be found in deterministic polynomial time: (1) 2-local Hamiltonians on any planar graph, solving an open problem of Bansal, Bravyi, and Terhal, (2) dense k -local Hamiltonians for any constant k, solving an open problem of Gharibian and Kempe, and (3) 2-local Hamiltonians on graphs with low threshold rank, via a quantum generalization of a recent result of Barak, Raghavendra and Steurer. Our work involves two new tools which may be of independent interest. First, we prove a new quantum version of the de Finetti theorem which does not require the usual assumption of symmetry. Second, we describe a way to analyze the application of the Lasserre/Parrilo SDP hierarchy to local quantum Hamiltonians.", acknowledgement = ack-nhfb, } @InProceedings{Ta-Shma:2013:IWC, author = "Amnon Ta-Shma", title = "Inverting well conditioned matrices in quantum logspace", crossref = "ACM:2013:SPF", pages = "881--890", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488720", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We show that quantum computers improve on the best known classical algorithms for matrix inversion (and singular value decomposition) as far as space is concerned. This adds to the (still short) list of important problems where quantum computers are of help. Specifically, we show that the inverse of a well conditioned matrix can be approximated in quantum logspace with intermediate measurements. This should be compared with the best known classical algorithm for the problem that requires $ \Omega (\log^2 n) $ space. We also show how to approximate the spectrum of a normal matrix, or the singular values of an arbitrary matrix, with $ \epsilon $ additive accuracy, and how to approximate the singular value decomposition (SVD) of a matrix whose singular values are well separated. The technique builds on ideas from several previous works, including simulating Hamiltonians in small quantum space (building on [2] and [10]), treating a Hermitian matrix as a Hamiltonian and running the quantum phase estimation procedure on it (building on [5]) and making small space probabilistic (and quantum) computation consistent through the use of offline randomness and the shift and truncate method (building on [8]).", acknowledgement = ack-nhfb, } @InProceedings{Ambainis:2013:SAE, author = "Andris Ambainis", title = "Superlinear advantage for exact quantum algorithms", crossref = "ACM:2013:SPF", pages = "891--900", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488721", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "A quantum algorithm is exact if, on any input data, it outputs the correct answer with certainty (probability 1). A key question is: how big is the advantage of exact quantum algorithms over their classical counterparts: deterministic algorithms. For total Boolean functions in the query model, the biggest known gap was just a factor of $2$: PARITY of $N$ input bits requires $N$ queries classically but can be computed with $ N / 2 $ queries by an exact quantum algorithm. We present the first example of a Boolean function $ f(x_1, \ldots {}, x_N) $ for which exact quantum algorithms have superlinear advantage over deterministic algorithms. Any deterministic algorithm that computes our function must use $N$ queries but an exact quantum algorithm can compute it with $ O(N^{0.8675 \ldots }) $ queries. A modification of our function gives a similar result for communication complexity: there is a function f which can be computed by an exact quantum protocol that communicates $ O(N^{0.8675 \ldots }) $ quantum bits but requires $ \Omega (N) $ bits of communication for classical protocols.", acknowledgement = ack-nhfb, } @InProceedings{Li:2013:AKM, author = "Shi Li and Ola Svensson", title = "Approximating $k$-median via pseudo-approximation", crossref = "ACM:2013:SPF", pages = "901--910", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488723", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We present a novel approximation algorithm for $k$-median that achieves an approximation guarantee of $ 1 + \sqrt 3 + \epsilon $ , improving upon the decade-old ratio of $ 3 + \epsilon $ . Our approach is based on two components, each of which, we believe, is of independent interest. First, we show that in order to give an $ \alpha $ -approximation algorithm for $k$-median, it is sufficient to give a pseudo-approximation algorithm that finds an $ \alpha $-approximate solution by opening $ k + O(1) $ facilities. This is a rather surprising result as there exist instances for which opening $ k + 1 $ facilities may lead to a significant smaller cost than if only $k$ facilities were opened. Second, we give such a pseudo-approximation algorithm with $ \alpha = 1 + \sqrt 3 + \epsilon $. Prior to our work, it was not even known whether opening $ k + o(k) $ facilities would help improve the approximation ratio.", acknowledgement = ack-nhfb, } @InProceedings{Kelner:2013:SCA, author = "Jonathan A. Kelner and Lorenzo Orecchia and Aaron Sidford and Zeyuan Allen Zhu", title = "A simple, combinatorial algorithm for solving {SDD} systems in nearly-linear time", crossref = "ACM:2013:SPF", pages = "911--920", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488724", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In this paper, we present a simple combinatorial algorithm that solves symmetric diagonally dominant (SDD) linear systems in nearly-linear time. It uses little of the machinery that previously appeared to be necessary for a such an algorithm. It does not require recursive preconditioning, spectral sparsification, or even the Chebyshev Method or Conjugate Gradient. After constructing a ``nice'' spanning tree of a graph associated with the linear system, the entire algorithm consists of the repeated application of a simple update rule, which it implements using a lightweight data structure. The algorithm is numerically stable and can be implemented without the increased bit-precision required by previous solvers. As such, the algorithm has the fastest known running time under the standard unit-cost RAM model. We hope the simplicity of the algorithm and the insights yielded by its analysis will be useful in both theory and practice.", acknowledgement = ack-nhfb, } @InProceedings{Sherstov:2013:CLB, author = "Alexander A. Sherstov", title = "Communication lower bounds using directional derivatives", crossref = "ACM:2013:SPF", pages = "921--930", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488725", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the set disjointness problem in the most powerful bounded-error model: the number-on-the-forehead model with $k$ parties and arbitrary classical or quantum communication. We obtain a communication lower bound of $ \Omega (\sqrt (n) / 2^k *k) $ bits, which is essentially optimal. Proving it was a longstanding open problem even in restricted settings, such as one-way classical protocols with $ k = 4 $ parties (Wigderson 1997). The proof contributes a novel technique for lower bounds on multiparty communication, based on directional derivatives of communication protocols over the reals.", acknowledgement = ack-nhfb, } @InProceedings{Andoni:2013:HFU, author = "Alexandr Andoni and Assaf Goldberger and Andrew McGregor and Ely Porat", title = "Homomorphic fingerprints under misalignments: sketching edit and shift distances", crossref = "ACM:2013:SPF", pages = "931--940", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488726", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "Fingerprinting is a widely-used technique for efficiently verifying that two files are identical. More generally, linear sketching is a form of lossy compression (based on random projections) that also enables the ``dissimilarity'' of non-identical files to be estimated. Many sketches have been proposed for dissimilarity measures that decompose coordinate-wise such as the Hamming distance between alphanumeric strings, or the Euclidean distance between vectors. However, virtually nothing is known on sketches that would accommodate alignment errors. With such errors, Hamming or Euclidean distances are rendered useless: a small misalignment may result in a file that looks very dissimilar to the original file according such measures. In this paper, we present the first linear sketch that is robust to a small number of alignment errors. Specifically, the sketch can be used to determine whether two files are within a small Hamming distance of being a cyclic shift of each other. Furthermore, the sketch is homomorphic with respect to rotations: it is possible to construct the sketch of a cyclic shift of a file given only the sketch of the original file. The relevant dissimilarity measure, known as the shift distance, arises in the context of embedding edit distance and our result addressed an open problem [Question 13 in Indyk-McGregor-Newman-Onak'11] with a rather surprising outcome. Our sketch projects a length $n$ file into $ D(n) \cdot \polylog n $ dimensions where $ D(n)l n $ is the number of divisors of $n$. The striking fact is that this is near-optimal, i.e., the $ D(n) $ dependence is inherent to a problem that is ostensibly about lossy compression. In contrast, we then show that any sketch for estimating the edit distance between two files, even when small, requires sketches whose size is nearly linear in $n$. This lower bound addresses a long-standing open problem on the low distortion embeddings of edit distance [Question 2.15 in Naor-Matousek'11, Indyk'01], for the case of linear embeddings.", acknowledgement = ack-nhfb, } @InProceedings{Chonev:2013:OPH, author = "Ventsislav Chonev and Jo{\"e}l Ouaknine and James Worrell", title = "The orbit problem in higher dimensions", crossref = "ACM:2013:SPF", pages = "941--950", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488728", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We consider higher-dimensional versions of Kannan and Lipton's Orbit Problem---determining whether a target vector space V may be reached from a starting point x under repeated applications of a linear transformation A. Answering two questions posed by Kannan and Lipton in the 1980s, we show that when V has dimension one, this problem is solvable in polynomial time, and when V has dimension two or three, the problem is in NP$^{RP}$.", acknowledgement = ack-nhfb, } @InProceedings{Azar:2013:LSD, author = "Yossi Azar and Ilan Reuven Cohen and Iftah Gamzu", title = "The loss of serving in the dark", crossref = "ACM:2013:SPF", pages = "951--960", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488729", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study the following balls and bins stochastic process: There is a buffer with B bins, and there is a stream of balls X = {X$_1$, X$_2$, \ldots{} ,X$_T$ } such that X$_i$ is the number of balls that arrive before time i but after time i-1. Once a ball arrives, it is stored in one of the unoccupied bins. If all the bins are occupied then the ball is thrown away. In each time step, we select a bin uniformly at random, clear it, and gain its content. Once the stream of balls ends, all the remaining balls in the buffer are cleared and added to our gain. We are interested in analyzing the expected gain of this randomized process with respect to that of an optimal gain-maximizing strategy, which gets the same online stream of balls, and clears a ball from a bin, if exists, at any step. We name this gain ratio the loss of serving in the dark. In this paper, we determine the exact loss of serving in the dark. We prove that the expected gain of the randomized process is worse by a factor of $ \rho + \epsilon $ from that of the optimal gain-maximizing strategy for any $ \epsilon > 0 $, where $ \rho = \max_{ \alpha > 1} \alpha e^{ \alpha } / ((\alpha - 1)e^{ \alpha } + e - 1) \sim 1.69996 $ and $ B = \Omega (1 / \epsilon^3) $. We also demonstrate that this bound is essentially tight as there are specific ball streams for which the above-mentioned gain ratio tends to $ \rho $. Our stochastic process occurs naturally in many applications. We present a prompt and truthful mechanism for bounded capacity auctions, and an application relating to packets scheduling.", acknowledgement = ack-nhfb, } @InProceedings{Azar:2013:TBO, author = "Yossi Azar and Ilan Reuven Cohen and Seny Kamara and Bruce Shepherd", title = "Tight bounds for online vector bin packing", crossref = "ACM:2013:SPF", pages = "961--970", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488730", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "In the $d$-dimensional bin packing problem (VBP), one is given vectors $ x_1, x_2, \ldots {}, x_n \in R^d $ and the goal is to find a partition into a minimum number of feasible sets: $ \{ 1, 2, \ldots {}, n \} = \cup_i^s B_i $. A set $ B_i $ is feasible if $ \Sigma_{j \in B i} x_j \leq 1 $, where $1$ denotes the all $1$'s vector. For online VBP, it has been outstanding for almost 20 years to clarify the gap between the best lower bound $ \Omega (1) $ on the competitive ratio versus the best upper bound of $ O(d) $. We settle this by describing a $ \Omega (d^{1 - \epsilon }) $ lower bound. We also give strong lower bounds (of $ \Omega (d^{1 / B - \epsilon }) $) if the bin size $ B \in Z_+ $ is allowed to grow. Finally, we discuss almost-matching upper bound results for general values of $B$; we show an upper bound whose exponent is additively ``shifted by 1'' from the lower bound exponent.", acknowledgement = ack-nhfb, } @InProceedings{Li:2013:SCO, author = "Jian Li and Wen Yuan", title = "Stochastic combinatorial optimization via {Poisson} approximation", crossref = "ACM:2013:SPF", pages = "971--980", year = "2013", DOI = "http://dx.doi.org/10.1145/2488608.2488731", bibdate = "Mon Mar 3 06:30:33 MST 2014", bibsource = "http://portal.acm.org/; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", abstract = "We study several stochastic combinatorial problems, including the expected utility maximization problem, the stochastic knapsack problem and the stochastic bin packing problem. A common technical challenge in these problems is to optimize some function (other than the expectation) of the sum of a set of random variables. The difficulty is mainly due to the fact that the probability distribution of the sum is the convolution of a set of distributions, which is not an easy objective function to work with. To tackle this difficulty, we introduce the Poisson approximation technique. The technique is based on the Poisson approximation theorem discovered by Le Cam, which enables us to approximate the distribution of the sum of a set of random variables using a compound Poisson distribution. Using the technique, we can reduce a variety of stochastic problems to the corresponding deterministic multiple-objective problems, which either can be solved by standard dynamic programming or have known solutions in the literature. For the problems mentioned above, we obtain the following results: We first study the expected utility maximization problem introduced recently [Li and Despande, FOCS11]. For monotone and Lipschitz utility functions, we obtain an additive PTAS if there is a multidimensional PTAS for the multi-objective version of the problem, strictly generalizing the previous result. The result implies the first additive PTAS for maximizing threshold probability for the stochastic versions of global min-cut, matroid base and matroid intersection. For the stochastic bin packing problem (introduced in [Kleinberg, Rabani and Tardos, STOC97]), we show there is a polynomial time algorithm which uses at most the optimal number of bins, if we relax the size of each bin and the overflow probability by e for any constant $ \epsilon > 0 $. Based on this result, we obtain a 3-approximation if only the size of each bin can be relaxed by $ \epsilon $ , improving the known $ O(1 / \epsilon) $ factor for constant overflow probability. For stochastic knapsack, we show a $ (1 + \epsilon) $-approximation using $ \epsilon $ extra capacity for any $ \epsilon > 0 $, even when the size and reward of each item may be correlated and cancelations of items are allowed. This generalizes the previous work [Balghat, Goel and Khanna, SODA11] for the case without correlation and cancelation. Our algorithm is also simpler. We also present a factor $ 2 + \epsilon $ approximation algorithm for stochastic knapsack with cancelations, for any constant $ \epsilon > 0 $, improving the current known approximation factor of 8 [Gupta, Krishnaswamy, Molinaro and Ravi, FOCS11]. We also study an interesting variant of the stochastic knapsack problem, where the size and the profit of each item are revealed before the decision is made. The problem falls into the framework of Bayesian online selection problems, which has been studied a lot recently.", acknowledgement = ack-nhfb, } %%% ==================================================================== %%% Cross-referenced entries must come last: @Proceedings{ACM:2006:SPT, editor = "{ACM}", booktitle = "{STOC'06: Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing 2006, Seattle, WA, USA, May 21--23, 2006}", title = "{STOC'06: Proceedings of the Thirty-Eighth Annual ACM Symposium on Theory of Computing 2006, Seattle, WA, USA, May 21--23, 2006}", publisher = pub-ACM, address = pub-ACM:adr, pages = "770 (est.)", year = "2006", ISBN = "1-59593-134-1", ISBN-13 = "978-1-59593-134-4", LCCN = "QA75.5 .A22 2006", bibdate = "Thu May 25 06:13:58 2006", bibsource = "http://www.math.utah.edu/pub/tex/bib/stoc.bib; http://www.math.utah.edu/pub/tex/bib/stoc2000.bib; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib; z3950.gbv.de:20011/gvk", note = "ACM order number 508060.", URL = "http://portal.acm.org/citation.cfm?id=1132516", acknowledgement = ack-nhfb, } @Proceedings{ACM:2010:SPA, editor = "{ACM}", booktitle = "{STOC'10: Proceedings of the 2010 ACM International Symposium on Theory of Computing: June 5--8, 2010, Cambridge, MA, USA}", title = "{STOC'10: Proceedings of the 2010 ACM International Symposium on Theory of Computing: June 5--8, 2010, Cambridge, MA, USA}", publisher = pub-ACM, address = pub-ACM:adr, pages = "xiv + 797", year = "2010", ISBN = "1-60558-817-2", ISBN-13 = "978-1-60558-817-9", LCCN = "QA 76.6 .A152 2010", bibdate = "Wed Sep 1 10:37:53 MDT 2010", bibsource = "z3950.gbv.de:20011/gvk; http://www.math.utah.edu/pub/tex/bib/stoc.bib; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", URL = "http://www.gbv.de/dms/tib-ub-hannover/63314455x.", acknowledgement = ack-nhfb, remark = "42nd annual STOC meeting.", } @Proceedings{ACM:2011:SPA, editor = "{ACM}", booktitle = "{STOC'11: Proceedings of the 2011 ACM International Symposium on Theory of Computing: June 6--8, 2011, San Jose, CA, USA}", title = "{STOC'11: Proceedings of the 2011 ACM International Symposium on Theory of Computing: June 6--8, 2011, San Jose, CA, USA}", publisher = pub-ACM, address = pub-ACM:adr, pages = "xxx + 822 (est.)", year = "2011", ISBN = "1-4503-0691-8", ISBN-13 = "978-1-4503-0691-1", LCCN = "????", bibdate = "Wed Sep 1 10:37:53 MDT 2010", bibsource = "z3950.gbv.de:20011/gvk; http://www.math.utah.edu/pub/tex/bib/stoc.bib; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", URL = "http://www.gbv.de/dms/tib-ub-hannover/63314455x.", acknowledgement = ack-nhfb, remark = "43rd annual STOC meeting.", } @Proceedings{ACM:2012:SPA, editor = "{ACM}", booktitle = "{STOC'12: Proceedings of the 2012 ACM International Symposium on Theory of Computing: May 19--22, 2012, New York, NY, USA}", title = "{STOC'12: Proceedings of the 2012 ACM International Symposium on Theory of Computing: May 19--22, 2012, New York, NY, USA}", publisher = pub-ACM, address = pub-ACM:adr, pages = "1292 (est.)", year = "2012", ISBN = "1-4503-1245-4", ISBN-13 = "978-1-4503-1245-5", LCCN = "????", bibdate = "Thu Nov 08 19:12:21 2012", bibsource = "http://www.math.utah.edu/pub/tex/bib/stoc2010.bib; http://www.math.utah.edu/pub/tex/bib/stoc.bib; z3950.gbv.de:20011/gvk", URL = "http://www.gbv.de/dms/tib-ub-hannover/63314455x.", acknowledgement = ack-nhfb, remark = "44th annual STOC meeting.", } @Proceedings{ACM:2013:SPF, editor = "{ACM}", booktitle = "{STOC '13: Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing: June 1--4, 2013, Palo Alto, California, USA}", title = "{STOC '13: Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing: June 1--4, 2013, Palo Alto, California, USA}", publisher = pub-ACM, address = pub-ACM:adr, pages = "980 (est.)", year = "2013", ISBN = "1-4503-2029-5", ISBN-13 = "978-1-4503-2029-0", bibdate = "Mon Mar 3 06:36:05 2014", bibsource = "http://www.math.utah.edu/pub/tex/bib/datacompression.bib; http://www.math.utah.edu/pub/tex/bib/prng.bib; http://www.math.utah.edu/pub/tex/bib/stoc.bib; http://www.math.utah.edu/pub/tex/bib/stoc2010.bib", acknowledgement = ack-nhfb, remark = "45th annual STOC meeting.", }