Valid HTML 4.0! Valid CSS!
%%% -*-BibTeX-*-
%%% ====================================================================
%%%  BibTeX-file{
%%%     author          = "Nelson H. F. Beebe",
%%%     version         = "2.93",
%%%     date            = "13 March 2024",
%%%     time            = "07:32:03 MST",
%%%     filename        = "tods.bib",
%%%     address         = "University of Utah
%%%                        Department of Mathematics, 110 LCB
%%%                        155 S 1400 E RM 233
%%%                        Salt Lake City, UT 84112-0090
%%%                        USA",
%%%     telephone       = "+1 801 581 5254",
%%%     FAX             = "+1 801 581 4148",
%%%     URL             = "https://www.math.utah.edu/~beebe",
%%%     checksum        = "20506 52462 281122 2774199",
%%%     email           = "beebe at math.utah.edu, beebe at acm.org,
%%%                        beebe at computer.org (Internet)",
%%%     codetable       = "ISO/ASCII",
%%%     keywords        = "bibliography, BibTeX, database systems,
%%%                        TODS",
%%%     license         = "public domain",
%%%     supported       = "yes",
%%%     docstring       = "This is a COMPLETE BibTeX bibliography for
%%%                        ACM Transactions on Database Systems (TODS)
%%%                        (CODEN ATDSD3, ISSN 0362-5915 (print),
%%%                        1557-4644 (electronic)), which began
%%%                        publishing in March 1976.
%%%
%%%                        The companion bibliography pods.bib covers
%%%                        the ACM SIGACT-SIGMOD Symposia on
%%%                        Principles of Database Systems, and the
%%%                        companion bibliography vldb.bib covers the
%%%                        International Conferences on Very Large
%%%                        Data Bases.  The companion bibliography
%%%                        sigmod.bib covers the ACM Special Interest
%%%                        Group on Management of Data SIGMOD Record
%%%                        newsletter.
%%%
%%%                        The journal has a World Wide Web site at
%%%
%%%                            http://www.acm.org/tods/
%%%                            http://www.acm.org/pubs/contents/journals/tods/
%%%                            http://portal.acm.org/browse_dl.cfm?idx=J777
%%%
%%%                        At version 2.93, the year coverage looked
%%%                        like this:
%%%
%%%                             1975 (   1)    1992 (  20)    2009 (  25)
%%%                             1976 (  20)    1993 (  20)    2010 (  29)
%%%                             1977 (  23)    1994 (  17)    2011 (  27)
%%%                             1978 (  21)    1995 (  13)    2012 (  32)
%%%                             1979 (  27)    1996 (  14)    2013 (  28)
%%%                             1980 (  25)    1997 (  14)    2014 (  35)
%%%                             1981 (  30)    1998 (  14)    2015 (  21)
%%%                             1982 (  31)    1999 (  13)    2016 (  33)
%%%                             1983 (  30)    2000 (  12)    2017 (  26)
%%%                             1984 (  32)    2001 (  12)    2018 (  18)
%%%                             1985 (  26)    2002 (  11)    2019 (  16)
%%%                             1986 (  25)    2003 (  15)    2020 (  20)
%%%                             1987 (  25)    2004 (  22)    2021 (  16)
%%%                             1988 (  18)    2005 (  29)    2022 (  16)
%%%                             1989 (  23)    2006 (  38)    2023 (  11)
%%%                             1990 (  22)    2007 (  30)    2024 (   4)
%%%                             1991 (  24)    2008 (  31)
%%%                             19xx (   2)
%%%
%%%                             Article:       1078
%%%                             Book:             3
%%%                             InProceedings:    1
%%%                             Proceedings:      5
%%%
%%%                             Total entries: 1087
%%%
%%%                        This bibliography was initially built from
%%%                        searches in the OCLC Content1st database.
%%%                        Additions were then made from all of the
%%%                        bibliographies in the TeX User Group
%%%                        collection, from bibliographies in the
%%%                        author's personal files, from the IEEE
%%%                        INSPEC CD-ROM database (1989--1995), from
%%%                        the Compendex database, from the American
%%%                        Mathematical Society MathSciNet database,
%%%                        and from the computer science bibliography
%%%                        collection on ftp.ira.uka.de in
%%%                        /pub/bibliography to which many people of
%%%                        have contributed.  The snapshot of this
%%%                        collection was taken on 5-May-1994, and it
%%%                        consists of 441 BibTeX files, 2,672,675
%%%                        lines, 205,289 entries, and 6,375
%%%                        <at>String{} abbreviations, occupying
%%%                        94.8MB of disk space.  Missing data in many
%%%                        entries were supplied after consulting
%%%                        original journal issues.
%%%
%%%                        Numerous errors in the sources noted above
%%%                        have been corrected.  Spelling has been
%%%                        verified with the UNIX spell and GNU ispell
%%%                        programs using the exception dictionary
%%%                        stored in the companion file with extension
%%%                        .sok.
%%%
%%%                        BibTeX citation tags are uniformly chosen as
%%%                        name:year:abbrev, where name is the family
%%%                        name of the first author or editor, year is a
%%%                        4-digit number, and abbrev is a 3-letter
%%%                        condensation of important title words.
%%%                        Citation labels were automatically generated
%%%                        by software developed for the BibNet Project.
%%%
%%%                        In this bibliography, entries are sorted in
%%%                        publication order, with the help of
%%%                        ``bibsort -byvolume''.  The bibsort utility
%%%                        is available from ftp.math.utah.edu in
%%%                        /pub/tex/bib.
%%%
%%%                        The checksum field above contains a CRC-16
%%%                        checksum as the first value, followed by the
%%%                        equivalent of the standard UNIX wc (word
%%%                        count) utility output of lines, words, and
%%%                        characters.  This is produced by Robert
%%%                        Solovay's checksum utility.",
%%%  }
%%% ====================================================================
@Preamble{
    "\hyphenation{ }"
}

%%% ====================================================================
%%% Acknowledgement abbreviations:
@String{ack-nhfb = "Nelson H. F. Beebe,
                    University of Utah,
                    Department of Mathematics, 110 LCB,
                    155 S 1400 E RM 233,
                    Salt Lake City, UT 84112-0090, USA,
                    Tel: +1 801 581 5254,
                    FAX: +1 801 581 4148,
                    e-mail: \path|beebe@math.utah.edu|,
                            \path|beebe@acm.org|,
                            \path|beebe@computer.org| (Internet),
                    URL: \path|https://www.math.utah.edu/~beebe/|"}

%%% ====================================================================
%%% Journal abbreviations:
@String{j-BIT                   = "BIT"}

@String{j-CACM                  = "Communications of the ACM"}

@String{j-TODS                  = "ACM Transactions on Database Systems"}

%%% ====================================================================
%%% Publishers and their addresses:
@String{pub-ACM                 = "ACM Press"}

@String{pub-ACM:adr             = "New York, NY 10036, USA"}

@String{pub-IEEE                = "IEEE Computer Society Press"}

@String{pub-IEEE:adr            = "1109 Spring Street, Suite 300, Silver
                                   Spring, MD 20910, USA"}

@String{pub-MORGAN-KAUFMANN     = "Morgan Kaufmann Publishers"}

@String{pub-MORGAN-KAUFMANN:adr = "Los Altos, CA 94022, USA"}

%%% ====================================================================
%%% Bibliography entries:
@Article{Yao:1977:ABA,
  author =       "S. B. Yao",
  title =        "Approximating Block Accesses in Database
                 Organization",
  journal =      j-CACM,
  volume =       "20",
  number =       "4",
  pages =        "260--261",
  month =        apr,
  year =         "1977",
  CODEN =        "CACMA2",
  ISSN =         "0001-0782 (print), 1557-7317 (electronic)",
  bibdate =      "Tue Sep 20 23:14:33 1994",
  bibsource =    "ftp://ftp.ira.uka.de/pub/bibliography/Database/Graefe.bib;
                 ftp://ftp.ira.uka.de/pub/bibliography/Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite{Yao:1977:ABM}.",
  ajournal =     "Commun. ACM",
  fjournal =     "Communications of the ACM",
  journal-URL =  "https://dl.acm.org/loi/tods",
  keywords =     "selectivity estimation I/O cost query optimization
                 CACM",
}

@Article{Hsiao:1976:ATD,
  author =       "David K. Hsiao",
  title =        "{ACM Transactions on Database Systems}: aim and
                 scope",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "1--2",
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p1-hsiao/p1-hsiao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p1-hsiao/",
  abstract =     "Record-keeping and decision-making in industry and
                 government are increasingly based on data stored in
                 computer processable databases. Thus the need for
                 improved computer technology for building, managing,
                 and using these databases is clearly evident. This need
                 is particularly acute in a complex society where the
                 interrelationships among various aspects of the society
                 must be identified and represented. The data which must
                 be used to represent these relationships are growing
                 more complex in nature and becoming greater in size.
                 Furthermore, the increasing on-line use of computer
                 systems and the proliferation and mass introduction of
                 multilevel secondary storage suggests that future
                 computer systems will be primarily oriented toward
                 database management. The large size of future on-line
                 databases will require the computer system to manage
                 local as well as physical resources. The management of
                 logical resources is concerned with the organization,
                 access, update, storage, and sharing of the data and
                 programs in the database. In addition, the sharing of
                 data means that the database system must be capable of
                 providing privacy protection and of controlling access
                 to the users' data. The term {\em data\/} is
                 interpreted broadly to include textual, numeric, and
                 signal data as well as data found in structured
                 records.\par

                 The aim of {\em ACM Transactions on Database Systems\/}
                 (TODS) is to serve as a focal point for an integrated
                 dissemination of database research and development on
                 storage and processor hardware, system software,
                 applications, information science, information
                 analysis, and file management. These areas are
                 particularly relevant to the following ACM Special
                 Interest Groups: Business Data Processing (SIGBDP),
                 Information Retrieval (SIGIR), and Management of Data
                 (SIGMOD). TODS will also embrace parts of the
                 Management/Database Systems and the Information
                 Retrieval and Language Processing sections of {\em
                 Communications of the ACM}.\par

                 High quality papers on all aspects of computer database
                 systems will be published in TODS. The scope of TODS
                 emphasizes data structures; storage organization; data
                 collection and dissemination; search and retrieval
                 strategies; update strategies; access control
                 techniques; data integrity; security and protection;
                 design and implementation of database software;
                 database related languages including data description
                 languages, query languages, and procedural and
                 nonprocedural data manipulation languages; language
                 processing; analysis and classification of data;
                 database utilities; data translation techniques;
                 distributed database problems and techniques; database
                 recovery and restart; database restructuring; adaptive
                 data structures; concurrent access techniques; database
                 computer hardware architecture; performance and
                 evaluation; intelligent front ends; and related
                 subjects such as privacy and economic issues.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4)",
}

@Article{Chen:1976:ERM,
  author =       "Peter Pin-Shan S. Chen",
  title =        "The Entity-Relationship Model: Toward a Unified View
                 of Data",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "9--36",
  month =        mar,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compiler/prog.lang.theory.bib; Database/Graefe.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib; Object/Nierstrasz.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p9-chen/p9-chen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p9-chen/",
  abstract =     "A data model, called the entity-relationship model, is
                 proposed. This model incorporates some of the important
                 semantic information about the real world. A special
                 diagrammatic technique is introduced as a tool for
                 database design. An example of database design and
                 description using the model and the diagrammatic
                 technique is given. Some implications for data
                 integrity, information retrieval, and data manipulation
                 are discussed.\par

                 The entity-relationship model can be used as a basis
                 for unification of different views of data: the network
                 model, the relational model, and the entity set model.
                 Semantic ambiguities in these models are analyzed.
                 Possible ways to derive their views of data from the
                 entity-relationship model are presented.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data Base Task Group; data definition and
                 manipulation; data integrity and consistency; data
                 models; database design; dblit; entity set model;
                 entity-relationship; entity-relationship model; logical
                 view of data; network model; relational model;
                 semantics of data; TODS",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Bayer:1976:EST,
  author =       "R. Bayer and J. K. Metzger",
  title =        "On the Encipherment of Search Trees and Random Access
                 Files",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "37--52",
  month =        mar,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite[p.~508--510]{Kerr:1975:PIC}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p37-bayer/p37-bayer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p37-bayer/",
  abstract =     "The securing of information in indexed, random access
                 files by means of privacy transformations must be
                 considered as a problem distinct from that for
                 sequential files. Not only must processing overhead due
                 to encrypting be considered, but also threats to
                 encipherment arising from updating and the file
                 structure itself must be countered. A general
                 encipherment scheme is proposed for files maintained in
                 a paged structure in secondary storage. This is applied
                 to the encipherment of indexes organized as $B$-trees;
                 a $B$-tree is a particular type of multiway search
                 tree. Threats to the encipherment of $B$-trees,
                 especially relating to updating, are examined, and
                 countermeasures are proposed for each. In addition, the
                 effect of encipherment on file access and update, on
                 paging mechanisms, and on files related to the
                 enciphered index are discussed. Many of the concepts
                 presented may be readily transferred to other forms of
                 multiway index trees and to binary search trees.",
  acknowledgement = ack-nhfb,
  annote =       "Trees versus hashing as his 1974 IFIP paper?",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "",
  subject =      "Software --- Operating Systems --- Security and
                 Protection (D.4.6): {\bf Access controls}; Software ---
                 Operating Systems --- Security and Protection (D.4.6):
                 {\bf Cryptographic controls}",
}

@Article{Lin:1976:DRA,
  author =       "Chyuan Shiun Lin and Diane C. P. Smith and John Miles
                 Smith",
  title =        "The design of a rotating associative memory for
                 relational database applications",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "53--65",
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p53-lin/p53-lin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p53-lin/",
  abstract =     "The design and motivation for a rotating associative
                 relational store (RARES) is described. RARES is
                 designed to enhance the performance of an optimizing
                 relational query interface by supporting important high
                 level optimization techniques. In particular, it can
                 perform tuple selection operations at the storage
                 device and also can provide a mechanism for efficient
                 sorting. Like other designs for rotating associative
                 stores, RARES contains search logic which is attached
                 to the heads of a rotating head-per-track storage
                 device. RARES is distinct from other designs in that it
                 utilizes a novel ``orthogonal'' storage layout. This
                 layout allows a high output rate of selected tuples
                 even when a sort order in the stored relation must be
                 preserved. As in certain other designs, RARES can
                 usually output a tuple as soon as it is found to
                 satisfy the selection criteria. However, relative to
                 these designs, the orthogonal layout allows an order of
                 magnitude reduction in the capacity of storage local to
                 the search logic.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative memory; content addressability; data
                 organization; head-per-track disks; memory systems;
                 relational database; rotating devices; search logic;
                 sorting technique",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Mahmoud:1976:OAR,
  author =       "Samy Mahmoud and J. S. Riordon",
  title =        "Optimal Allocation of Resources in Distributed
                 Information Networks",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "66--78",
  month =        mar,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p66-mahmoud/p66-mahmoud.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p66-mahmoud/",
  abstract =     "The problems of file allocation and capacity
                 assignment in a fixed topology distributed computer
                 network are examined. These two aspects of the design
                 are tightly coupled by means of an average message
                 delay constraint. The objective is to allocate copies
                 of information files to network nodes and capacities to
                 network links so that a minimum cost is achieved
                 subject to network delay and file availability
                 constraints. A model for solving the problem is
                 formulated and the resulting optimization problem is
                 shown to fall into a class of nonlinear integer
                 programming problems. Deterministic techniques for
                 solving this class of problems are computationally
                 cumbersome, even for small size problems. A new
                 heuristic algorithm is developed, which is based on a
                 decomposition technique that greatly reduces the
                 computational complexity of the problem. Numerical
                 results for a variety of network configurations
                 indicate that the heuristic algorithm, while not
                 theoretically convergent, yields practicable low cost
                 solutions with substantial savings in computer
                 processing time and storage requirements. Moreover, it
                 is shown that this algorithm is capable of solving
                 realistic network problems whose solutions using
                 deterministic techniques are computationally
                 intractable.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data files; distributed computed; information
                 networks; link capacities; resource sharing",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2)",
}

@Article{Stemple:1976:DMF,
  author =       "David W. Stemple",
  title =        "A Database Management Facility for Automatic
                 Generation of Database Managers",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "79--94",
  month =        mar,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite[p.~252]{Kerr:1975:PIC}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-1/p79-stemple/p79-stemple.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-1/p79-stemple/",
  abstract =     "A facility is described for the implementation of
                 database management systems having high degrees of {\em
                 horizontal\/} data independence, i.e. independence from
                 chosen logical properties of a database as opposed to
                 {\em vertical\/} independence from storage structures.
                 The facility consists of a high level language for the
                 specification of virtual database managers, a compiler
                 from this language to a pseudomachine language, and an
                 interpreter for the pseudomachine language.\par

                 It is shown how this facility can be used to produce
                 efficient database management systems with any degree
                 of both horizontal and vertical data independence. Two
                 key features of this tool are the compilation of
                 tailored database managers from individual schemas and
                 multiple levels of optional binding.",
  acknowledgement = ack-nhfb,
  annote =       "Describes SLUSH and SLIM, a proposed compiler and
                 interpreter to operate on network schemas with
                 adjustable binding times.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data independence; database management systems",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management (H.2); Software --- Operating Systems ---
                 Systems Programs and Utilities (D.4.9): {\bf make}",
}

@Article{Astrahan:1976:SRR,
  author =       "M. M. Astrahan and M. W. Blasgen and D. D. Chamberlin
                 and K. P. Eswaran and J. N. Gray and P. P. Griffiths
                 and W. F. King and R. A. Lorie and P. R. McJones and J.
                 W. Mehl and G. R. Putzolu and I. L. Traiger and B. W.
                 Wade and V. Watson",
  title =        "{System R}: a Relational Approach to Database
                 Management",
  journal =      j-TODS,
  volume =       "1",
  number =       "2",
  pages =        "97--137",
  month =        jun,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Object/Nierstrasz.bib",
  note =         "Also published in/as: IBM, San Jose, Research Report.
                 No. RJ-1738, Feb. 1976. Reprinted in
                 \cite{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-2/p97-astrahan/p97-astrahan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-2/p97-astrahan/",
  abstract =     "System R is a database management system which
                 provides a high level relational data interface. The
                 systems provides a high level of data independence by
                 isolating the end user as much as possible from
                 underlying storage structures. The system permits
                 definition of a variety of relational views on common
                 underlying data. Data control features are provided,
                 including authorization, integrity assertions,
                 triggered transactions, a logging and recovery
                 subsystem, and facilities for maintaining data
                 consistency in a shared-update environment.\par

                 This paper contains a description of the overall
                 architecture and design of the system. At the present
                 time the system is being implemented and the design
                 evaluated. We emphasize that System R is a vehicle for
                 research in database architecture, and is not planned
                 as a product.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "authorization; data structures; database; dblit; index
                 structures; locking; nonprocedural language; recovery;
                 relational model; TODS relation database IBM San Jose",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf System R}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Relational databases}; Information Systems --- Database
                 Management (H.2)",
}

@Article{Navathe:1976:RLD,
  author =       "Shamkant B. Navathe and James P. Fry",
  title =        "Restructuring for Large Data Bases: Three Levels of
                 Abstraction",
  journal =      j-TODS,
  volume =       "1",
  number =       "2",
  pages =        "138--158",
  month =        mar,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite[p.~174]{Kerr:1975:PIC}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-2/p138-navathe/p138-navathe.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-2/p138-navathe/",
  abstract =     "The development of a powerful restructuring function
                 involves two important components--the unambiguous
                 specification of the restructuring operations and the
                 realization of these operations in a software system.
                 This paper is directed to the first component in the
                 belief that a precise specification will provide a firm
                 foundation for the development of restructuring
                 algorithms and, subsequently, their implementation. The
                 paper completely defines the semantics of the
                 restructuring of tree structured databases.\par

                 The delineation of the restructuring function is
                 accomplished by formulating three different levels of
                 abstraction, with each level of abstraction
                 representing successively more detailed semantics of
                 the function.\par

                 At the first level of abstraction, the schema
                 modification, three types are identified--naming,
                 combining, and relating; these three types are further
                 divided into eight schema operations. The second level
                 of abstraction, the instance operations, constitutes
                 the transformations on the data instances; they are
                 divided into group operations such as replication,
                 factoring, union, etc., and group relation operations
                 such as collapsing, refinement, fusion, etc. The final
                 level, the item value operations, includes the actual
                 item operations, such as copy value, delete value, or
                 create a null value.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data definition; data translation; database; database
                 management systems; logical restructuring",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management ---
                 Heterogeneous Databases (H.2.5): {\bf Data
                 translation**}",
}

@Article{Yao:1976:DDR,
  author =       "S. B. Yao and K. S. Das and T. J. Teorey",
  title =        "A Dynamic Database Reorganization Algorithm",
  journal =      j-TODS,
  volume =       "1",
  number =       "2",
  pages =        "159--174",
  month =        jun,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Purdue Un., TR-168, Nov. 1975.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-2/p159-yao/p159-yao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-2/p159-yao/",
  abstract =     "Reorganization is necessary in some databases for
                 overcoming the performance deterioration caused by
                 updates. The paper presents a dynamic reorganization
                 algorithm which makes the reorganization decision by
                 measuring the database search costs. Previously, the
                 reorganization intervals could only be determined for
                 linear deterioration and known database lifetime. It is
                 shown that the dynamic reorganization algorithm is near
                 optimum for constant reorganization cost and is
                 superior for increasing reorganization cost. In
                 addition, it can be applied to cases of unknown
                 database lifetime and nonlinear performance
                 deterioration. The simplicity, generality, and
                 efficiency appear to make this good heuristic for
                 database reorganization.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database; file organization; information retrieval;
                 reorganization",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3): {\bf Retrieval models}",
}

@Article{Burkhard:1976:HTA,
  author =       "Walter A. Burkhard",
  title =        "Hashing and Trie Algorithms for Partial-Match
                 Retrieval",
  journal =      j-TODS,
  volume =       "1",
  number =       "2",
  pages =        "175--187",
  month =        jun,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; Graphics/siggraph/76.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: UCSD, Appl. Physics and Inf. Sc,
                 CS TR.2, Jun. 1975.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-2/p175-burkhard/p175-burkhard.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-2/p175-burkhard/",
  abstract =     "File designs suitable for retrieval from a file of
                 $k$-letter words when queries may be only partially
                 specified are examined. A new class of partial match
                 file designs (called PMF designs) based upon hash
                 coding and trie search algorithms which provide good
                 worst-case performance is introduced. Upper bounds on
                 the worst-case performance of these designs are given
                 along with examples of files achieving the bound. Other
                 instances of PMF designs are known to have better
                 worst-case performances. The implementation of the file
                 designs with associated retrieval algorithms is
                 considered. The amount of storage required is
                 essentially that required of the records themselves.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; analysis; associative retrieval; hash
                 coding; partial match; retrieval; searching; trie
                 search",
  oldlabel =     "geom-96",
  subject =      "Mathematics of Computing --- Mathematical Software
                 (G.4): {\bf Algorithm design and analysis}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3): {\bf
                 Retrieval models}",
}

@Article{Stonebraker:1976:DII,
  author =       "Michael Stonebraker and Eugene Wong and Peter Kreps
                 and Gerald Held",
  title =        "The Design and Implementation of {INGRES}",
  journal =      j-TODS,
  volume =       "1",
  number =       "3",
  pages =        "189--222",
  month =        sep,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}. Also
                 published in/as: UCB, Elec. Res. Lab, Memo No.
                 ERL-M577, Jan. 1976.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-3/p189-stonebraker/p189-stonebraker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-3/p189-stonebraker/",
  abstract =     "The currently operational (March 1976) version of the
                 INGRES database management system is described. This
                 multiuser system gives a relational view of data,
                 supports two high level nonprocedural data
                 sublanguages, and runs as a collection of user
                 processes on top of the UNIX operating system for
                 Digital Equipment Corporation PDP 11/40, 11/45, and
                 11/70 computers. Emphasis is on the design decisions
                 and tradeoffs related to (1) structuring the system
                 into processes, (2) embedding one command language in a
                 general purpose programming language, (3) the
                 algorithms implemented to process interactions, (4) the
                 access methods implemented, (5) the concurrency and
                 recovery control currently provided, and (6) the data
                 structures used for system catalogs and the role of the
                 database administrator.\par

                 Also discussed are (1) support for integrity
                 constraints (which is only partly operational), (2) the
                 not yet supported features concerning views and
                 protection, and (3) future plans concerning the
                 system.",
  acknowledgement = ack-nhfb,
  annote =       "Describes implementation of INGRES, a non-distributed
                 relational database system. This paper is useful for
                 understanding the distributed INGRES paper.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; data integrity; data organization; data
                 sublanguage; database optimization; nonprocedural
                 language; protection; QUEL EQUEL query modification
                 process structure Halloween problem TODS; query
                 decomposition; query language; relational database",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3); Information Systems --- Database
                 Management --- General (H.2.0): {\bf Security,
                 integrity, and protection**}",
}

@Article{Wong:1976:DSQ,
  author =       "Eugene Wong and Karel Youssefi",
  title =        "Decomposition --- {A} Strategy for Query Processing",
  journal =      j-TODS,
  volume =       "1",
  number =       "3",
  pages =        "223--241",
  month =        sep,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: UCB, Elec. Res. Lab, Memo No.
                 ERL-574, Jan. 1976",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-3/p223-wong/p223-wong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-3/p223-wong/",
  abstract =     "Strategy for processing multivariable queries in the
                 database management system INGRES is considered. The
                 general procedure is to decompose the query into a
                 sequence of one-variable queries by alternating between
                 (a) reduction: breaking off components of the query
                 which are joined to it by a single variable, and (b)
                 tuple substitution: substituting for one of the
                 variables a tuple at a time. Algorithms for reduction
                 and for choosing the variable to be substituted are
                 given. In most cases the latter decision depends on
                 estimation of costs; heuristic procedures for making
                 such estimates are outlined.",
  acknowledgement = ack-nhfb,
  annote =       "INGRES query decomposition by reduction to single
                 variable queries, and tuple substitution --- choosing a
                 variable and for it from all tuples, generating a
                 family of queries in one fewer variable.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "connected query; decomposition; detachment; Ingres
                 TODS; irreducible query; joining (overlapping)
                 variable; query processing; relational database; tuple
                 substitution; variable selection",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}",
}

@Article{Griffiths:1976:AMR,
  author =       "Patricia P. Griffiths and Bradford W. Wade",
  title =        "An Authorization Mechanism for a Relational Database
                 System",
  journal =      j-TODS,
  volume =       "1",
  number =       "3",
  pages =        "242--255",
  month =        sep,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-3/p242-griffiths/p242-griffiths.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-3/p242-griffiths/",
  abstract =     "A multiuser database system must selectively permit
                 users to share data, while retaining the ability to
                 restrict data access. There must be a mechanism to
                 provide protection and security, permitting information
                 to be accessed only by properly authorized users.
                 Further, when tables or restricted views of tables are
                 created and destroyed dynamically, the granting,
                 authentication, and revocation of authorization to use
                 them must also be dynamic. Each of these issues and
                 their solutions in the context of the relational
                 database management system System R are discussed.
                 \par

                 When a database user creates a table, he is fully and
                 solely authorized to perform upon it actions such as
                 read, insert, update, and delete. He may explicitly
                 grant to any other user any or all of his privileges on
                 the table. In addition he may specify that that user is
                 authorized to further grant these privileges to still
                 other users. The result is a directed graph of granted
                 privileges originating from the table creator.\par

                 At some later time a user A may revoke some or all of
                 the privileges which he previously granted to another
                 user B. This action usually revokes the entire subgraph
                 of the grants originating from A's grant to B. It may
                 be, however, that B will still possess the revoked
                 privileges by means of a grant from another user C, and
                 therefore some or all of B's grants should not be
                 revoked. This problem is discussed in detail, and an
                 algorithm for detecting exactly which of B's grants
                 should be revoked is presented.",
  acknowledgement = ack-nhfb,
  annote =       "Defines a dynamic authorization mechanism. A database
                 user can grant or revoke privileges (such as to read,
                 insert, or delete) on a file that he has created.
                 Furthermore, he can authorize others to grant these
                 same privileges. The database management system keeps
                 track of a directed graph, emanating from the creator
                 of granted privileges.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access control; authorization; data dependent
                 authorization; database systems; privacy; protection in
                 databases; revocation of authorization; security",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4); Information Systems --- Database Management
                 --- General (H.2.0): {\bf Security, integrity, and
                 protection**}",
}

@Article{Severance:1976:DFT,
  author =       "Dennis G. Severance and Guy M. Lohman",
  title =        "Differential Files: Their Application to the
                 Maintenance of Large Databases",
  journal =      j-TODS,
  volume =       "1",
  number =       "3",
  pages =        "256--267",
  month =        sep,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-3/p256-severance/p256-severance.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-3/p256-severance/",
  abstract =     "The representation of a collection of data in terms of
                 its differences from some preestablished point of
                 reference is a basic storage compaction technique which
                 finds wide applicability. This paper describes a
                 differential database representation which is shown to
                 be an efficient method for storing large and volatile
                 databases. The technique confines database
                 modifications to a relatively small area of physical
                 storage and as a result offers two significant
                 operational advantages. First, because the ``reference
                 point'' for the database is inherently static, it can
                 be simply and efficiently stored. Second, since all
                 modifications to the database are physically localized,
                 the process of backup and the process of recovery are
                 relatively fast and inexpensive.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backup and recovery; data sharing; database
                 maintenance; differential files",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Shneiderman:1976:BSS,
  author =       "Ben Shneiderman and Victor Goodman",
  title =        "Batched Searching of Sequential and Tree Structured
                 Files",
  journal =      j-TODS,
  volume =       "1",
  number =       "3",
  pages =        "268--275",
  month =        sep,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See comments in \cite{Piwowarski:1985:CBS}. Also
                 published in/as: Indiana Un., CSD Tech. Ref. 0132.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-3/p268-shneiderman/p268-shneiderman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-3/p268-shneiderman/",
  abstract =     "The technique of batching searches has been ignored in
                 the context of disk based online data retrieval
                 systems. This paper suggests that batching be
                 reconsidered for such systems since the potential
                 reduction in processor demand may actually reduce
                 response time. An analysis with sample numerical
                 results and algorithms is presented.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval (H.3)",
}

@Article{Bernstein:1976:STN,
  author =       "Philip A. Bernstein",
  title =        "Synthesizing Third Normal Form Relations from
                 Functional Dependencies",
  journal =      j-TODS,
  volume =       "1",
  number =       "4",
  pages =        "277--298",
  month =        dec,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-4/p277-bernstein/p277-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-4/p277-bernstein/",
  abstract =     "It has been proposed that the description of a
                 relational database can be formulated as a set of
                 functional relationships among database attributes.
                 These functional relationships can then be used to
                 synthesize algorithmically a relational scheme. It is
                 the purpose of this paper to present an effective
                 procedure for performing such a synthesis. The schema
                 that results from this procedure is proved to be in
                 Codd's third normal form and to contain the fewest
                 possible number of relations. Problems with earlier
                 attempts to construct such a procedure are also
                 discussed.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database schema; functional dependency; relational
                 model; semantics of data; third normal form",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Normal forms}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Schema and subschema}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Data models}",
}

@Article{Liu:1976:APS,
  author =       "Jane W. S. Liu",
  title =        "Algorithms for parsing search queries in systems with
                 inverted file organization",
  journal =      j-TODS,
  volume =       "1",
  number =       "4",
  pages =        "299--316",
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-4/p299-liu/p299-liu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-4/p299-liu/",
  abstract =     "In an inverted file system a query is in the form of a
                 Boolean expression of index terms. In response to a
                 query the system accesses the inverted lists
                 corresponding to the index terms, merges them, and
                 selects from the merged list those records that satisfy
                 the search logic. Considered in this paper is the
                 problem of determining a Boolean expression which leads
                 to the minimum total merge time among all Boolean
                 expressions that are equivalent to the expression given
                 in the query. This problem is the same as finding an
                 optimal merge tree among all trees that realize the
                 truth function determined by the Boolean expression in
                 the query. Several algorithms are described which
                 generate optimal merge trees when the sizes of overlaps
                 between different lists are small compared with the
                 length of the lists.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "inverted file systems; merge algorithms; parsing
                 Boolean queries",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Mathematics of Computing ---
                 Mathematical Software (G.4): {\bf Algorithm design and
                 analysis}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Query processing}",
}

@Article{Sherman:1976:PDM,
  author =       "Stephen W. Sherman and Richard S. Brice",
  title =        "Performance of a Database Manager in a Virtual Memory
                 System",
  journal =      j-TODS,
  volume =       "1",
  number =       "4",
  pages =        "317--343",
  month =        dec,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-4/p317-sherman/p317-sherman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-4/p317-sherman/",
  abstract =     "Buffer space is created and managed in database
                 systems in order to reduce accesses to the I/O devices
                 for database information. In systems using virtual
                 memory any increase in the buffer space may be
                 accompanied by an increase in paging. The effects of
                 these factors on system performance are quantified
                 where system performance is a function of page faults
                 and database accesses to I/O devices. This phenomenon
                 is examined through the analysis of empirical data
                 gathered in a multifactor experiment. The factors
                 considered are memory size, size of buffer space,
                 memory replacement algorithm, and buffer management
                 algorithm. The improvement of system performance
                 through an increase in the size of the buffer space is
                 demonstrated. It is also shown that for certain values
                 of the other factors an increase in the size of the
                 buffer space can cause performance to deteriorate.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer manager; Buffer operating system support TODS;
                 database management; double paging; page faults; page
                 replacement algorithm; performance; virtual buffer;
                 virtual memory",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Database Manager}; Mathematics of
                 Computing --- Mathematical Software (G.4): {\bf
                 Algorithm design and analysis}; Computer Systems
                 Organization --- Performance of Systems (C.4)",
}

@Article{Donovan:1976:DSA,
  author =       "John J. Donovan",
  title =        "Database System Approach to Management Decision
                 Support",
  journal =      j-TODS,
  volume =       "1",
  number =       "4",
  pages =        "344--369",
  month =        dec,
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-4/p344-donovan/p344-donovan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-4/p344-donovan/",
  abstract =     "Traditional intuitive methods of decision-making are
                 no longer adequate to deal with the complex problems
                 faced by the modern policymaker. Thus systems must be
                 developed to provide the information and analysis
                 necessary for the decisions which must be made. These
                 systems are called decision support systems. Although
                 database systems provide a key ingredient to decision
                 support systems, the problems now facing the
                 policymaker are different from those problems to which
                 database systems have been applied in the past. The
                 problems are usually not known in advance, they are
                 constantly changing, and answers are needed quickly.
                 Hence additional technologies, methodologies, and
                 approaches must expand the traditional areas of
                 database and operating systems research (as well as
                 other software and hardware research) in order for them
                 to become truly effective in supporting policymakers.
                 \par

                 This paper describes recent work in this area and
                 indicates where future work is needed. Specifically the
                 paper discusses: (1) why there exists a vital need for
                 decision support systems; (2) examples from work in the
                 field of energy which make explicit the characteristics
                 which distinguish these decision support systems from
                 traditional operational and managerial systems; (3) how
                 an awareness of decision support systems has evolved,
                 including a brief review of work done by others and a
                 statement of the computational needs of decision
                 support systems which are consistent with contemporary
                 technology; (4) an approach which has been made to meet
                 many of these computational needs through the
                 development and implementation of a computational
                 facility, the Generalized Management Information System
                 (GMIS); and (5) the application of this computational
                 facility to a complex and important energy problem
                 facing New England in a typical study within the New
                 England Energy Management Information System (NEEMIS)
                 Project.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; decision support systems; management
                 applications; modeling; networking; relational; virtual
                 machines",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management --- Systems
                 (H.2.4)",
}

@Article{McGee:1976:UCD,
  author =       "William C. McGee",
  title =        "On user criteria for data model evaluation",
  journal =      j-TODS,
  volume =       "1",
  number =       "4",
  pages =        "370--387",
  year =         "1976",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1976-1-4/p370-mcgee/p370-mcgee.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1976-1-4/p370-mcgee/",
  abstract =     "The emergence of a database technology in recent years
                 has focused interest on the subject of data models. A
                 data model is the class of logical data structures
                 which a computer system or language makes available to
                 the user for the purpose of formulating data processing
                 applications. The diversity of computer systems and
                 languages has resulted in a corresponding diversity of
                 data models, and has created a problem for the user in
                 selecting a data model which is in some sense
                 appropriate to a given application. An evaluation
                 procedure is needed which will allow the user to
                 evaluate alternative models in the context of a
                 specific set of applications. This paper takes a first
                 step toward such a procedure by identifying the
                 attributes of a data model which can be used as
                 criteria for evaluating the model. Two kinds of
                 criteria are presented: use criteria, which measure the
                 usability of the model; and implementation criteria,
                 which measure the implementability of the model and the
                 efficiency of the resulting implementation. The use of
                 the criteria is illustrated by applying them to three
                 specific models: an $n$-ary relational model, a
                 hierarchic model, and a network model.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data model; data model evaluation; data model
                 selection; hierarchic model; network model; relational
                 model",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}",
}

@Article{Kam:1977:MSD,
  author =       "John B. Kam and Jeffrey D. Ullman",
  title =        "A Model of Statistical Databases and Their Security",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "1--10",
  month =        mar,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-1/p1-kam/p1-kam.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-1/p1-kam/",
  abstract =     "Considered here, for a particular model of databases
                 in which only information about relatively large sets
                 of records can be obtained, is the question of whether
                 one can from statistical information obtain information
                 about individuals. Under the assumption that the data
                 in the database is taken from arbitrary integers, it is
                 shown that essentially nothing can be inferred. It is
                 also shown that when the values are known to be
                 imprecise in some fixed range, one can often deduce the
                 values of individual records.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromisability; data security; linear independence;
                 statistical database; vector spece",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- General (H.2.0): {\bf Security, integrity, and
                 protection**}",
}

@Article{Bayer:1977:PBT,
  author =       "Rudolf Bayer and Karl Unterauer",
  title =        "Prefix {B}-trees",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "11--26",
  month =        mar,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: IBM Yorktwon, Technical Report
                 RJ1796, Jun. 1976.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-1/p11-bayer/p11-bayer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-1/p11-bayer/",
  abstract =     "Two modifications of $B$-trees are described, simple
                 prefix $B$-trees and prefix $B$-trees. Both store only
                 parts of keys, namely prefixes, in the index part of a
                 $B$ *-tree. In simple prefix $B$-trees those prefixes
                 are selected carefully to minimize their length. In
                 prefix $B$-trees the prefixes need not be fully stored,
                 but are reconstructed as the tree is searched. Prefix
                 $B$-trees are designed to combine some of the
                 advantages of $B$-trees, digital search trees, and key
                 compression techniques while reducing the processing
                 overhead of compression techniques.",
  acknowledgement = ack-nhfb,
  annote =       "Index Btree structures can easily be compressed.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "truncation compression TODS",
  subject =      "Data --- Data Structures (E.1): {\bf Trees}",
}

@Article{Schkolnick:1977:CAH,
  author =       "Mario Schkolnick",
  title =        "A Clustering Algorithm for Hierarchical Structures",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "27--44",
  month =        may,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 09:36:45 1996",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  annote =       "Optimal file partitioning, applied to IMS.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yao:1977:ABM,
  author =       "S. B. Yao",
  title =        "An Attribute Based Model for Database Access Cost
                 Analysis",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "45--67",
  month =        mar,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite{Yao:1977:ABA}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-1/p45-yao/p45-yao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-1/p45-yao/",
  abstract =     "A generalized model for physical database
                 organizations is presented. Existing database
                 organizations are shown to fit easily into the model as
                 special cases. Generalized access algorithms and cost
                 equations associated with the model are developed and
                 analyzed. The model provides a general design framework
                 in which the distinguishing properties of database
                 organizations are made explicit and their performances
                 can be compared.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "B-tree; database model; database organization;
                 database performance; estimation approximation TODS;
                 evaluation; index organization; index sequential;
                 inverted file; multilist",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1); Information Systems ---
                 Information Storage and Retrieval --- Content Analysis
                 and Indexing (H.3.1); Data --- Data Structures (E.1):
                 {\bf Trees}",
}

@Article{Anderson:1977:MCS,
  author =       "Henry D. Anderson and P. Bruce Berra",
  title =        "Minimum Cost Selection of Secondary Indexes for
                 Formatted Files",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "68--90",
  month =        mar,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-1/p68-anderson/p68-anderson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-1/p68-anderson/",
  abstract =     "Secondary indexes are often used in database
                 management systems for secondary key retrieval.
                 Although their use can improve retrieval time
                 significantly, the cost of index maintenance and
                 storage increases the overhead of the file processing
                 application. The optimal set of indexed secondary keys
                 for a particular application depends on a number of
                 application dependent factors. In this paper a cost
                 function is developed for the evaluation of candidate
                 indexing choices and applied to the optimization of
                 index selection. Factors accounted for include file
                 size, the relative rates of retrieval and maintenance
                 and the distribution of retrieval and maintenance over
                 the candidate keys, index structure, and system
                 charging rates. Among the results demonstrated are the
                 increased effectiveness of secondary indexes for large
                 files, the effect of the relative rates of retrieval
                 and maintenance, the greater cost of allowing for
                 arbitrarily formulated queries, and the impact on cost
                 of the use of different index structures.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access methods; access path; Boolean query; cost
                 function; data management; database; file design; file
                 organization; inverted file; inverted index;
                 maintenance; optimization; retrieval; secondary index;
                 secondary key; secondary key access",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Content Analysis and Indexing (H.3.1):
                 {\bf Indexing methods}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Storage (H.3.2): {\bf File organization}; Information
                 Systems --- Database Management --- Physical Design
                 (H.2.2): {\bf Access methods}; Information Systems ---
                 Database Management (H.2)",
}

@Article{Lorie:1977:PIL,
  author =       "Raymond A. Lorie",
  title =        "Physical Integrity in a Large Segmented Database",
  journal =      j-TODS,
  volume =       "2",
  number =       "1",
  pages =        "91--104",
  month =        mar,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-1/p91-lorie/p91-lorie.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-1/p91-lorie/",
  abstract =     "A database system can generally be divided into three
                 major components. One component supports the logical
                 database as seen by the user. Another component maps
                 the information into physical records. The third
                 component, called the storage component, is responsible
                 for mapping these records onto auxiliary storage
                 (generally disks) and controlling their transfer to and
                 from main storage.\par

                 This paper is primarily concerned with the
                 implementation of a storage component. It considers a
                 simple and classical interface to the storage
                 component: Seen at this level the database is a
                 collection of segments. Each segment is a linear
                 address space.\par

                 A recovery scheme is first proposed for system failure
                 (hardware or software error which causes the contents
                 of main storage to be lost). It is based on maintaining
                 a dual mapping between pages and their location on
                 disk. One mapping represents the current state of a
                 segment being modified; the other represents a previous
                 backup state. At any time the backup state can be
                 replaced by the current state without any data merging.
                 Procedures for segment modification, save, and restore
                 are analyzed. Another section proposes a facility for
                 protection against damage to the auxiliary storage
                 itself. It is shown how such protection can be obtained
                 by copying on a tape (checkpoint) only those pages that
                 have been modified since the last checkpoint.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "checkpoint-restart; database; recovery; storage
                 management",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0): {\bf Security, integrity, and
                 protection**}; Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2);
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart}",
}

@Article{Smith:1977:DAA,
  author =       "John Miles Smith and Diane C. P. Smith",
  title =        "Database abstractions: Aggregation and
                 Generalization",
  journal =      j-TODS,
  volume =       "2",
  number =       "2",
  pages =        "105--133",
  month =        jun,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Object/Nierstrasz.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-2/p105-smith/p105-smith.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-2/p105-smith/",
  abstract =     "Two kinds of abstraction that are fundamentally
                 important in database design and usage are defined.
                 Aggregation is an abstraction which turns a
                 relationship between objects into an aggregate object.
                 Generalization is an abstraction which turns a class of
                 objects into a generic object. It is suggested that all
                 objects (individual, aggregate, generic) should be
                 given uniform treatment in models of the real world. A
                 new data type, called generic, is developed as a
                 primitive for defining such models. Models defined with
                 this primitive are structured as a set of aggregation
                 hierarchies intersecting with a set of generalization
                 hierarchies. Abstract objects occur at the points of
                 intersection. This high level structure provides a
                 discipline for the organization of relational
                 databases. In particular this discipline allows: (i) an
                 important class of views to be integrated and
                 maintained; (ii) stability of data and programs under
                 certain evolutionary changes; (iii) easier
                 understanding of complex models and more natural {\em
                 query formulation;\/} (iv) {\em a more systematic
                 approach to database design;\/} (v) {\em more
                 optimization\/} to be performed at lower implementation
                 levels. The generic type is formalized by a set of
                 invariant properties. These properties should be
                 satisfied by all relations in a database if
                 abstractions are to be preserved. A triggering
                 mechanism for automatically maintaining these
                 invariants during update operations is proposed. A
                 simple mapping of aggregation/generalization
                 hierarchies onto owner-coupled set structures is
                 given.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregation; data abstraction; data model; data type;
                 database design; dblit data abstraction;
                 generalization; integrity constraints; knowledge
                 representation; relational database",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Relational databases}; Software --- Software
                 Engineering --- Software Architectures (D.2.11): {\bf
                 Data abstraction}",
}

@Article{Shu:1977:EDE,
  author =       "N. C. Shu and B. C. Housel and R. W. Taylor and S. P.
                 Ghosh and V. Y. Lum",
  title =        "{EXPRESS}: a data {EXtraction, Processing, and
                 Restructuring System}",
  journal =      j-TODS,
  volume =       "2",
  number =       "2",
  pages =        "134--174",
  month =        jun,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-2/p134-shu/p134-shu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-2/p134-shu/",
  abstract =     "EXPRESS is an experimental prototype data translation
                 system which can access a wide variety of data and
                 restructure it for new uses. The system is driven by
                 two very high level nonprocedural languages: DEFINE for
                 data description and CONVERT for data restructuring.
                 Program generation and cooperating process techniques
                 are used to achieve efficient operation.\par

                 This paper describes the design and implementation of
                 EXPRESS. DEFINE and CONVERT are summarized and the
                 implementation architecture presented.\par

                 The DEFINE description is compiled into a customized
                 PL/1 program for accessing source data. The
                 restructuring specified in CONVERT is compiled into a
                 set of customized PL/1 procedures to derive multiple
                 target files from multiple input files. Job steps and
                 job control statements are generated automatically.
                 During execution, the generated procedures run under
                 control of a process supervisor, which coordinates
                 buffer management and handles file allocation,
                 deallocation, and all input/output requests.\par

                 The architecture of EXPRESS allows efficiency in
                 execution by avoiding unnecessary secondary storage
                 references while at the same time allowing the
                 individual procedures to be independent of each other.
                 Its modular structure permits the system to be extended
                 or transferred to another environment easily.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data conversion; data description languages; data
                 manipulation languages; data restructuring; data
                 translation; file conversion; program generation; very
                 high level languages",
  subject =      "Information Systems --- Database Management ---
                 Heterogeneous Databases (H.2.5): {\bf Data
                 translation**}; Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2):
                 {\bf File organization}; Information Systems ---
                 Database Management --- Languages (H.2.3)",
}

@Article{Ozkarahan:1977:PER,
  author =       "E. A. Ozkarahan and S. A. Schuster and K. C. Sevcik",
  title =        "Performance Evaluation of a Relational Associative
                 Processor",
  journal =      j-TODS,
  volume =       "2",
  number =       "2",
  pages =        "175--195",
  month =        jun,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-2/p175-ozkarahan/p175-ozkarahan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-2/p175-ozkarahan/",
  abstract =     "An associative processor called RAP has been designed
                 to provide hardware support for the use and
                 manipulation of databases. RAP is particularly suited
                 for supporting relational databases. In this paper, the
                 relational operations provided by the RAP hardware are
                 described, and a representative approach to providing
                 the same relational operations with conventional
                 software and hardware is devised. Analytic models are
                 constructed for RAP and the conventional system. The
                 execution times of several of the operations are shown
                 to be vastly improved with RAP for large relations.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative processors; database machines; performance
                 evaluation; RAP hardware support database machine TODS;
                 relational databases",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}; Hardware
                 --- Control Structures and Microprogramming --- Control
                 Structure Performance Analysis and Design Aids
                 (B.1.2)",
}

@Article{Brice:1977:EPD,
  author =       "Richard S. Brice and Stephen W. Sherman",
  title =        "An Extension on the Performance of a Database Manager
                 in a Virtual Memory System Using Partially Locked
                 Virtual Buffers",
  journal =      j-TODS,
  volume =       "2",
  number =       "2",
  pages =        "196--207",
  month =        jun,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-2/p196-brice/p196-brice.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-2/p196-brice/",
  abstract =     "Buffer pools are created and managed in database
                 systems in order to reduce the total number of accesses
                 to the I/O devices. In systems using virtual memory,
                 any reduction in I/O accesses may be accompanied by an
                 increase in paging. The effects of these factors on
                 system performance are quantified, where system
                 performance is a function of page faults and database
                 accesses to the I/O devices. A previous study of this
                 phenomenon is extended through the analysis of
                 empirical data gathered in a multifactor experiment. In
                 this study memory is partitioned between the program
                 and the buffer so that the impact of the controlled
                 factors can be more effectively evaluated. It is
                 possible to improve system performance through the use
                 of different paging algorithms in the program partition
                 and the buffer partition. Also, the effects on system
                 performance as the virtual buffer size is increased
                 beyond the real memory allocated to the buffer
                 partition are investigated.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer manager; database management; double paging;
                 locked buffer; page faults; page replacement algorithm;
                 performance; pinning fixing TODS; virtual buffer;
                 virtual memory",
  subject =      "Hardware --- Control Structures and Microprogramming
                 --- Control Structure Performance Analysis and Design
                 Aids (B.1.2); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Database
                 Manager}",
}

@Article{Lohman:1977:OPB,
  author =       "Guy M. Lohman and John A. Muckstadt",
  title =        "Optimal Policy for Batch Operations: Backup,
                 Checkpointing, Reorganization, and Updating",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "209--222",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p209-lohman/p209-lohman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p209-lohman/",
  abstract =     "Many database maintenance operations are performed
                 periodically in batches, even in realtime systems. The
                 purpose of this paper is to present a general model for
                 determining the optimal frequency of these batch
                 operations. Specifically, optimal backup,
                 checkpointing, batch updating, and reorganization
                 policies are derived. The approach used exploits
                 inventory parallels by seeking the optimal number of
                 items--rather than a time interval--to trigger a batch.
                 The Renewal Reward Theorem is used to find the average
                 long run costs for backup, recovery, and item storage,
                 per unit time, which is then minimized to find the
                 optimal backup policy. This approach permits far less
                 restrictive assumptions about the update arrival
                 process than did previous models, as well as inclusion
                 of storage costs for the updates. The optimal
                 checkpointing, batch updating, and reorganization
                 policies are shown to be special cases of this optimal
                 backup policy. The derivation of previous results as
                 special cases of this model, and an example,
                 demonstrate the generality of the methodology
                 developed.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backup frequency; batch operations; batch update;
                 checkpoint interval; data base systems; database
                 maintenance; file reorganization; inventory theory;
                 real-time systems; renewal theory",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0)",
}

@Article{Wong:1977:IHT,
  author =       "Kai C. Wong and Murray Edelberg",
  title =        "Interval Hierarchies and Their Application to
                 Predicate Files",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "223--232",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p223-wong/p223-wong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p223-wong/",
  abstract =     "Predicates are used extensively in modern database
                 systems for purposes ranging from user specification of
                 associative accesses to data, to user-invisible system
                 control functions such as concurrency control and data
                 distribution. Collections of predicates, or predicate
                 files, must be maintained and accessed efficiently. A
                 dynamic index is described, called an interval
                 hierarchy, which supports several important retrieval
                 operations on files of simple conjunctive predicates.
                 Search and maintenance algorithms for interval
                 hierarchies are given. For a file of n predicates,
                 typical of the kind expected in practice, these
                 algorithms require time equal to $ O(\log n) $.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; data base systems; database
                 system; distributed data; index; interval; predicate
                 file",
  subject =      "Software --- Operating Systems --- Storage Management
                 (D.4.2): {\bf Storage hierarchies}; Information Systems
                 --- Information Storage and Retrieval --- Information
                 Storage (H.3.2): {\bf File organization}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Distributed databases}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Concurrency}; Information Systems --- Information
                 Storage and Retrieval --- Content Analysis and Indexing
                 (H.3.1): {\bf Indexing methods}",
}

@Article{Ries:1977:ELG,
  author =       "Daniel R. Ries and Michael Stonebraker",
  title =        "Effects of Locking Granularity in a Database
                 Management System",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "233--246",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p233-ries/p233-ries.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p233-ries/",
  abstract =     "Many database systems guarantee some form of integrity
                 control upon multiple concurrent updates by some form
                 of locking. Some ``granule'' of the database is chosen
                 as the unit which is individually locked, and a lock
                 management algorithm is used to ensure integrity. Using
                 a simulation model, this paper explores the desired
                 size of a granule. Under a wide variety of seemingly
                 realistic conditions, surprisingly coarse granularity
                 is called for. The paper concludes with some
                 implications of these results concerning the viability
                 of so-called predicate locking.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; consistency; data base systems; database
                 management; locking granularity; multiple updates;
                 predicate locks",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management (H.2); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Concurrency}",
}

@Article{Schmidt:1977:SHL,
  author =       "Joachim W. Schmidt",
  title =        "Some High Level Language Constructs for Data of Type
                 Relation",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "247--261",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p247-schmidt/p247-schmidt.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p247-schmidt/",
  abstract =     "For the extension of high level languages by data
                 types of mode relation, three language constructs are
                 proposed and discussed: a repetition statement
                 controlled by relations, predicates as a generalization
                 of Boolean expressions, and a constructor for relations
                 using predicates. The language constructs are developed
                 step by step starting with a set of elementary
                 operations on relations. They are designed to fit into
                 PASCAL without introducing too many additional
                 concepts.",
  acknowledgement = ack-nhfb,
  annote =       "PASCAL/R",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data type; database;
                 high level language; language extension; nonprocedural
                 language; relational calculus; relational model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3)",
}

@Article{Fagin:1977:MVD,
  author =       "Ronald Fagin",
  title =        "Multi-Valued Dependencies and a New Normal Form for
                 Relational Databases",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "262--278",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p262-fagin/p262-fagin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p262-fagin/",
  abstract =     "A new type of dependency, which includes the
                 well-known functional dependencies as a special case,
                 is defined for relational databases. By using this
                 concept, a new (``fourth'') normal form for relation
                 schemata is defined. This fourth normal form is
                 strictly stronger than Codd's ``improved third normal
                 form'' (or ``Boyce-Codd normal form''). It is shown
                 that every relation schema can be decomposed into a
                 family of relation schemata in fourth normal form
                 without loss of information (that is, the original
                 relation can be obtained from the new relations by
                 taking joins).",
  acknowledgement = ack-nhfb,
  annote =       "Multivalued dependency is defined for relational
                 databases, a new (``fourth'') normal form is strictly
                 stronger than Codd's.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "3NF; 4NF; Boyce-Codd normal form; data base systems;
                 database design; decomposition; fourth normal form;
                 functional dependency; multivalued dependency;
                 normalization; relational database; third normal form",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Normal forms}",
}

@Article{March:1977:DER,
  author =       "Salvatore T. March and Dennis G. Severance",
  title =        "The Determination of Efficient Record Segmentations
                 and Blocking Factors for Shared Data Files",
  journal =      j-TODS,
  volume =       "2",
  number =       "3",
  pages =        "279--296",
  month =        sep,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-3/p279-march/p279-march.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-3/p279-march/",
  abstract =     "It is generally believed that 80 percent of all
                 retrieval from a commercial database is directed at
                 only 20 percent of the stored data items. By
                 partitioning data items into primary and secondary
                 record segments, storing them in physically separate
                 files, and judiciously allocating available buffer
                 space to the two files, it is possible to significantly
                 reduce the average cost of information retrieval from a
                 shared database. An analytic model, based upon
                 knowledge of data item lengths, data access costs, and
                 user retrieval patterns, is developed to assist an
                 analyst with this assignment problem. A computationally
                 tractable design algorithm is presented and results of
                 its application are described.",
  acknowledgement = ack-nhfb,
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bicriterion mathematical programs; branch and bound;
                 buffer allocation; data base systems; data management;
                 information science --- information retrieval; network
                 flows; record design; record segmentation",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Ozkarahan:1977:AAF,
  author =       "E. A. Ozkarahan and K. C. Sevcik",
  title =        "Analysis of Architectural Features for Enhancing the
                 Performance of a Database Machine",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "297--316",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p297-ozkarahan/p297-ozkarahan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p297-ozkarahan/",
  abstract =     "RAP (Relational Associative Processor) is a
                 ``back-end'' database processor that is intended to
                 take over much of the effort of database management in
                 a computer system. In order to enhance RAP's
                 performance its design includes mechanisms for
                 permitting features analogous to multiprogramming and
                 virtual memory as in general purpose computer systems.
                 It is the purpose of this paper to present the detailed
                 design of these mechanisms, along with some analysis
                 that supports their value. Specifically, (1) the
                 response time provided by RAP under several scheduling
                 disciplines involving priority by class is analyzed,
                 (2) the cost effectiveness of the additional hardware
                 in RAP necessary to support multiprogramming is
                 assessed, and (3) a detailed design of the RAP virtual
                 memory system and its monitor is presented.",
  acknowledgement = ack-nhfb,
  annote =       "RAP (Relational Associative Processor) is a ``back-end
                 database processor''; its design includes mechanisms
                 for multiprogramming and virtual memory.",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative processors; computer architecture;
                 computer architecture, hardware support TODS; data base
                 systems; database machines; database management",
  subject =      "Information Systems --- Database Management (H.2)",
}

@Article{Rissanen:1977:ICR,
  author =       "Jorma Rissanen",
  title =        "Independent Components of Relations",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "317--325",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p317-rissanen/p317-rissanen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p317-rissanen/",
  abstract =     "In a multiattribute relation or, equivalently, a
                 multicolumn table a certain collection of the
                 projections can be shown to be independent in much the
                 same way as the factors in a Cartesian product or
                 orthogonal components of a vector. A precise notion of
                 independence for relations is defined and studied. The
                 main result states that the operator which reconstructs
                 the original relation from its independent components
                 is the natural join, and that independent components
                 split the full family of functional dependencies into
                 corresponding component families. These give an
                 easy-to-check criterion for independence.",
  acknowledgement = ack-nhfb,
  annote =       "In a multi-attribute relation a certain collection of
                 projections can be shown to be independent. The
                 operator which reconstructs the original relation is
                 the natural join. Independent components split the full
                 family of functional dependencies into corresponding
                 component families.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database; functional dependencies;
                 relations",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Bonczek:1977:TGB,
  author =       "Robert H. Bonczek and James I. Cash and Andrew B.
                 Whinston",
  title =        "A Transformational Grammar-Based Query Processor for
                 Access Control in a Planning System",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "326--338",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p326-bonczek/p326-bonczek.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p326-bonczek/",
  abstract =     "Providing computer facilities and data availability to
                 larger numbers of users generates increased system
                 vulnerability which is partially offset by software
                 security systems. Much too often these systems are
                 presented as ad hoc additions to the basic data
                 management system. One very important constituent of
                 software security systems is the access control
                 mechanism which may be the last resource available to
                 prohibit unauthorized data retrieval. This paper
                 presents a specification for an access control
                 mechanism. The mechanism is specified in a context for
                 use with the GPLAN decision support system by a
                 theoretical description consistent with the formal
                 definition of GPLAN's query language. Incorporation of
                 the mechanism into the language guarantees it will not
                 be an ad hoc addition. Furthermore, it provides a
                 facile introduction of data security dictates into the
                 language processor.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access control; data processing; data security;
                 database; decision support system; planning system",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- General (H.2.0):
                 {\bf Security, integrity, and protection**};
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods}",
}

@Article{Lang:1977:DBP,
  author =       "Tom{\'a}s Lang and Christopher Wood and Eduardo B.
                 Fern{\'a}ndez",
  title =        "Database Buffer Paging in Virtual Storage Systems",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "339--351",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p339-lang/p339-lang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p339-lang/",
  abstract =     "Three models, corresponding to different sets of
                 assumptions, are analyzed to study the behavior of a
                 database buffer in a paging environment. The models
                 correspond to practical situations and vary in their
                 search strategies and replacement algorithms. The
                 variation of I/O cost with respect to buffer size is
                 determined for the three models. The analysis is valid
                 for arbitrary database and buffer sizes, and the I/O
                 cost is obtained in terms of the miss ratio, the buffer
                 size, the number of main memory pages available for the
                 buffer, and the relative buffer and database access
                 costs.",
  acknowledgement = ack-nhfb,
  annote =       "The variation of I/O cost with respect to buffer size
                 is determined for three models: the IMS/360 database
                 buffer, with LRU memory replacement, and a prefix table
                 in main memory indicating which database pages are in
                 the VSAM buffer.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer management; computer systems performance; data
                 base systems; database performance; page replacement
                 algorithm; virtual memory",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0); Information Systems --- Database
                 Management --- Systems (H.2.4)",
}

@Article{Thomas:1977:VAP,
  author =       "D. A. Thomas and B. Pagurek and R. J. Buhr",
  title =        "Validation Algorithms for Pointer Values in {DBTG}
                 Databases",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "352--369",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p352-thomas/p352-thomas.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p352-thomas/",
  abstract =     "This paper develops algorithms for verifying pointer
                 values in DBTG (Data Base Task Group) type databases.
                 To validate pointer implemented access paths and set
                 structures, two algorithms are developed. The first
                 procedure exploits the ``typed pointer'' concept
                 employed in modern programming languages to diagnose
                 abnormalities in directories and set instances. The
                 second algorithm completes pointer validation by
                 examining set instances to ensure that each DBTG set
                 has a unique owner. Sequential processing is used by
                 both algorithms, allowing a straightforward
                 implementation which is efficient in both time and
                 space. As presented, the algorithms are independent of
                 implementation schema and physical structure.",
  acknowledgement = ack-nhfb,
  annote =       "Type Checking algorithm detects and locates errors in
                 the pointers which are used to represent chained and
                 pointer array implemented sets. In addition to invalid
                 set pointers, the algorithm has been extended to check
                 index sequential and inverted access directories
                 provided by EDMS.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database integrity; database
                 utilities; type checking; validation",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management --- General
                 (H.2.0): {\bf Security, integrity, and protection**}",
}

@Article{Claybrook:1977:FDM,
  author =       "Billy G. Claybrook",
  title =        "A Facility for Defining and Manipulating Generalized
                 Data Structures",
  journal =      j-TODS,
  volume =       "2",
  number =       "4",
  pages =        "370--406",
  month =        dec,
  year =         "1977",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1977-2-4/p370-claybrook/p370-claybrook.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1977-2-4/p370-claybrook/",
  abstract =     "A data structure definition facility (DSDF) is
                 described that provides definitions for several
                 primitive data types, homogeneous and heterogeneous
                 arrays, cells, stacks, queues, trees, and general
                 lists. Each nonprimitive data structure consists of two
                 separate entities--a head and a body. The head contains
                 the entry point(s) to the body of the structure; by
                 treating the head like a cell, the DSDF operations are
                 capable of creating and manipulating very general data
                 structures. A template structure is described that
                 permits data structures to share templates.\par

                 The primary objectives of the DSDF are: (1) to develop
                 a definition facility that permits the programmer to
                 explicitly define and manipulate generalized data
                 structures in a consistent manner, (2) to detect
                 mistakes and prevent the programmer from creating
                 (either inadvertently or intentionally) undesirable (or
                 illegal) data structures, (3) to provide a syntactic
                 construction mechanism that separates the
                 implementation of a data structure from its use in the
                 program in which it is defined, and (4) to facilitate
                 the development of reliable software.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data definition languages; data processing; data
                 structure definition facility; data structures;
                 database management",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management ---
                 Languages (H.2.3)",
}

@Article{Minker:1978:SSS,
  author =       "Jack Minker",
  title =        "Search Strategy and Selection Function for an
                 Inferential Relational System",
  journal =      j-TODS,
  volume =       "3",
  number =       "1",
  pages =        "1--31",
  month =        mar,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-1/p1-minker/p1-minker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-1/p1-minker/",
  abstract =     "An inferential relational system is one in which data
                 in the system consists of both explicit facts and
                 general axioms (or ``views''). The general axioms are
                 used together with the explicit facts to derive the
                 facts that are implicit (virtual relations) within the
                 system. A top-down algorithm, as used in artificial
                 intelligence work, is described to develop inferences
                 within the system. The top-down approach starts with
                 the query, a conjunction of relations, to be answered.
                 Either a relational fact solves a given relation in a
                 conjunct, or the relation is replaced by a conjunct of
                 relations which must be solved to solve the given
                 relation. The approach requires that one and only one
                 relation in a conjunction be replaced (or expanded) by
                 the given facts and general axioms. The decision to
                 expand only a single relation is termed a selection
                 function. It is shown for relational systems that such
                 a restriction still guarantees that a solution to the
                 problem will be found if one exists.\par

                 The algorithm provides for heuristic direction in the
                 search process. Experimental results are presented
                 which illustrate the techniques. A bookkeeping
                 mechanism is described which permits one to know when
                 subproblems are solved. It further facilitates the
                 outputting of reasons for the deductively found answer
                 in a coherent fashion.",
  acknowledgement = ack-nhfb,
  annote =       "Data in the system consists of both explicit facts and
                 general axioms. The top-down approach starts with the
                 query, a conjunction of relations, to be answered.
                 Either a relational fact solves a given relation in a
                 conjunct, or the relation is replaced by a conjunct of
                 relations which must be solved to solve the given
                 relation. Experimental results are presented which
                 illustrate the techniques.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "answer and reason extraction; data base systems;
                 heuristics; inference mechanism; logic; predicate
                 calculus; relational databases; search strategy;
                 selection function; top-down search; virtual
                 relations",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Search process}",
}

@Article{Tuel:1978:ORP,
  author =       "William G. {Tuel, Jr.}",
  title =        "Optimum Reorganization Points for Linearly Growing
                 Files",
  journal =      j-TODS,
  volume =       "3",
  number =       "1",
  pages =        "32--40",
  month =        mar,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-1/p32-tuel/p32-tuel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-1/p32-tuel/",
  abstract =     "The problem of finding optimal reorganization
                 intervals for linearly growing files is solved. An
                 approximate reorganization policy, independent of file
                 lifetime, is obtained. Both the optimum and approximate
                 policies are compared to previously published results
                 using a numerical example.",
  acknowledgement = ack-nhfb,
  annote =       "The problem of finding optimal reorganization
                 intervals for linearly growing files is solved.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing --- file organization; database; file
                 organization; optimization; physical database design
                 TODS, data base systems; reorganization",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Yu:1978:END,
  author =       "C. T. Yu and W. S. Luk and M. K. Siu",
  title =        "On the Estimation of the Number of Desired Records
                 with Respect to a Given Query",
  journal =      j-TODS,
  volume =       "3",
  number =       "1",
  pages =        "41--56",
  month =        mar,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-1/p41-yu/p41-yu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-1/p41-yu/",
  abstract =     "The importance of the estimation of the number of
                 desired records for a given query is outlined. Two
                 algorithms for the estimation in the ``closest
                 neighbors problem'' are presented. The numbers of
                 operations of the algorithms are $ O(m \ell^2) $ and $
                 O(m \ell) $, where $m$ is the number of clusters and $
                 \ell $ is the ``length'' of the query.",
  acknowledgement = ack-nhfb,
  annote =       "Two Algorithms for the estimation in the `closest
                 neighbors problem'",
  classification = "901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "closest neighbors; database; estimate; information
                 science, CTYu selectivity TODS; query",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}",
}

@Article{Su:1978:CCS,
  author =       "Stanley Y. W. Su and Ahmed Emam",
  title =        "{CASDAL}: {{\em CAS\/}SM}'s {{\em DA\/}}ta {{\em
                 L\/}}anguage",
  journal =      j-TODS,
  volume =       "3",
  number =       "1",
  pages =        "57--91",
  month =        mar,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1978-3-1/p57-su/",
  abstract =     "CASDAL is a high level data language designed and
                 implemented for the database machine CASSM. The
                 language is used for the manipulation and maintenance
                 of a database using an unnormalized (hierarchically
                 structured) relational data model. It also has
                 facilities to define, modify, and maintain the data
                 model definition. The uniqueness of CASDAL lies in its
                 power to specify complex operations in terms of several
                 new language constructs and its concepts of tagging or
                 marking tuples and of matching values when walking from
                 relation to relation. The language is a result of a
                 top-down design and development effort for a database
                 machine in which high level language constructs are
                 directly supported by the hardware. This paper (1)
                 gives justifications for the use of an unnormalized
                 relational model on which the language is based, (2)
                 presents the CASDAL language constructs with examples,
                 and (3) describes CASSM's architecture and hardware
                 primitives which match closely with the high level
                 language constructs and facilitate the translation
                 process. This paper also attempts to show how the
                 efficiency of the language and the translation task can
                 be achieved and simplified in a system in which the
                 language is the result of a top-down system design and
                 development.",
  acknowledgement = ack-nhfb,
  annote =       "CASDAL is a high level data language for the database
                 machine CASSM. It uses an unnormalized (hierarchically
                 structured) relational data model. This paper (1)
                 justifies the use of this model (2) presents the Casdal
                 language constructs with examples, and (3) describes
                 CASSM's architecture.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative memory; computer programming languages;
                 data language; database; nonprocedural language; query
                 language; relational model; SYWSu hardware support
                 database machine TODS, data base systems",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3); Information Systems --- Database
                 Management --- Languages (H.2.3): {\bf Query
                 languages}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Relational databases}",
}

@Article{Chin:1978:SSD,
  author =       "Francis Y. Chin",
  title =        "Security in Statistical Databases for Queries with
                 Small Counts",
  journal =      j-TODS,
  volume =       "3",
  number =       "1",
  pages =        "92--104",
  month =        mar,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-1/p92-chin/p92-chin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-1/p92-chin/",
  abstract =     "The security problem of statistical databases
                 containing anonymous but individual records which may
                 be evaluated by queries about sums and averages is
                 considered. A model, more realistic than the previous
                 ones, is proposed, in which nonexisting records for
                 some keys can be allowed. Under the assumption that the
                 system protects the individual's information by the
                 well-known technique which avoids publishing summaries
                 with small counts, several properties about the system
                 and a necessary and sufficient condition for
                 compromising the database have been derived. The
                 minimum number of queries needed to compromise the
                 database is also discussed.",
  acknowledgement = ack-nhfb,
  annote =       "Under the assumption that the system protects the
                 individual's information by the technique which avoids
                 publishing summaries with small counts, properties
                 about the system and a necessary and sufficient
                 condition for compromising the database have been
                 derived.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromisability; data base systems; data processing
                 --- security of data; data security; protection;
                 statistical databases",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Query processing};
                 Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Hendrix:1978:DNL,
  author =       "Gary G. Hendrix and Earl D. Sacerdoti and Daniel
                 Sagalowicz and Jonathan Slocum",
  title =        "Developing a Natural Language Interface to Complex
                 Data",
  journal =      j-TODS,
  volume =       "3",
  number =       "2",
  pages =        "105--147",
  month =        jun,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/Ai.misc.bib; Compendex database;
                 Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-2/p105-hendrix/p105-hendrix.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-2/p105-hendrix/",
  abstract =     "Aspects of an intelligent interface that provides
                 natural language access to a large body of data
                 distributed over a computer network are described. The
                 overall system architecture is presented, showing how a
                 user is buffered from the actual database management
                 systems (DBMSs) by three layers of insulating
                 components. These layers operate in series to convert
                 natural language queries into calls to DBMSs at remote
                 sites. Attention is then focused on the first of the
                 insulating components, the natural language system. A
                 pragmatic approach to language access that has proved
                 useful for building interfaces to databases is
                 described and illustrated by examples. Special language
                 features that increase system usability, such as
                 spelling correction, processing of incomplete inputs,
                 and run-time system personalization, are also
                 discussed. The language system is contrasted with other
                 work in applied natural language processing, and the
                 system's limitations are analyzed.",
  acknowledgement = ack-nhfb,
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database access; human engineering;
                 intelligent access semantic grammar human engineering
                 run-time personalization, computer interfaces;
                 intelligent interface; natural language; Natural
                 Language, Intelligent Interface, Database Access,
                 Semantic Grammar, Human Engineering, Runtime
                 Personalization; run-time personalization; semantic
                 grammar",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3); Information Systems --- Database
                 Management --- Physical Design (H.2.2): {\bf Access
                 methods}",
}

@Article{Langdon:1978:NAP,
  author =       "Glen G. {Langdon, Jr.}",
  title =        "A Note on Associative Processors for Data Management",
  journal =      j-TODS,
  volume =       "3",
  number =       "2",
  pages =        "148--158",
  month =        jun,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-2/p148-langdon/p148-langdon.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-2/p148-langdon/",
  abstract =     "Associative ``logic-per-track'' processors for data
                 management are examined from a technological and
                 engineering point of view. Architectural and design
                 decisions are discussed. Some alternatives to the
                 design of comparators, garbage collection, and domain
                 extraction for architectures like the Relational
                 Associative Processor (RAP) are offered.",
  acknowledgement = ack-nhfb,
  annote =       "Associative ``logic-per-track'' processors for data
                 management are examined from a technological and
                 engineering point of view (RAP).",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative processors; computer operating systems;
                 data base systems, hardware support database machine
                 TODS; database machines",
  subject =      "Information Systems --- Database Management (H.2)",
}

@Article{Kluge:1978:DFM,
  author =       "Werner E. Kluge",
  title =        "Data File Management in Shift-Register Memories",
  journal =      j-TODS,
  volume =       "3",
  number =       "2",
  pages =        "159--177",
  month =        jun,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-2/p159-kluge/p159-kluge.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-2/p159-kluge/",
  abstract =     "The paper proposes a shift-register memory, structured
                 as a two-dimensional array of uniform shift-register
                 loops which are linked by flow-steering switches, whose
                 switch control scheme is tailored to perform with great
                 efficiency data management operations on sequentially
                 organized files. The memory operates in a linear
                 input/output mode to perform record insertion,
                 deletion, and relocation on an existing file, and in a
                 sublinear mode for rapid internal file movement to
                 expedite file positioning and record retrieval and
                 update operations.\par

                 The memory, implemented as a large capacity
                 charge-coupled device or magnetic domain memory,
                 permits efficient data management on very large
                 databases at the level of secondary storage and lends
                 itself to applications as a universal disk replacement,
                 particularly in database computers.",
  acknowledgement = ack-nhfb,
  annote =       "Shift-register memory, structured as a two-dimensional
                 array tailored to perform with great efficiency data
                 management operations on sequentially organized files",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data transformations; deletion; hardware support
                 database machine TODS, computer operating systems;
                 insertion; LIFO/FIFO operation modes; management of
                 sequentially organized files; record retrieval;
                 relocation; shift-register memories; updating",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3)",
}

@Article{Rosenkrantz:1978:SLC,
  author =       "David J. Rosenkrantz and Richard E. Stearns and Philip
                 M. {Lewis, II}",
  title =        "System Level Concurrency Control for Distributed
                 Database Systems",
  journal =      j-TODS,
  volume =       "3",
  number =       "2",
  pages =        "178--198",
  month =        jun,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-2/p178-rosenkrantz/p178-rosenkrantz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-2/p178-rosenkrantz/",
  abstract =     "A distributed database system is one in which the
                 database is spread among several sites and application
                 programs ``move'' from site to site to access and
                 update the data they need. The concurrency control is
                 that portion of the system that responds to the read
                 and write requests of the application programs. Its job
                 is to maintain the global consistency of the
                 distributed database while ensuring that the
                 termination of the application programs is not
                 prevented by phenomena such as deadlock. We assume each
                 individual site has its own local concurrency control
                 which responds to requests at that site and can only
                 communicate with concurrency controls at other sites
                 when an application program moves from site to site,
                 terminates, or aborts.\par

                 This paper presents designs for several distributed
                 concurrency controls and demonstrates that they work
                 correctly. It also investigates some of the
                 implications of global consistency of a distributed
                 database and discusses phenomena that can prevent
                 termination of application programs.",
  acknowledgement = ack-nhfb,
  annote =       "Later arriving transactions may be aborted if not yet
                 in the commit stage.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; consistency; data base systems; database;
                 deadlock; deadly embrace; distributed; integrity; lock;
                 readers and writers; restart; rollback; transaction",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Concurrency}",
}

@Article{Delobel:1978:NHD,
  author =       "Claude Delobel",
  title =        "Normalization and Hierarchical Dependencies in the
                 Relational Data Model",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "201--222",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p201-delobel/p201-delobel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p201-delobel/",
  abstract =     "The purpose of this paper is to present a new approach
                 to the conceptual design of logical schemata for
                 relational databases. One-to-one, one-to-many, and
                 many-to-many relationships between the attributes of
                 database relations are modeled by means of functional
                 dependencies and multivalued dependencies. A new type
                 of dependency is introduced: first-order hierarchical
                 decomposition. The properties of this new type of
                 dependency are studied and related to the normalization
                 process of relations. The relationship between the
                 concept of first-order hierarchical decomposition and
                 the notion of hierarchical organization of data is
                 discussed through the normalization process.",
  acknowledgement = ack-nhfb,
  annote =       "One-to-one, one-to-many relationships between the
                 attributes of database relations are modeled by means
                 of functional dependencies and multivalued
                 dependencies. A new type of dependency is first-order
                 hierarchical.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; data model; first-order
                 hierarchical dependency; functional dependency;
                 hierarchical schema; multivalued dependency;
                 normalization process; relational database; relational
                 model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Information Systems
                 --- Database Management --- Logical Design (H.2.1):
                 {\bf Schema and subschema}",
}

@Article{Smith:1978:SPD,
  author =       "Alan Jay Smith",
  title =        "Sequentiality and Prefetching in Database Systems",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "223--247",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p223-smith/p223-smith.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p223-smith/",
  abstract =     "Sequentiality of access is an inherent characteristic
                 of many database systems. We use this observation to
                 develop an algorithm which selectively prefetches data
                 blocks ahead of the point of reference. The number of
                 blocks prefetched is chosen by using the empirical run
                 length distribution and conditioning on the observed
                 number of sequential block references immediately
                 preceding reference to the current block. The optimal
                 number of blocks to prefetch is estimated as a function
                 of a number of ``costs,'' including the cost of
                 accessing a block not resident in the buffer (a miss),
                 the cost of fetching additional data blocks at fault
                 times, and the cost of fetching blocks that are never
                 referenced. We estimate this latter cost, described as
                 memory pollution, in two ways. We consider the
                 treatment (in the replacement algorithm) of prefetched
                 blocks, whether they are treated as referenced or not,
                 and find that it makes very little difference. Trace
                 data taken from an operational IMS database system is
                 analyzed and the results are presented. We show how to
                 determine optimal block sizes. We find that
                 anticipatory fetching of data can lead to significant
                 improvements in system operation.",
  acknowledgement = ack-nhfb,
  annote =       "An algorithm which selectively prefetches data blocks
                 ahead of the point of reference. The optimal number of
                 blocks to prefetch is estimated as a function, the cost
                 (a miss), the cost of fetching blocks that are never
                 referenced of fetching additional data blocks, and the
                 cost of fetching blocks that are never referenced.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer management; database systems; dynamic
                 programming; IMS; paging; prefetching; read-ahead
                 caches caching buffer management TODS, data base
                 systems; sequentiality",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4)",
}

@Article{Schlageter:1978:PSD,
  author =       "Gunter Schlageter",
  title =        "Process Synchronization in Database Systems",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "248--271",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See errata report in \cite{Bernstein:1979:CPS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p248-schlageter/p248-schlageter.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p248-schlageter/",
  abstract =     "The problem of process synchronization in database
                 systems is analyzed in a strictly systematic way, on a
                 rather abstract level; the abstraction is chosen such
                 that the essential characteristics of the problem can
                 be distinctly modeled and investigated. Using a small
                 set of concepts, a consistent description of the whole
                 problem is developed; many widely used, but only
                 vaguely defined, notions are defined exactly within
                 this framework. The abstract treatment of the problem
                 immediately leads to practically useful insights with
                 respect to possible solutions, although
                 implementational aspects are not discussed in detail.",
  acknowledgement = ack-nhfb,
  annote =       "Process synchronization in database systems is
                 analyzed on a rather abstract level. [see Bernstein for
                 comments]",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database consistency; database systems; integrity;
                 locking; operating system support TODS, data base
                 systems; parallel process systems; process
                 synchronization",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management (H.2)",
}

@Article{Hollaar:1978:SMP,
  author =       "Lee A. Hollaar",
  title =        "Specialized Merge Processor Networks for Combining
                 Sorted Lists",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "272--284",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p272-hollaar/p272-hollaar.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p272-hollaar/",
  abstract =     "In inverted file database systems, index lists
                 consisting of pointers to items within the database are
                 combined to form a list of items which potentially
                 satisfy a user's query. This list merging is similar to
                 the common data processing operation of combining two
                 or more sorted input files to form a sorted output
                 file, and generally represents a large percentage of
                 the computer time used by the retrieval system.
                 Unfortunately, a general purpose digital computer is
                 better suited for complicated numeric processing rather
                 than the simple combining of data. The overhead of
                 adjusting and checking pointers, aligning data, and
                 testing for completion of the operation overwhelm the
                 processing of the data.\par

                 A specialized processor can perform most of these
                 overhead operations in parallel with the processing of
                 the data, thereby offering speed increases by a factor
                 from 10 to 100 over conventional computers, depending
                 on whether a higher speed memory is used for storing
                 the lists. These processors can also be combined into
                 networks capable of directly forming the result of a
                 complex expression, with another order of magnitude
                 speed increase possible. The programming and operation
                 of these processors and networks is discussed, and
                 comparisons are made with the speed and efficiency of
                 conventional general purpose computers.",
  acknowledgement = ack-nhfb,
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backend processors; binary tree networks; computer
                 architecture --- program processors; computer system
                 architecture; full text retrieval systems; hardware
                 support database machine TODS, data base systems;
                 inverted file databases; nonnumeric processing;
                 pipelined networks; sorted list merging",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3);
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Haerder:1978:IGA,
  author =       "Theo Haerder",
  title =        "Implementing a Generalized Access Path Structure for a
                 Relational Database System",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "285--298",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p285-haerder/p285-haerder.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p285-haerder/",
  abstract =     "A new kind of implementation technique for access
                 paths connecting sets of tuples qualified by attribute
                 values is described. It combines the advantages of
                 pointer chain and multilevel index implementation
                 techniques. Compared to these structures the
                 generalized access path structure is at least
                 competitive in performing retrieval and update
                 operations, while a considerable storage space saving
                 is gained. Some additional features of this structure
                 support $m$-way joins and the evaluation of
                 multirelation queries, and allow efficient checks of
                 integrity assertions and simple reorganization
                 schemes.",
  acknowledgement = ack-nhfb,
  annote =       "Implementation technique for access paths connecting
                 sets of tuples qualified by attribute values combines
                 the advantages of pointer chains and multilevel
                 indexes. Features of this structure support m-way
                 joins.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems, Harder multi-relation indices
                 TODS",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Yu:1978:PP,
  author =       "C. T. Yu and M. K. Siu and K. Lam",
  title =        "On a Partitioning Problem",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "299--309",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p299-yu/p299-yu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p299-yu/",
  abstract =     "This paper investigates the problem of locating a set
                 of ``boundary points'' of a large number of records.
                 Conceptually, the boundary points partition the records
                 into subsets of roughly the same number of elements,
                 such that the key values of the records in one subset
                 are all smaller or all larger than those of the records
                 in another subset. We guess the locations of the
                 boundary points by linear interpolation and check their
                 accuracy by reading the key values of the records on
                 one pass. This process is repeated until all boundary
                 points are determined. Clearly, this problem can also
                 be solved by performing an external tape sort. Both
                 analytical and empirical results indicate that the
                 number of passes required is small in comparison with
                 that in an external tape sort. This kind of record
                 partitioning may be of interest in setting up a
                 statistical database system.",
  acknowledgement = ack-nhfb,
  annote =       "Boundary points partition the records into subsets of
                 roughly the same number of elements. We guess the
                 locations of the boundary points by linear
                 interpolation and check their accuracy by reading the
                 key values of the records on one pass. This process is
                 repeated until all boundary points are determined.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CTYu TODS, data base systems; external sort; key
                 value; partition; passes; tape probability",
  subject =      "Computing Methodologies --- Image Processing And
                 Computer Vision --- Segmentation (I.4.6): {\bf Region
                 growing, partitioning}",
}

@Article{Fagin:1978:AM,
  author =       "Ronald Fagin",
  title =        "On an Authorization Mechanism",
  journal =      j-TODS,
  volume =       "3",
  number =       "3",
  pages =        "310--319",
  month =        sep,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-3/p310-fagin/p310-fagin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-3/p310-fagin/",
  abstract =     "Griffiths and Wade ({\em ACM Trans. Database Syst.
                 1,3}, (Sept. 1976), 242-255) have defined a dynamic
                 authorization mechanism that goes beyond the
                 traditional password approach. A database user can
                 grant or revoke privileges (such as to read, insert, or
                 delete) on a file that he has created. Furthermore, he
                 can authorize others to grant these same privileges.
                 The database management system keeps track of a
                 directed graph, emanating from the creator, of granted
                 privileges. The nodes of the graph correspond to users,
                 and the edges (each of which is labeled with a
                 timestamp) correspond to grants. The edges are of two
                 types, corresponding to whether or not the recipient of
                 the grant has been given the option to make further
                 grants of this privilege. Furthermore, for each pair $
                 A, B $ of nodes, there can be no more than one edge of
                 each type from $A$ to $B$. We modify this approach by
                 allowing graphs in which there can be multiple edges of
                 each type from one node to another. We prove
                 correctness (in a certain strong sense) for our
                 modified authorization mechanism. Further, we show by
                 example that under the original mechanism, the system
                 might forbid some user from exercising or granting a
                 privilege that he ``should'' be allowed to exercise or
                 grant.",
  acknowledgement = ack-nhfb,
  annote =       "We prove correctness for our modified authorization
                 mechanism",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access control; authorization; data base systems;
                 database; privacy; proof of correctness; protection;
                 revocation; security",
  subject =      "Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Salton:1978:GSC,
  author =       "G. Salton and A. Wong",
  title =        "Generation and Search of Clustered Files",
  journal =      j-TODS,
  volume =       "3",
  number =       "4",
  pages =        "321--346",
  month =        dec,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-4/p321-salton/p321-salton.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-4/p321-salton/",
  abstract =     "A classified, or clustered file is one where related,
                 or similar records are grouped into classes, or
                 clusters of items in such a way that all items within a
                 cluster are jointly retrievable. Clustered files are
                 easily adapted to broad and narrow search strategies,
                 and simple file updating methods are available. An
                 inexpensive file clustering method applicable to large
                 files is given together with appropriate file search
                 methods. An abstract model is then introduced to
                 predict the retrieval effectiveness of various search
                 methods in a clustered file environment. Experimental
                 evidence is included to test the versatility of the
                 model and to demonstrate the role of various parameters
                 in the cluster search process.",
  acknowledgement = ack-nhfb,
  annote =       "Automatic classification for information retrieval",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "automatic classification; cluster searching; clustered
                 files; data processing; fast classification; file
                 organization; probabilistic models",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3): {\bf Clustering}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3): {\bf Search
                 process}",
}

@Article{Banerjee:1978:CCD,
  author =       "Jayanta Banerjee and Richard I. Baum and David K.
                 Hsiao",
  title =        "Concepts and Capabilities of a Database Computer",
  journal =      j-TODS,
  volume =       "3",
  number =       "4",
  pages =        "347--384",
  month =        dec,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-4/p347-banerjee/p347-banerjee.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-4/p347-banerjee/",
  abstract =     "The concepts and capabilities of a database computer
                 (DBC) are given in this paper. The proposed design
                 overcomes many of the traditional problems of database
                 system software and is one of the first to describe a
                 complete data-secure computer capable of handling large
                 databases.\par

                 This paper begins by characterizing the major problems
                 facing today's database system designers. These
                 problems are intrinsically related to the nature of
                 conventional hardware and can only be solved by
                 introducing new architectural concepts. Several such
                 concepts are brought to bear in the later sections of
                 this paper. These architectural principles have a major
                 impact upon the design of the system and so they are
                 discussed in some detail. A key aspect of these
                 principles is that they can be implemented with
                 near-term technology. The rest of the paper is devoted
                 to the functional characteristics and the theory of
                 operation of the DBC. The theory of operation is based
                 on a series of abstract models of the components and
                 data structures employed by the DBC. These models are
                 used to illustrate how the DBC performs access
                 operations, manages data structures and security
                 specifications, and enforces security requirements.
                 Short Algol-like algorithms are used to show how these
                 operations are carried out. This part of the paper
                 concludes with a high-level description of the DBC
                 organization. The actual details of the DBC hardware
                 are quite involved and so their presentation is not the
                 subject of this paper.\par

                 A sample database is included in the Appendix to
                 illustrate the working of the security and clustering
                 mechanisms of the DBC.",
  acknowledgement = ack-nhfb,
  annote-1 =     "The concepts of a database computer (DBC) are given.
                 The theory of operation is based on abstract models.
                 The DBC performs access operations, manages data
                 structures and security specifications.",
  annote-2 =     "The correct author order (from the running heads and
                 table of contents) is Banerjee, Baum, Hsiao: the
                 article cover page has Banerjee, Hsiao, Baum, because
                 the first two share a common address.",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "clustering; content-addressable memory; data base
                 systems; database computers; hardware support machine
                 TODS, computer architecture; keywords; mass memory;
                 performance; security; structure memory",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Clustering}",
}

@Article{Bradley:1978:EOC,
  author =       "J. Bradley",
  title =        "An Extended Owner-Coupled Set Data Model and Predicate
                 Calculus for Database Management",
  journal =      j-TODS,
  volume =       "3",
  number =       "4",
  pages =        "385--416",
  month =        dec,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-4/p385-bradley/p385-bradley.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-4/p385-bradley/",
  abstract =     "A data model is presented, based on the extension of
                 the concept of a DBTG owner-coupled set to permit {\em
                 static\/} and {\em dynamic\/} sets and a new kind of
                 set referred to as a {\em virtual\/} set. The notion of
                 {\em connection fields\/} is introduced, and it is
                 shown how connection fields may be used to construct
                 derived information bearing set names, and hence permit
                 the specification of (dynamic) sets which are not
                 predeclared in a schema. Virtual sets are shown to
                 reflect the functional dependencies which can exist
                 within a file. A technique which permits the data model
                 to be fully described diagrammatically by {\em extended
                 Bachman diagrams\/} is described. A predicate calculus
                 for manipulation of this data model is presented.
                 Expressions written in this calculus are compared with
                 corresponding expressions in a relational predicate
                 calculus, DSL ALPHA. An argument for the relational
                 completeness of the language is given.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Codasyl DBTG; connection field; data base systems; DSL
                 AlPHA; dynamic set; extended Bachman diagram; extended
                 owner-coupled set data model; extended owner-coupled
                 set predicate calculus; functional dependency;
                 information bearing set name; owner-coupled set; static
                 set; virtual set",
  subject =      "Information Systems --- Database Management (H.2)",
}

@Article{Shneiderman:1978:IHF,
  author =       "Ben Shneiderman",
  title =        "Improving the Human Factors Aspect of Database
                 Interactions",
  journal =      j-TODS,
  volume =       "3",
  number =       "4",
  pages =        "417--439",
  month =        dec,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-4/p417-shneiderman/p417-shneiderman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-4/p417-shneiderman/",
  abstract =     "The widespread dissemination of computer and
                 information systems to nontechnically trained
                 individuals requires a new approach to the design and
                 development of database interfaces. This paper provides
                 the motivational background for controlled
                 psychological experimentation in exploring the
                 person\slash machine interface. Frameworks for the
                 reductionist approach are given, research methods
                 discussed, research issues presented, and a small
                 experiment is offered as an example of what can be
                 accomplished. This experiment is a comparison of
                 natural and artificial language query facilities.
                 Although subjects posed approximately equal numbers of
                 valid queries with either facility, natural language
                 users made significantly more invalid queries which
                 could not be answered from the database that was
                 described.",
  acknowledgement = ack-nhfb,
  classification = "461; 723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; data models; database systems;
                 experimentation; human engineering; human factors;
                 natural language interfaces; psychology; query
                 languages; systems science and cybernetics --- man
                 machine systems",
  subject =      "Information Systems --- Models and Principles ---
                 User/Machine Systems (H.1.2): {\bf Human factors};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Information Systems
                 --- Database Management --- Languages (H.2.3): {\bf
                 Query languages}; Information Systems --- Database
                 Management --- Systems (H.2.4)",
}

@Article{Comer:1978:DOI,
  author =       "Douglas Comer",
  title =        "The Difficulty of Optimum Index Selection",
  journal =      j-TODS,
  volume =       "3",
  number =       "4",
  pages =        "440--445",
  month =        dec,
  year =         "1978",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1978-3-4/p440-comer/p440-comer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1978-3-4/p440-comer/",
  abstract =     "Given a file on a secondary store in which each record
                 has several attributes, it is usually advantageous to
                 build an index mechanism to decrease the cost of
                 conducting transactions to the file. The problem of
                 selecting attributes over which to index has been
                 studied in the context of various storage structures
                 and access assumptions. One algorithm to make an
                 optimum index selection requires 2 $k$ steps in the
                 worst case, where $k$ is the number of attributes in
                 the file. We examine the question of whether a more
                 efficient algorithm might exist and show that even
                 under a simple cost criterion the problem is
                 computationally difficult in a precise sense. Our
                 results extend directly to other related problems where
                 the cost of the index depends on fixed values which are
                 assigned to each attribute. Some practical implications
                 are discussed.",
  acknowledgement = ack-nhfb,
  annote =       "Theorem: optimum index selection problem OISP is
                 NP-complete for files of degree $ d \ge 2 $.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "attribute selection; complexity; index selection;
                 physical database design; secondary index",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Content Analysis and Indexing (H.3.1):
                 {\bf Indexing methods}",
}

@Article{Babb:1979:IRD,
  author =       "E. Babb",
  title =        "Implementing a Relational Database by Means of
                 Specialized Hardware",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "1--29",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p1-babb/p1-babb.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p1-babb/",
  abstract =     "New hardware is described which allows the rapid
                 execution of queries demanding the joining of
                 physically stored relations. The main feature of the
                 hardware is a special store which can rapidly remember
                 or recall data. This data might be pointers from one
                 file to another, in which case the memory helps with
                 queries on joins of files. Alternatively, the memory
                 can help remove redundant data during projection[s??],
                 giving a considerable speed advantage over conventional
                 hardware.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bit array; CAFS; content addressing; database;
                 hardware support machine bit vector filter
                 probabilistic semi-join TODS, data base systems;
                 hashing; information retrieval; join; projection;
                 relational model; selection; special hardware",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Beeri:1979:CPR,
  author =       "Catriel Beeri and Philip A. Bernstein",
  title =        "Computational Problems Related to the Design of Normal
                 Form Relational Schemas",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "30--59",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Compiler/prog.lang.theory.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: errata in ACM Transactions on
                 Database Systems, Vol. 4 No. 3, Sep. 1979, pp. 396.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p30-beeri/p30-beeri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p30-beeri/",
  abstract =     "Problems related to functional dependencies and the
                 algorithmic design of relational schemas are examined.
                 Specifically, the following results are presented: (1)
                 a tree model of derivations of functional dependencies
                 from other functional dependencies; (2) a linear-time
                 algorithm to test if a functional dependency is in the
                 closure of a set of functional dependencies; (3) a
                 quadratic-time implementation of Bernstein's third
                 normal form schema synthesis algorithm.
                 \par

                 Furthermore, it is shown that most interesting
                 algorithmic questions about Boyce-Codd normal form and
                 keys are {\em NP\/}-complete and are therefore probably
                 not amenable to fast algorithmic solutions.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1); Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Lockemann:1979:DAD,
  author =       "Peter C. Lockemann and Heinrich C. Mayr and Wolfgang
                 H. Weil and Wolfgang H. Wohlleber",
  title =        "Data Abstractions for Database Systems",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "60--75",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p60-lockemann/p60-lockemann.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p60-lockemann/",
  abstract =     "Data abstractions were originally conceived as a
                 specification tool in programming. They also appear to
                 be useful for exploring and explaining the capabilities
                 and shortcomings of the data definition and
                 manipulation facilities of present-day database
                 systems. Moreover they may lead to new approaches to
                 the design of these facilities. In the first section
                 the paper introduces an axiomatic method for specifying
                 data abstractions and, on that basis, gives precise
                 meaning to familiar notions such as data model, data
                 type, and database schema. In a second step the various
                 possibilities for specifying data types within a given
                 data model are examined and illustrated. It is shown
                 that data types prescribe the individual operations
                 that are allowed within a database. Finally, some
                 additions to the method are discussed which permit the
                 formulation of interrelationships between arbitrary
                 operations.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data type; data abstraction; data base
                 systems; data definition language; data manipulation
                 language; data model; data structure; data type;
                 database consistency; database design; database schema;
                 integrity constraints; specification",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Languages (H.2.3): {\bf Data
                 manipulation languages (DML)}; Information Systems ---
                 Database Management --- Logical Design (H.2.1): {\bf
                 Data models}; Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Schema and
                 subschema}",
}

@Article{Denning:1979:TTS,
  author =       "Dorothy E. Denning and Peter J. Denning and Mayer D.
                 Schwartz",
  title =        "The Tracker: a Threat to Statistical Database
                 Security",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "76--96",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p76-denning/p76-denning.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p76-denning/",
  abstract =     "The query programs of certain databases report raw
                 statistics for query sets, which are groups of records
                 specified implicitly by a characteristic formula. The
                 raw statistics include query set size and sums of
                 powers of values in the query set. Many users and
                 designers believe that the individual records will
                 remain confidential as long as query programs refuse to
                 report the statistics of query sets which are too
                 small. It is shown that the compromise of small query
                 sets can in fact almost always be accomplished with the
                 help of characteristic formulas called trackers. J.
                 Schl{\"o}rer's individual tracker is reviewed; it is
                 derived from known characteristics of a given
                 individual and permits deducing additional
                 characteristics he may have. The general tracker is
                 introduced: It permits calculating statistics for
                 arbitrary query sets, without requiring preknowledge of
                 anything in the database. General trackers always exist
                 if there are enough distinguishable classes of
                 individuals in the database, in which case the trackers
                 have a simple form. Almost all databases have a general
                 tracker, and general trackers are almost always easy to
                 find. Security is not guaranteed by the lack of a
                 general tracker.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; data base systems; data processing;
                 data security; database security; secure query
                 functions; statistical database; tracker",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0): {\bf Security, integrity, and
                 protection**}; Information Systems --- Database
                 Management --- Database Applications (H.2.8): {\bf
                 Statistical databases}",
}

@Article{Dobkin:1979:SDP,
  author =       "David Dobkin and Anita K. Jones and Richard J.
                 Lipton",
  title =        "Secure Databases: Protection Against User Influence",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "97--106",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p97-dobkin/p97-dobkin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p97-dobkin/",
  abstract =     "Users may be able to compromise databases by asking a
                 series of questions and then inferring new information
                 from the answers. The complexity of protecting a
                 database against this technique is discussed here.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromise; data base systems; database; inference;
                 information flow; protection; security; Security TODS,
                 data processing; statistical query",
  subject =      "Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Kent:1979:LRB,
  author =       "William Kent",
  title =        "Limitations of Record-Based Information Models",
  journal =      j-TODS,
  volume =       "4",
  number =       "1",
  pages =        "107--131",
  month =        mar,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-1/p107-kent/p107-kent.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-1/p107-kent/",
  abstract =     "Record structures are generally efficient, familiar,
                 and easy to use for most current data processing
                 applications. But they are not complete in their
                 ability to represent information, nor are they fully
                 self-describing.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual model; data model; entities; first normal
                 form; information model; normalization; records;
                 relationships; semantic model",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Normal forms}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Yao:1979:OQE,
  author =       "S. Bing Yao",
  title =        "Optimization of Query Evaluation Algorithms",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "133--155",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p133-yao/p133-yao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p133-yao/",
  abstract =     "A model of database storage and access is presented.
                 The model represents many evaluation algorithms as
                 special cases, and helps to break a complex algorithm
                 into simple access operations. Generalized access cost
                 equations associated with the model are developed and
                 analyzed. Optimization of these cost equations yields
                 an optimal access algorithm which can be synthesized by
                 a query subsystem whose design is based on the modular
                 access operations.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data base systems;
                 data manipulation language; database optimization;
                 inverted file; query language; query languages; query
                 optimization; relational data model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Data models}; Information Systems ---
                 Database Management --- Languages (H.2.3): {\bf Data
                 manipulation languages (DML)}",
}

@Article{Schwartz:1979:LQS,
  author =       "M. D. Schwartz and D. E. Denning and P. J. Denning",
  title =        "Linear Queries in Statistical Databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "156--167",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p156-schwartz/p156-schwartz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p156-schwartz/",
  abstract =     "A database is compromised if a user can determine the
                 data elements associated with keys which he did not
                 know previously. If it is possible, compromise can be
                 achieved by posing a finite set of queries over sets of
                 data elements and employing initial information to
                 solve the resulting system of equations. Assuming the
                 allowable queries are linear, that is, weighted sums of
                 data elements, we show how compromise can be achieved
                 and we characterize the maximal initial information
                 permitted of a user in a secure system. When compromise
                 is possible, the initial information and the number of
                 queries required to achieve it is surprisingly small.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; data base systems; data processing
                 --- security of data; data security; database security;
                 inference; linear query; secure query functions;
                 statistical database",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Query processing};
                 Information Systems --- Database Management --- General
                 (H.2.0): {\bf Security, integrity, and protection**}",
}

@Article{Aho:1979:OPM,
  author =       "Alfred V. Aho and Jeffrey D. Ullman",
  title =        "Optimal Partial-Match Retrieval When Fields are
                 Independently Specified",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "168--179",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Graphics/siggraph/79.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p168-aho/p168-aho.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p168-aho/",
  abstract =     "This paper considers the design of a system to answer
                 partial-match queries from a file containing a
                 collection of records, each record consisting of a
                 sequence of fields. A partial-match query is a
                 specification of values for zero or more fields of a
                 record, and the answer to a query is a listing of all
                 records in the file whose fields match the specified
                 values.\par

                 A design is considered in which the file is stored in a
                 set of bins. A formula is derived for the optimal
                 number of bits in a bin address to assign to each
                 field, assuming the probability that a given field is
                 specified in a query is independent of what other
                 fields are specified. Implications of the optimality
                 criterion on the size of bins are also discussed.",
  acknowledgement = ack-nhfb,
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative searching; data processing --- file
                 organization; file organization; hashing; information
                 retrieval; information science; partial-match
                 retrieval; searching",
  oldlabel =     "geom-2",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Retrieval models}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Search process};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Thomas:1979:MCA,
  author =       "Robert H. Thomas",
  title =        "A Majority Consensus Approach to Concurrency Control
                 for Multiple Copy Databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "180--209",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p180-thomas/p180-thomas.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p180-thomas/",
  abstract =     "A ``majority consensus'' algorithm which represents a
                 new solution to the update synchronization problem for
                 multiple copy databases is presented. The algorithm
                 embodies distributed control and can function
                 effectively in the presence of communication and
                 database site outages. The correctness of the algorithm
                 is demonstrated and the cost of using it is analyzed.
                 Several examples that illustrate aspects of the
                 algorithm operation are included in the Appendix.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "clock synchronization; computer networks; concurrency
                 control; data base systems; distributed computation;
                 distributed control; distributed databases;
                 multiprocess systems; update synchronization",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}; Information Systems
                 --- Information Storage and Retrieval --- Systems and
                 Software (H.3.4): {\bf Distributed systems}",
}

@Article{Ries:1979:LGR,
  author =       "Daniel R. Ries and Michael R. Stonebraker",
  title =        "Locking Granularity Revisited",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "210--227",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p210-ries/p210-ries.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p210-ries/",
  abstract =     "Locking granularity refers to the size and hence the
                 number of locks used to ensure the consistency of a
                 database during multiple concurrent updates. In an
                 earlier simulation study we concluded that coarse
                 granularity, such as area or file locking, is to be
                 preferred to fine granularity such as individual page
                 or record locking.\par

                 However, alternate assumptions than those used in the
                 original paper can change that conclusion. First, we
                 modified the assumptions concerning the placement of
                 the locks on the database with respect to the accessing
                 transactions. In the original model the locks were
                 assumed to be well placed. Under worse case and random
                 placement assumptions when only very small transactions
                 access the database, fine granularity is preferable.
                 \par

                 Second, we extended the simulation to model a lock
                 hierarchy where large transactions use large locks and
                 small transactions use small locks. In this scenario,
                 again under the random and worse case lock placement
                 assumptions, fine granularity is preferable if all
                 transactions accessing more than 1 percent of the
                 database use large locks.\par

                 Finally, the simulation was extended to model a ``claim
                 as needed'' locking strategy together with the
                 resultant possibility of deadlock. In the original
                 study all locks were claimed in one atomic operation at
                 the beginning of a transaction. The claim as needed
                 strategy does not change the conclusions concerning the
                 desired granularity.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; database management; locking granularity;
                 locking hierarchies; multiple updates; TODS Ingres,
                 data base systems",
  subject =      "Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Concurrency}; Information Systems --- Database
                 Management (H.2); Information Systems --- Database
                 Management --- Physical Design (H.2.2): {\bf Deadlock
                 avoidance}",
}

@Article{Burkhard:1979:PMH,
  author =       "Walter A. Burkhard",
  title =        "Partial-Match Hash Coding: Benefits of Redundancy",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "228--239",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Graphics/siggraph/79.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p228-burkhard/p228-burkhard.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p228-burkhard/",
  abstract =     "File designs suitable for retrieval from a file of
                 $k$-field records when queries may be partially
                 specified are examined. Storage redundancy is
                 introduced to obtain improved worst-case and
                 average-case performances. The resulting storage
                 schemes are appropriate for replicated distributed
                 database environments; it is possible to improve the
                 overall average and worst-case behavior for query
                 response as well as provide an environment with very
                 high reliability. Within practical systems it will be
                 possible to improve the query response time performance
                 as well as reliability over comparable systems without
                 replication.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access methods; algorithms; analysis; data base
                 systems; data processing --- file organization; data
                 structures; database systems; replication; searching",
  oldlabel =     "geom-100",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Search process}",
}

@Article{Raghavan:1979:EDR,
  author =       "Vijay V. Raghavan and C. T. Yu",
  title =        "Experiments on the Determination of the Relationships
                 Between Terms",
  journal =      j-TODS,
  volume =       "4",
  number =       "2",
  pages =        "240--260",
  month =        jun,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-2/p240-raghavan/p240-raghavan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-2/p240-raghavan/",
  abstract =     "The retrieval effectiveness of an automatic method
                 that uses relevance judgments for the determination of
                 positive as well as negative relationships between
                 terms is evaluated. The term relationships are
                 incorporated into the retrieval process by using a
                 generalized similarity function that has a term match
                 component, a positive term relationship component, and
                 a negative term relationship component. Two strategies,
                 query partitioning and query clustering, for the
                 evaluation of the effectiveness of the term
                 relationships are investigated. The latter appears to
                 be more attractive from linguistic as well as economic
                 points of view. The positive and the negative
                 relationships are verified to be effective both when
                 used individually, and in combination. The importance
                 attached to the term relationship components relative
                 to that of term match component is found to have a
                 substantial effect on the retrieval performance. The
                 usefulness of discriminant analysis as a technique for
                 determining the relative importance of these components
                 is investigated.",
  acknowledgement = ack-nhfb,
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "antonym; document retrieval; feedback; information
                 science; pseudoclassification; semantics; statistical
                 discrimination; synonym; term associations; thesaurus",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3);
                 Information Systems --- Information Storage and
                 Retrieval --- Content Analysis and Indexing (H.3.1):
                 {\bf Thesauruses}; Information Systems --- Database
                 Management --- Database Applications (H.2.8): {\bf
                 Statistical databases}",
}

@Article{Lipski:1979:SIC,
  author =       "Witold {Lipski, Jr.}",
  title =        "On Semantic Issues Connected with Incomplete
                 Information Databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "262--296",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Compendex database;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-3/p262-lipski/p262-lipski.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p262-lipski/",
  abstract =     "Various approaches to interpreting queries in a
                 database with incomplete information are discussed. A
                 simple model of a database is described, based on
                 attributes which can take values in specified attribute
                 domains. Information incompleteness means that instead
                 of having a single value of an attribute, we have a
                 subset of the attribute domain, which represents our
                 knowledge that the actual value, though unknown, is one
                 of the values in this subset. This extends the idea of
                 Codd's null value, corresponding to the case when this
                 subset is the whole attribute domain. A simple query
                 language to communicate with such a system is described
                 and its various semantics are precisely defined. We
                 emphasize the distinction between two different
                 interpretations of the query language--the external
                 one, which refers the queries directly to the real
                 world modeled in an incomplete way by the system, and
                 the internal one, under which the queries refer to the
                 system's information about this world, rather than to
                 the world itself. Both external and internal
                 interpretations are provided with the corresponding
                 sets of axioms which serve as a basis for equivalent
                 transformations of queries. The technique of equivalent
                 transformations of queries is then extensively
                 exploited for evaluating the interpretation of (i.e.,
                 the response to) a query.",
  acknowledgement = ack-nhfb,
  annote =       "Attributes can take values in specified attribute
                 domains. Instead a single value of an attribute, we
                 have a subset of the attribute domain, which represents
                 our knowledge that the actual value, though unknown, is
                 one of the values in this subset. This extends the idea
                 of Codd's null value, corresponding to the case when
                 this subset is the whole attribute domain.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database; incomplete information;
                 model logic; null values; query language semantics;
                 relational model",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Relational databases}; Information Systems ---
                 Database Management --- Languages (H.2.3): {\bf Query
                 languages}",
}

@Article{Aho:1979:TJR,
  author =       "A. V. Aho and C. Beeri and J. D. Ullman",
  title =        "The theory of joins in relational databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "297--314",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See corrigendum \cite{Ullman:1983:CTJ}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-3/p297-aho/p297-aho.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p297-aho/",
  abstract =     "Answering queries in a relational database often
                 requires that the natural join of two or more relations
                 be computed. However, the result of a join may not be
                 what one expects. In this paper we give efficient
                 algorithms to determine whether the join of several
                 relations has the intuitively expected value (is {\em
                 lossless\/}) and to determine whether a set of
                 relations has a subset with a lossy join. These
                 algorithms assume that all data dependencies are
                 functional. We then discuss the extension of our
                 techniques to the case where data dependencies are
                 multivalued.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; decomposition of database schemes;
                 functional dependencies; lossless join; multivalued
                 dependencies; natural join; projection of dependencies;
                 relational databases",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Models and Principles ---
                 Systems and Information Theory (H.1.1)",
}

@Article{Fagin:1979:EHF,
  author =       "Ronald Fagin and J{\"u}rg Nievergelt and Nicholas
                 Pippenger and H. Raymond Strong",
  key =          "Fagin et al.",
  title =        "Extendible Hashing --- a Fast Access Method for
                 Dynamic Files",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "315--344",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/bin-packing.bib; Misc/is.bib",
  note =         "Also published in/as: IBM Research Report RJ2305, Jul.
                 1978. See \cite{Regnier:1985:AGF}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-3/p315-fagin/p315-fagin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p315-fagin/",
  abstract =     "Extendible hashing is a new access technique, in which
                 the user is guaranteed no more than two page faults to
                 locate the data associated with a given unique
                 identifier, or key. Unlike conventional hashing,
                 extendible hashing has a dynamic structure that grows
                 and shrinks gracefully as the database grows and
                 shrinks. This approach simultaneously solves the
                 problem of making hash tables that are extendible and
                 of making radix search trees that are balanced. We
                 study, by analysis and simulation, the performance of
                 extendible hashing. The results indicate that
                 extendible hashing provides an attractive alternative
                 to other access methods, such as balanced trees.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  journalabr =   "ACM Trans Database Syst",
  keywords =     "access method; B-tree; data processing; directory;
                 extendible hashing; external hashing; file
                 organization; hashing; index; radix search; searching;
                 trie",
  remark =       "The user is guaranteed no more than two page faults to
                 locate the data associated with a given unique
                 identifier, or key. Extendible hashing has a dynamic
                 structure that grows and shrinks as the database grows
                 and shrinks.",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3): {\bf Search process}",
}

@Article{Lam:1979:PSH,
  author =       "Chat Yu Lam and Stuart E. Madnick",
  title =        "Properties of Storage Hierarchy Systems with Multiple
                 Page Sizes and Redundant Data",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "345--367",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p345-lam/",
  abstract =     "The need for high performance, highly reliable storage
                 for very large on-line databases, coupled with rapid
                 advances in storage device technology, has made the
                 study of generalized storage hierarchies an important
                 area of research.\par

                 This paper analyzes properties of a data storage
                 hierarchy system specifically designed for handling
                 very large on-line databases. To attain high
                 performance and high reliability, the data storage
                 hierarchy makes use of multiple page sizes in different
                 storage levels and maintains multiple copies of the
                 same information across the storage levels. Such a
                 storage hierarchy system is currently being designed as
                 part of the INFOPLEX database computer project.
                 Previous studies of storage hierarchies have primarily
                 focused on virtual memories for program storage and
                 hierarchies with a single page size across all storage
                 levels and/or a single copy of information in the
                 hierarchy.\par

                 In the INFOPLEX design, extensions to the least
                 recently used (LRU) algorithm are used to manage the
                 storage levels. The read-through technique is used to
                 initially load a referenced page of the appropriate
                 size into all storage levels above the one in which the
                 page is found. Since each storage level is viewed as an
                 extension of the immediate higher level, an overflow
                 page from level $i$ is always placed in level $ i + 1
                 $. Important properties of these algorithms are
                 derived. It is shown that depending on the types of
                 algorithms used and the relative sizes of the storage
                 levels, it is not always possible to guarantee that the
                 contents of a given storage level $i$ is always a
                 superset of the contents of its immediate higher
                 storage level $ i - 1 $. The necessary and sufficient
                 conditions for this property to hold are identified and
                 proved. Furthermore, it is possible that increasing the
                 size of intermediate storage levels may actually
                 increase the number of references to lower storage
                 levels, resulting in reduced performance. Conditions
                 necessary to avoid such an anomaly are also identified
                 and proved.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; data storage hierarchy; database
                 computer; inclusion properties; modeling; perform and
                 reliability analysis; storage management algorithms;
                 very large databases",
  subject =      "Information Systems --- Database Management (H.2);
                 Software --- Operating Systems --- Storage Management
                 (D.4.2): {\bf Storage hierarchies}",
}

@Article{Buneman:1979:EMR,
  author =       "O. Peter Buneman and Eric K. Clemons",
  title =        "Efficiently Monitoring Relational Databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "368--382",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Working paper, 76-10-08, Dep.
                 Decision Sciences, The Wharton School, Un. Penn, PA,
                 Jun. 1977.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-3/p368-buneman/p368-buneman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p368-buneman/",
  abstract =     "An alerter is a program which monitors a database and
                 reports to some user or program when a specified
                 condition occurs. It may be that the condition is a
                 complicated expression involving several entities in
                 the database; in this case the evaluation of the
                 expression may be computationally expensive. A scheme
                 is presented in which alerters may be placed on a
                 complex query involving a relational database, and a
                 method is demonstrated for reducing the amount of
                 computation involved in checking whether an alerter
                 should be triggered.",
  acknowledgement = ack-nhfb,
  annote =       "An alerter monitors a database and reports when a
                 specific condition occurs. Alerters may be placed on a
                 query, a method is demonstrated for reducing the amount
                 of computation involved in checking whether an alerter
                 should be triggered. Recomputation of derived data with
                 pruning, viz. identity connection.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "alerters; data base systems; exception reporting;
                 integrity constraints; programming techniques;
                 relational databases",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Comer:1979:HTI,
  author =       "Douglas Comer",
  title =        "Heuristics For Trie Index Minimization",
  journal =      j-TODS,
  volume =       "4",
  number =       "3",
  pages =        "383--395",
  month =        sep,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-3/p383-comer/p383-comer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-3/p383-comer/",
  abstract =     "A trie is a digital search tree in which leaves
                 correspond to records in a file. Searching proceeds
                 from the root to a leaf, where the edge taken at each
                 node depends on the value of an attribute in the query.
                 Trie implementations have the advantage of being fast,
                 but the disadvantage of achieving that speed at great
                 expense in storage space. Of primary concern in making
                 a trie practical, therefore, is the problem of
                 minimizing storage requirements. One method for
                 reducing the space required is to reorder attribute
                 testing. Unfortunately, the problem of finding an
                 ordering which guarantees a minimum-size trie is
                 NP-complete. In this paper we investigate several
                 heuristics for reordering attributes, and derive bounds
                 on the sizes of the worst tries produced by them in
                 terms of the underlying file. Although the analysis is
                 presented for a binary file, extensions to files of
                 higher degree are shown.\par

                 Another alternative for reducing the space required by
                 a trie is an implementation, called an $ \Omega $-trie,
                 in which the order of attribute testing is contained in
                 the trie itself. We show that for most applications, $
                 \Omega $-tries are smaller than other implementations
                 of tries, even when heuristics for improving storage
                 requirements are employed.",
  acknowledgement = ack-nhfb,
  annote =       "Of primary concern in making a trie practical is the
                 problem of minimizing storage requirements. One method
                 for reducing the space is attribute testing which is
                 NP-complete. Another alternative is an $ \Omega $-trie,
                 in which the order of attribute testing is contained in
                 the trie itself. $ \Omega $-tries are smaller than
                 other implementations of tries.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; doubly chained tree; index; trie;
                 trie minimization",
  subject =      "Data --- Data Structures (E.1); Information Systems
                 --- Database Management (H.2)",
}

@Article{Codd:1979:EDR,
  author =       "E. F. Codd",
  title =        "Extending the Database Relational Model to Capture
                 More Meaning",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "397--434",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Compendex database;
                 Compiler/prog.lang.theory.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Reprinted in
                 \cite[pp.~457--475]{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p397-codd/p397-codd.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p397-codd/",
  abstract =     "During the last three or four years several
                 investigators have been exploring ``semantic models''
                 for formatted databases. The intent is to capture (in a
                 more or less formal way) more of the meaning of the
                 data so that database design can become more systematic
                 and the database system itself can behave more
                 intelligently. Two major thrusts are clear.\par

                 (1) the search for meaningful units that are as small
                 as possible-- {\em atomic semantics\/};\par

                 (2) the search for meaningful units that are larger
                 than the usual $n$-ary relation-- {\em molecular
                 semantics}.\par

                 In this paper we propose extensions to the relational
                 model to support certain atomic and molecular
                 semantics. These extensions represent a synthesis of
                 many ideas from the published work in semantic modeling
                 plus the introduction of new rules for insertion,
                 update, and deletion, as well as new algebraic
                 operators.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8905-0330",
  annote =       "``Semantic models'' for formatted databases, to
                 capture in a more or less formal way more of the
                 meaning of the data. Two major thrusts: relation and
                 molecular semantics. Extensions to the relational model
                 (RM/T). New rules for insertion, update, and deletion,
                 as well as new algebraic operators (Theta-select, outer
                 join,\ldots{}.).",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual model; conceptual schema; data base
                 systems; data model; data semantics; database; database
                 schema; entity model; knowledge base; knowledge
                 representation; relation; relational database;
                 relational model; relational schema; semantic model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Information Systems
                 --- Database Management --- Logical Design (H.2.1):
                 {\bf Schema and subschema}",
}

@Article{Aho:1979:EOC,
  author =       "A. V. Aho and Y. Sagiv and J. D. Ullman",
  title =        "Efficient Optimization of a Class of Relational
                 Expressions",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "435--454",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p435-aho/p435-aho.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p435-aho/",
  abstract =     "The design of several database query languages has
                 been influenced by Codd's relational algebra. This
                 paper discusses the difficulty of optimizing queries
                 based on the relational algebra operations select,
                 project, and join. A matrix, called a tableau, is
                 proposed as a useful device for representing the value
                 of a query, and optimization of queries is couched in
                 terms of finding a minimal tableau equivalent to a
                 given one. Functional dependencies can be used to imply
                 additional equivalences among tableaux. Although the
                 optimization problem is NP-complete, a polynomial time
                 algorithm exists to optimize tableaux that correspond
                 to an important subclass of queries.",
  acknowledgement = ack-nhfb,
  annote =       "Optimizing queries based on select, project, and
                 join.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems, TODS tableaux optimization;
                 equivalence of queries; NP-completeness; query
                 optimization; relational algebra; relational database;
                 tableaux",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}",
}

@Article{Maier:1979:TID,
  author =       "David Maier and Alberto O. Mendelzon and Yehoshua
                 Sagiv",
  title =        "Testing Implications of Data Dependencies",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "455--469",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p455-maier/p455-maier.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p455-maier/",
  abstract =     "Presented is a computation method --- the {\em
                 chase\/} --- for testing implication of data
                 dependencies by a set of data dependencies. The chase
                 operates on tableaux similar to those of Aho, Sagiv,
                 and Ullman. The chase includes previous tableau
                 computation methods as special cases. By interpreting
                 tableaux alternately as mappings or as templates for
                 relations, it is possible to test implication of join
                 dependencies (including multivalued dependencies) and
                 functional dependencies by a set of dependencies.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "chase; data base systems; data dependencies;
                 functional dependencies; join dependencies; multivalued
                 dependencies; relational databases; tableaux",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Housel:1979:PTI,
  author =       "Barron C. Housel",
  title =        "Pipelining: a Technique for Implementing Data
                 Restructurers",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "470--492",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p470-housel/p470-housel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p470-housel/",
  abstract =     "In the past several years much attention has been
                 given to the problem of data translation. The focus has
                 been mainly on methodologies and specification
                 languages for accomplishing this task. Recently,
                 several prototype systems have emerged, and now the
                 issues of implementation and performance must be
                 addressed. In general, a data restructuring
                 specification may contain multiple source and target
                 files. This specification can be viewed as a ``process
                 graph'' which is a network of restructuring operations
                 subject to precedence constraints. One technique used
                 to achieve good performance is that of pipelining data
                 in the process graph.\par

                 In this paper we address a number of issues pertinent
                 to a pipelining architecture. Specifically, we give
                 algorithms for resolving deadlock situations which can
                 arise, and partitioning the process graph to achieve an
                 optimal schedule for executing the restructuring steps.
                 In addition, we discuss how pipelining has influenced
                 the design of the restructuring operations and the file
                 structures used in an actual system.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; data translation; database
                 conversion; deadlock; pipelining; process scheduling",
  subject =      "Information Systems --- Database Management ---
                 Heterogeneous Databases (H.2.5): {\bf Data
                 translation**}; Information Systems --- Database
                 Management --- Physical Design (H.2.2): {\bf Deadlock
                 avoidance}",
}

@Article{Shopiro:1979:TPL,
  author =       "Jonathan E. Shopiro",
  title =        "{Theseus} --- {A} Programming Language for Relational
                 Databases",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "493--517",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p493-shopiro/p493-shopiro.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p493-shopiro/",
  abstract =     "Theseus, a very high-level programming language
                 extending EUCLID, is described. Data objects in Theseus
                 include relations and a-sets, a generalization of
                 records. The primary design goals of Theseus are to
                 facilitate the writing of well-structured programs for
                 database applications and to serve as a vehicle for
                 research in automatic program optimization.",
  acknowledgement = ack-nhfb,
  annote =       "Extending EUCLID. Data objects in Theseus include
                 relations and a-sets",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compiler organization; computer programming languages;
                 data base systems; relational database languages; very
                 high-level languages",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Database (persistent)
                 programming languages}",
}

@Article{Yamamoto:1979:DBM,
  author =       "Sumiyasu Yamamoto and Shinsei Tazawa and Kazuhiko
                 Ushio and Hideto Ikeda",
  title =        "Design of a Balanced Multiple-Valued File-Organization
                 Scheme with the Least Redundancy",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "518--530",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p518-yamamoto/p518-yamamoto.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p518-yamamoto/",
  abstract =     "A new balanced file-organization scheme of order two
                 for multiple-valued records is presented. This scheme
                 is called HUBMFS 2 (Hiroshima University Balanced
                 Multiple-valued File-organization Scheme of order two).
                 It is assumed that records are characterized by $m$
                 attributes having $n$ possible values each, and the
                 query set consists of queries which specify values of
                 two attributes. It is shown that the redundancy of the
                 bucket (the probability of storing a record in the
                 bucket) is minimized if and only if the structure of
                 the bucket is a partite-claw. A necessary and
                 sufficient condition for the existence of an HUBMFS 2,
                 which is composed exclusively of partite-claw buckets,
                 is given. A construction algorithm is also given. The
                 proposed HUBMFS 2 is superior to existing BMFS 2
                 (Balanced Multiple-valued File-organization Schemes of
                 order two) in that it has the least redundancy among
                 all possible BMFS 2 's having the same parameters and
                 that it can be constructed for a less restrictive set
                 of parameters.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "balanced filing scheme; bucket; claw; data processing;
                 file organization; graph decomposition; information
                 retrieval; information storage; inverted file;
                 multipartite graph; multiple-valued attributes;
                 redundancy; secondary index",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Schema and
                 subschema}; Information Systems --- Information Storage
                 and Retrieval --- Information Storage (H.3.2)",
}

@Article{Batory:1979:STF,
  author =       "Don S. Batory",
  title =        "On Searching Transposed Files",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "531--544",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p531-batory/p531-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p531-batory/",
  abstract =     "A transposed file is a collection of nonsequential
                 files called subfiles. Each subfile contains selected
                 attribute data for all records. It is shown that
                 transposed file performance can be enhanced by using a
                 proper strategy to process queries. Analytic cost
                 expressions for processing conjunctive, disjunctive,
                 and batched queries are developed and an effective
                 heuristic for minimizing query processing costs is
                 presented. Formulations of the problem of optimally
                 processing queries for a particular family of
                 transposed files are shown to be NP-complete. Query
                 processing performance comparisons of multilist,
                 inverted, and nonsequential files with transposed files
                 are also considered.",
  acknowledgement = ack-nhfb,
  annote =       "Analytic cost expressions for processing conjunctive,
                 disjunctive, and batch queries are developed and an
                 effective heuristic for minimizing query processing
                 costs is presented.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; file searching; inverted file;
                 multilist; NP-complete; query processing; transposed
                 file",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3): {\bf Search process}",
}

@Article{Bernstein:1979:CPS,
  author =       "Philip A. Bernstein and Marco A. Casanova and Nathan
                 Goodman",
  title =        "Comments on {``Process Synchronization in Database
                 Systems''}",
  journal =      j-TODS,
  volume =       "4",
  number =       "4",
  pages =        "545--546",
  month =        dec,
  year =         "1979",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Schlageter:1978:PSD}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1979-4-4/p545-bernstein/p545-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1979-4-4/p545-bernstein/",
  acknowledgement = ack-nhfb,
  annote =       "The results of Schlageter are in error.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "Information Systems --- Database Management (H.2)",
}

@Article{Rothnie:1980:ISD,
  author =       "James B. {Rothnie, Jr.} and Philip A. Bernstein and S.
                 Fox and N. Goodman and M. Hammer and T. A. Landers and
                 C. Reeve and David W. Shipman and E. Wong",
  title =        "Introduction to a System for Distributed Databases
                 ({SDD-1})",
  journal =      j-TODS,
  volume =       "5",
  number =       "1",
  pages =        "1--17",
  month =        mar,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-1/p1-rothnie/p1-rothnie.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-1/p1-rothnie/",
  abstract =     "The declining cost of computer hardware and the
                 increasing data processing needs of geographically
                 dispersed organizations have led to substantial
                 interest in distributed data management. SDD-1 is a
                 distributed database management system currently being
                 developed by Computer Corporation of America. Users
                 interact with SDD-1 precisely as if it were a
                 nondistributed database system because SDD-1 handles
                 all issues arising from the distribution of data. These
                 issues include distributed concurrency control,
                 distributed query processing, resiliency to component
                 failure, and distributed directory management. This
                 paper presents an overview of the SDD-1 design and its
                 solutions to the above problems.\par

                 This paper is the first of a series of companion papers
                 on SDD-1 (Bernstein and Shipman [2], Bernstein et al.
                 [4], and Hammer and Shipman [14]).",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; data base systems; database
                 reliability; distributed database system; query
                 processing; relational data model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Relational databases}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Concurrency}; Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Query
                 processing}",
}

@Article{Bernstein:1980:CCS,
  author =       "Philip A. Bernstein and David W. Shipman and James B.
                 {Rothnie, Jr.}",
  title =        "Concurrency Control in a System for Distributed
                 Databases ({SDD-1})",
  journal =      j-TODS,
  volume =       "5",
  number =       "1",
  pages =        "18--51",
  month =        mar,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-1/p18-bernstein/p18-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-1/p18-bernstein/",
  abstract =     "This paper presents the concurrency control strategy
                 of SDD-1. SDD-1, a System for Distributed Databases, is
                 a prototype distributed database system being developed
                 by Computer Corporation of America. In SDD-1, portions
                 of data distributed throughout a network may be
                 replicated at multiple sites. The SDD-1 concurrency
                 control guarantees database consistency in the face of
                 such distribution and replication.\par

                 This paper is one of a series of companion papers on
                 SDD-1 [4, 10, 12, 21].",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; conflict graph; data base
                 systems; distributed database system; serializability;
                 synchronization; timestamps",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Distributed databases}",
}

@Article{Bernstein:1980:CCC,
  author =       "Philip A. Bernstein and David W. Shipman",
  title =        "Correctness of Concurrency Control Mechanisms in a
                 System for Distributed Databases ({SDD-1})",
  journal =      j-TODS,
  volume =       "5",
  number =       "1",
  pages =        "52--68",
  month =        mar,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-1/p52-bernstein/p52-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-1/p52-bernstein/",
  abstract =     "This paper presents a formal analysis of the
                 concurrency control strategy of SDD-1. SDD-1, a System
                 for Distributed Databases, is a prototype distributed
                 database system being developed by Computer Corporation
                 of America. In SDD-1, portions of data distributed
                 throughout a network may be replicated at multiple
                 sites. The SDD-1 concurrency control guarantees
                 database consistency in the face of such distribution
                 and replication.\par

                 This paper is one of a series of companion papers on
                 SDD-1 [2, 8].",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conflict graph; correctness of concurrency control;
                 data base systems; distributed database system;
                 serializability theory",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Concurrency}",
}

@Article{Gopalakrishna:1980:PEA,
  author =       "V. Gopalakrishna and C. E. {Veni Madhavan}",
  title =        "Performance Evaluation of Attribute-Based Tree
                 Organization",
  journal =      j-TODS,
  volume =       "5",
  number =       "1",
  pages =        "69--87",
  month =        mar,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-1/p69-gopalakrishna/p69-gopalakrishna.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-1/p69-gopalakrishna/",
  abstract =     "A modified version of the multiple attribute tree
                 (MAT) database organization, which uses a compact
                 directory, is discussed. An efficient algorithm to
                 process the directory for carrying out the node
                 searches is presented. Statistical procedures are
                 developed to estimate the number of nodes searched and
                 the number of data blocks retrieved for most general
                 and complex queries. The performance of inverted file
                 and modified MAT organizations are compared using six
                 real-life databases and four types of query
                 complexities. Careful tradeoffs are established in
                 terms of storage and access times for directory and
                 data, query complexities, and database
                 characteristics.",
  acknowledgement = ack-nhfb,
  annote =       "A version of the multiple attribute tree (MAT)
                 database organization. Statistical procedures are
                 developed to estimate the number of nodes searched and
                 the number of data blocks retrieved. The performance of
                 inverted file and modified MAT organizations are
                 compared.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access time; average retrieval time per query; data
                 base systems; database organization; database
                 performance; directory search time; modified multiple
                 attribute tree; query complexity; secondary index
                 organization",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3)",
}

@Article{Denning:1980:FPF,
  author =       "Dorothy E. Denning and Jan Schl{\"o}rer",
  title =        "Fast Procedure for Finding a Tracker in a Statistical
                 Database",
  journal =      j-TODS,
  volume =       "5",
  number =       "1",
  pages =        "88--102",
  month =        mar,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-1/p88-denning/p88-denning.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-1/p88-denning/",
  abstract =     "To avoid trivial compromises, most on-line statistical
                 databases refuse to answer queries for statistics about
                 small subgroups. Previous research discovered a
                 powerful snooping tool, the tracker, with which the
                 answers to these unanswerable queries are easily
                 calculated. However, the extent of this threat was not
                 clear, for no one had shown that finding a tracker is
                 guaranteed to be easy.\par

                 This paper gives a simple algorithm for finding a
                 tracker when the maximum number of identical records is
                 not too large. The number of queries required to find a
                 tracker is at most {$ O(\log_2 S) $} queries, where
                 {$S$} is the number of distinct records possible.
                 Experimental results show that the procedure often
                 finds a tracker with just a few queries. The threat
                 posed by trackers is therefore considerable.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; data base systems; data security;
                 database security; statistical database; tracker",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Menasce:1980:LPR,
  author =       "Daniel A. Menasc{\'e} and Gerald J. Popek and Richard
                 R. Muntz",
  title =        "A Locking Protocol for Resource Coordination in
                 Distributed Databases",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "103--138",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p103-menasce/p103-menasce.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p103-menasce/",
  abstract =     "A locking protocol to coordinate access to a
                 distributed database and to maintain system consistency
                 throughout normal and abnormal conditions is presented.
                 The proposed protocol is robust in the face of crashes
                 of any participating site, as well as communication
                 failures. Recovery from any number of failures during
                 normal operation or any of the recovery stages is
                 supported. Recovery is done in such a way that maximum
                 forward progress is achieved by the recovery
                 procedures. Integration of virtually any locking
                 discipline including predicate lock methods is
                 permitted by this protocol. The locking algorithm
                 operates, and operates correctly, when the network is
                 partitioned, either intentionally or by failure of
                 communication lines. Each partition is able to continue
                 with work local to it, and operation merges gracefully
                 when the partitions are reconnected.\par

                 A subroutine of the protocol, that assures reliable
                 communication among sites, is shown to have better
                 performance than two-phase commit methods. For many
                 topologies of interest, the delay introduced by the
                 overall protocol is not a direct function of the size
                 of the network. The communications cost is shown to
                 grow in a relatively slow, linear fashion with the
                 number of sites participating in the transaction. An
                 informal proof of the correctness of the algorithm is
                 also presented in this paper.\par

                 The algorithm has as its core a centralized locking
                 protocol with distributed recovery procedures. A
                 centralized controller with local appendages at each
                 site coordinates all resource control, with requests
                 initiated by application programs at any site. However,
                 no site experiences undue load. Recovery is broken down
                 into three disjoint mechanisms: for single node
                 recovery, merge of partitions, and reconstruction of
                 the centralized controller and tables. The disjointness
                 of the mechanisms contributes to comprehensibility and
                 ease of proof.\par

                 The paper concludes with a proposal for an extension
                 aimed at optimizing operation of the algorithm to adapt
                 to highly skewed distributions of activity. The
                 extension applies nicely to interconnected computer
                 networks.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; consistency; crash recovery; distributed
                 databases; locking protocol",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Distributed databases}",
}

@Article{Bayer:1980:PRD,
  author =       "R. Bayer and H. Heller and A. Reiser",
  title =        "Parallelism and Recovery in Database Systems",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "139--156",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p139-bayer/p139-bayer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p139-bayer/",
  abstract =     "In this paper a new method to increase parallelism in
                 database systems is described. Use is made of the fact
                 that for recovery reasons, we often have two values for
                 one object in the database--the new one and the old
                 one. Introduced and discussed in detail is a certain
                 scheme by which readers and writers may work
                 simultaneously on the same object. It is proved that
                 transactions executed according to this scheme have the
                 correct effect; i.e., consistency is preserved. Several
                 variations of the basic scheme which are suitable
                 depending on the degree of parallelism required, are
                 described.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; consistency; data base systems; deadlock;
                 integrity; recovery; synchronization; transaction; two
                 phase locking",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart};
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Deadlock avoidance};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Concurrency}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Transaction processing}",
}

@Article{Navathe:1980:SAD,
  author =       "Shamkant B. Navathe",
  title =        "Schema Analysis for Database Restructuring",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "157--184",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Proceedings of the Third
                 Conference on Very Large Databases, Morgan Kaufman
                 pubs. (Los Altos CA), 1977.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p157-navathe/p157-navathe.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p157-navathe/",
  abstract =     "The problem of generalized restructuring of databases
                 has been addressed with two limitations: first, it is
                 assumed that the restructuring user is able to describe
                 the source and target databases in terms of the
                 implicit data model of a particular methodology;
                 second, the restructuring user is faced with the task
                 of judging the scope and applicability of the defined
                 types of restructuring to his database implementation
                 and then of actually specifying his restructuring needs
                 by translating them into the restructuring operations
                 on a foreign data model. A certain amount of analysis
                 of the logical and physical structure of databases must
                 be performed, and the basic ingredients for such an
                 analysis are developed here. The distinction between
                 hierarchical and nonhierarchical data relationships is
                 discussed, and a classification for database schemata
                 is proposed. Examples are given to illustrate how these
                 schemata arise in the conventional hierarchical and
                 network systems. Application of the schema analysis
                 methodology to restructuring specification is also
                 discussed. An example is presented to illustrate the
                 different implications of restructuring three seemingly
                 identical database structures.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; data model; data relationships;
                 data semantics; data structure; database; database
                 design; database management systems; database
                 restructuring; graphical representation of data;
                 schema; stored data",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Schema and subschema}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Data models}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Mylopoulos:1980:LFD,
  author =       "John Mylopoulos and Philip A. Bernstein and Harry K.
                 T. Wong",
  title =        "A Language Facility for Designing Database-Intensive
                 Applications",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "185--207",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib; Object/Nierstrasz.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p185-mylopoulos/p185-mylopoulos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p185-mylopoulos/",
  abstract =     "TAXIS, a language for the design of interactive
                 information systems (e.g., credit card verification,
                 student-course registration, and airline reservations)
                 is described. TAXIS offers (relational) database
                 management facilities, a means of specifying semantic
                 integrity constraints, and an exception-handling
                 mechanism, integrated into a single language through
                 the concepts of {\em class, property}, and the {\em
                 IS-A\/} (generalization) {\em relationship}. A
                 description of the main constructs of TAXIS is included
                 and their usefulness illustrated with examples.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data type; applications programming;
                 exception handling; information system; relational data
                 model; semantic network",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Information Systems
                 --- Database Management --- Languages (H.2.3)",
}

@Article{Lozinskii:1980:CRR,
  author =       "Eliezer L. Lozinskii",
  title =        "Construction of Relations in Relational Databases",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "208--224",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p208-lozinskii/p208-lozinskii.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p208-lozinskii/",
  abstract =     "Using a nonprocedural language for query formulation
                 requires certain automatization of a query answering
                 process. Given a query for creation of a new relation,
                 the problem is to find an efficient procedure which
                 produces this relation from a given relational
                 database. The author concentrates upon sequences of
                 join operations which losslessly produce a relation
                 required by a query. A new property of such sequences
                 is analyzed which provides a basis for the presented
                 algorithms that construct an efficient join procedure.
                 The algorithms have polynomial complexity. A modified
                 AND\slash OR graph is used for the display of a given
                 set of dependencies and a collection of relations
                 representing a database.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; AND/OR graphs; data base systems; lossless
                 joins; query answering; relational databases",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Stonebraker:1980:RDS,
  author =       "Michael Stonebraker",
  title =        "Retrospection on a Database System",
  journal =      j-TODS,
  volume =       "5",
  number =       "2",
  pages =        "225--240",
  month =        jun,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-2/p225-stonebraker/p225-stonebraker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-2/p225-stonebraker/",
  abstract =     "This paper describes the implementation history of the
                 INGRES database system. It focuses on mistakes that
                 were made in progress rather than on eventual
                 corrections. Some attention is also given to the role
                 of structured design in a database system
                 implementation and to the problem of supporting
                 nontrivial users. Lastly, miscellaneous impressions of
                 UNIX, the PDP-11, and data models are given.",
  acknowledgement = ack-nhfb,
  annote =       "The implementation history of the INGRES database
                 system. The role of structured design in a database
                 system implementation, impressions of UNIX, the PDP-11,
                 and data models are given.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency; data base systems, history evaluation,
                 Ingres, TODS; integrity; nonprocedural languages;
                 protection; recovery; relational databases",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Relational
                 databases}; Information Systems --- Database Management
                 --- Physical Design (H.2.2): {\bf Recovery and
                 restart}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Concurrency}",
}

@Article{Beeri:1980:MPF,
  author =       "Catriel Beeri",
  title =        "On the Membership Problem for Functional and
                 Multivalued Dependencies in Relational Databases",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "241--259",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p241-beeri/p241-beeri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p241-beeri/",
  abstract =     "The problem of whether a given dependency in a
                 database relation can be derived from a given set of
                 dependencies is investigated. We show that the problem
                 can be decided in polynomial time when the given set
                 consists of either multivalued dependencies only or of
                 both functional and multivalued dependencies and the
                 given dependency is also either a functional or a
                 multivalued dependency. These results hold when the
                 derivations are restricted not to use the
                 complementation rule.",
  acknowledgement = ack-nhfb,
  annote =       "The problem of whether a given dependency in a
                 database relation can be derived from a given set of
                 dependencies is investigated. We show that the problem
                 can be decided in polynomial time when the given set
                 consists of either multivalued dependencies only or of
                 both functional and multivalued dependencies and the
                 given dependency is also either a functional or a
                 multivalued dependency. These results hold when the
                 derivations are restricted not to use the
                 complementation rule.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; functional dependency; inference
                 rule; membership; multivalued dependency; relations",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Klug:1980:CCR,
  author =       "A. Klug",
  title =        "Calculating Constraints on Relational Expressions",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "260--290",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p260-klug/p260-klug.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p260-klug/",
  abstract =     "This paper deals with the problem of determining which
                 of a certain class of constraints hold on a given
                 relational algebra expression where the base relations
                 come from a given schema. The class of constraints
                 includes functional dependencies, equality of domains,
                 and constancy of domains. The relational algebra
                 consists of projection, selection, restriction, cross
                 product, union, and difference. The problem as given is
                 undecidable, but if set difference is removed from the
                 algebra, there is a solution. Operators specifying a
                 closure function (similar to functional dependency
                 closure on one relation) are defined; these will
                 generate exactly the set of constraints valid on the
                 given relational algebra expression. We prove that the
                 operators are sound and complete.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "completeness; constraints; data base systems;
                 derivation rules; functional dependencies; Views",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Denning:1980:SSD,
  author =       "Dorothy E. Denning",
  title =        "Secure Statistical Databases with Random Sample
                 Queries",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "291--315",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p291-denning/p291-denning.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p291-denning/",
  abstract =     "A new inference control, called random sample queries,
                 is proposed for safeguarding confidential data in
                 on-line statistical databases. The random sample
                 queries control deals directly with the basic principle
                 of compromise by making it impossible for a questioner
                 to control precisely the formation of query sets.
                 Queries for relative frequencies and averages are
                 computed using random samples drawn from the query
                 sets. The sampling strategy permits the release of
                 accurate and timely statistics and can be implemented
                 at very low cost. Analysis shows the relative error in
                 the statistics decreases as the query set size
                 increases; in contrast, the effort required to
                 compromise increases with the query set size due to
                 large absolute errors. Experiments performed on a
                 simulated database support the analysis.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; data base systems; database security;
                 disclosure controls; sampling; statistical database",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Query processing};
                 Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Beck:1980:SMS,
  author =       "Leland L. Beck",
  title =        "A security mechanism for statistical database",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "316--338",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p316-beck/p316-beck.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p316-beck/",
  abstract =     "The problem of user inference in statistical databases
                 is discussed and illustrated with several examples. It
                 is assumed that the database allows ``total'',
                 ``average'', ``count'', and ``percentile'' queries; a
                 query may refer to any arbitrary subset of the
                 database. Methods for protecting the security of such a
                 database are considered; it is shown that any scheme
                 which gives ``statistically correct'' answers is
                 vulnerable to penetration. A precise definition of
                 compromisability (in a statistical sense) is given. A
                 general model of user inference is proposed; two
                 special cases of this model appear to contain all
                 previously published strategies for compromising a
                 statistical database. A method for protecting the
                 security of such a statistical database against these
                 types of user inference is presented and discussed. It
                 is shown that the number of queries required to
                 compromise the database can be made arbitrarily large
                 by accepting moderate increases in the variance of
                 responses to queries. A numerical example is presented
                 to illustrate the application of the techniques
                 discussed.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromisability; data base systems; data security;
                 database inference; privacy protection; statistical
                 databases; statistical queries",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0): {\bf Security, integrity, and
                 protection**}; Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Query processing};
                 Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}",
}

@Article{Lee:1980:QTF,
  author =       "D. T. Lee and C. K. Wong",
  title =        "Quintary Trees: a File Structure for Multidimensional
                 Database Systems",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "339--353",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Graphics/siggraph/80.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p339-lee/p339-lee.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p339-lee/",
  abstract =     "A file structure is presented that was designed for a
                 database system in which four types of retrieval
                 requests (queries) are allowed: exact match, partial
                 match, range, and partial range queries. Outlines are
                 sketched for inserting and deleting records that
                 require O(k plus (log N)**k) time, on the average. This
                 structure achieves faster response time than previously
                 known structures (for many of the queries) at the cost
                 of extra storage.",
  acknowledgement = ack-nhfb,
  annote =       "Four types of retrieval (queries) are allowed: exact
                 match, partial match, range, and partial range queries.
                 Faster response time at the cost of extra storage.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; data processing --- data
                 structures; database system; exact match queries; file
                 maintenance; information retrieval; key;
                 multidimensional space; queries; range search; search",
  subject =      "Data --- Data Structures (E.1): {\bf Trees};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}; Information Systems
                 --- Information Storage and Retrieval --- Information
                 Storage (H.3.2): {\bf File organization}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3)",
}

@Article{Kung:1980:CMB,
  author =       "H. T. Kung and Philip L. Lehman",
  title =        "Concurrent Manipulation of Binary Search Trees",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "354--382",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-3/p354-kung/p354-kung.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-3/p354-kung/",
  abstract =     "The concurrent manipulation of a binary search tree is
                 considered in this paper. The systems presented can
                 support any number of concurrent processes which
                 perform searching, insertion, deletion, and rotation
                 (reorganization) on the tree, but allow any process to
                 lock only a constant number of nodes at any time. Also,
                 in the systems, searches are essentially never blocked.
                 The concurrency control techniques introduced in the
                 paper include the use of special nodes and pointers to
                 redirect searches, and the use of copies of sections of
                 the tree to introduce many changes simultaneously and
                 therefore avoid unpredictable interleaving. Methods
                 developed in this paper may provide new insights into
                 other problems in the area of concurrent database
                 manipulation.",
  acknowledgement = ack-nhfb,
  annote =       "Operations on tries are defined so that concurrency of
                 access is possible while the number of locked nodes is
                 minimal.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "binary search trees; concurrency controls; concurrent
                 algorithm; consistency; correctness; data processing;
                 data structures; databases; locking protocols",
  subject =      "Data --- Data Structures (E.1): {\bf Trees};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Concurrency}",
}

@Article{Denning:1980:CLQ,
  author =       "D. E. Denning",
  title =        "Corrigenda on Linear Queries in Statistical
                 Databases",
  journal =      j-TODS,
  volume =       "5",
  number =       "3",
  pages =        "383--383",
  month =        sep,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  annote =       "refers to Schwartz 1979 TODS.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hsiao:1980:TFT,
  author =       "David K. Hsiao",
  title =        "{TODS} --- the first three years {(1976\&ndash1978)}",
  journal =      j-TODS,
  volume =       "5",
  number =       "4",
  pages =        "385--403",
  month =        dec,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-4/p385-hsiao/p385-hsiao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-4/p385-hsiao/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "General Literature --- General (A.0)",
}

@Article{Armstrong:1980:DFD,
  author =       "W. W. Armstrong and C. Delobel",
  title =        "Decompositions and Functional Dependencies in
                 Relations",
  journal =      j-TODS,
  volume =       "5",
  number =       "4",
  pages =        "404--430",
  month =        dec,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-4/p404-armstrong/p404-armstrong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-4/p404-armstrong/",
  abstract =     "A general study is made of two basic integrity
                 constraints on relations: functional and multivalued
                 dependencies. The latter are studied via an equivalent
                 concept: decompositions. A model is constructed for any
                 possible combination of functional dependencies and
                 decompositions. The model embodies some decompositions
                 as unions of relations having different schemata of
                 functional dependencies. This suggests a new, stronger
                 integrity constraint, the degenerate decomposition.
                 More generally, the theory demonstrates the importance
                 of using the union operation in database design and of
                 allowing different schemata on the operands of a union.
                 Techniques based on the union lead to a method for
                 solving the problem of membership of a decomposition in
                 the closure of a given set of functional dependencies
                 and decompositions. The concept of antiroot is
                 introduced as a tool for describing families of
                 decompositions, and its fundamental importance for
                 database design is indicated.",
  acknowledgement = ack-nhfb,
  annote =       "A general study is made of two basic integrity
                 constrains, functional and multivalued dependencies,
                 via an equivalent concept: decompositions.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; decomposition; functional
                 dependency; integrity constraint; multivalued
                 dependency; relational database",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases}",
}

@Article{Hammer:1980:RMS,
  author =       "Michael Hammer and David Shipman",
  title =        "Reliability Mechanisms for {SDD-1}: a System for
                 Distributed Databases",
  journal =      j-TODS,
  volume =       "5",
  number =       "4",
  pages =        "431--466",
  month =        dec,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-4/p431-hammer/p431-hammer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-4/p431-hammer/",
  abstract =     "This paper presents the reliability mechanisms of
                 SDD-1, a prototype distributed database system being
                 developed by the Computer Corporation of America.
                 Reliability algorithms in SDD-1 center around the
                 concept of the Reliable Network (RelNet). The RelNet is
                 a communications medium incorporating facilities for
                 site status monitoring, event timestamping, multiply
                 buffered message delivery, and the atomic control of
                 distributed transactions.\par

                 This paper is one of a series of companion papers on
                 SDD-1 [3, 4, 6, 13].",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "atomicity; data base systems; distributed databases;
                 recovery; reliability",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart}",
}

@Article{Schloer:1980:SSD,
  author =       "Jan Schl{\"o}er",
  title =        "Security of statistical databases: multidimensional
                 transformation",
  journal =      j-TODS,
  volume =       "5",
  number =       "4",
  pages =        "467--492",
  month =        dec,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1980-5-4/p467-schler/",
  abstract =     "Statistical evaluation of databases which contain
                 personal records may entail risks for the
                 confidentiality of the individual records. The risk has
                 increased with the availability of flexible interactive
                 evaluation programs which permit the use of trackers,
                 the most dangerous class of snooping tools known. A
                 class of trackers, called union trackers, is described.
                 They permit reconstruction of the entire database
                 without supplementary knowledge and include the general
                 tracker recently described as a special case. For many
                 real statistical databases the overwhelming majority of
                 definable sets of records will form trackers. For such
                 databases a random search for a tracker is likely to
                 succeed rapidly. Individual trackers are redefined and
                 counted and their cardinalities are investigated. If
                 there are $n$ records in the database, then most
                 individual trackers employ innocent cardinalities near
                 $ n / 3 $, making them difficult to detect. Disclosure
                 with trackers usually requires little effort per
                 retrieved data element.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; database security; security;
                 statistical database; tracker",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Database Management
                 --- Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
  xxtitle =      "Disclosure from Statistical Databases: Quantitative
                 Aspects of Trackers",
}

@Article{Herot:1980:SMD,
  author =       "Christopher F. Herot",
  title =        "Spatial Management of Data",
  journal =      j-TODS,
  volume =       "5",
  number =       "4",
  pages =        "493--513",
  month =        dec,
  year =         "1980",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 Graphics/imager/imager.80.bib;
                 Graphics/siggraph/80.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1980-5-4/p493-herot/p493-herot.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1980-5-4/p493-herot/",
  abstract =     "Spatial data management is a technique for organizing
                 and retrieving information by positioning it in a
                 graphical data space (GDS). This graphical data space
                 is viewed through a color raster-scan display which
                 enables users to traverse the GDS surface or zoom into
                 the image to obtain greater detail. In contrast to
                 conventional database management systems, in which
                 users access data by asking questions in a formal query
                 language, a spatial data management system (SDMS)
                 presents the information graphically in a form that
                 seems to encourage browsing and to require less prior
                 knowledge of the contents and organization of the
                 database.\par

                 This paper presents an overview of the SDMS concept and
                 describes its implementation in a prototype system for
                 retrieving information from both a symbolic database
                 management system and an optical videodisk.",
  acknowledgement = ack-nhfb,
  annote =       "Organizing and retrieving information by positioning
                 it in a graphical data space viewed through a color
                 display. An overview of the SDMS concept and describes
                 its implementation in a prototype system.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer graphics; database query languages;
                 graphical/programming language, query language,
                 Man-Machine Communications interaction, data base
                 systems; graphics languages; man-machine interaction",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Spatial databases
                 and GIS}; Information Systems --- Database Management
                 --- Languages (H.2.3): {\bf Query languages};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}",
}

@Article{Zaniolo:1981:DRD,
  author =       "Carlo Zaniolo and Michel A. Melkanoff",
  title =        "On the Design of Relational Database Schemata",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "1--47",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15 (68H05)",
  MRnumber =     "82b:68019",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p1-zaniolo/p1-zaniolo.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p1-zaniolo/",
  abstract =     "The purpose of this paper is to present a new approach
                 to the conceptual design of relational databases based
                 on the complete relatability conditions (CRCs).\par

                 It is shown that current database design methodology
                 based upon the elimination of anomalies is not
                 adequate. In contradistinction, the CRCs are shown to
                 provide a powerful criticism for decomposition. A
                 decomposition algorithm is presented which (1) permits
                 decomposition of complex relations into simple,
                 well-defined primitives, (2) preserves all the original
                 information, and (3) minimizes redundancy.\par

                 The paper gives a complete derivation of the CRCs,
                 beginning with a unified treatment of functional and
                 multivalued dependencies, and introduces the concept of
                 elementary functional dependencies and multiple
                 elementary multivalued dependencies. Admissibility of
                 covers and validation of results are also discussed,
                 and it is shown how these concepts may be used to
                 improve the design of 3NF schemata. Finally, a
                 convenient graphical representation is proposed, and
                 several examples are described in detail to illustrate
                 the method.",
  acknowledgement = ack-nhfb,
  annote =       "The conceptual design of relational databases based on
                 the complete reliability conditions (CRCs). A unified
                 treatment of functional and multivalued dependencies.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; decomposition; functional
                 dependencies; minimal covers; multivalued dependencies;
                 relational databases; schema design",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Schema and subschema}",
}

@Article{Lien:1981:HSR,
  author =       "Y. Edmund Lien",
  title =        "Hierarchical Schemata for Relational Databases",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "48--69",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15 (68H05)",
  MRnumber =     "82b:68015",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p48-lien/p48-lien.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p48-lien/",
  abstract =     "Most database design methods for the relational model
                 produce a flat database, that is, a family of relations
                 with no explicit interrelational connections. The user
                 of a flat database is likely to be unaware of certain
                 interrelational semantics. In contrast, the
                 entity-relationship model provides schema graphs as a
                 description of the database, as well as for navigating
                 the database. Nevertheless, the user of an
                 entity-relationship database may still commit semantic
                 errors, such as performing a lossy join. This paper
                 proposes a nonflat, or hierarchical, view of relational
                 databases. Relations are grouped together to form {\em
                 relation hierarchies\/} in which lossless joins are
                 explicitly shown whereas lossy joins are excluded.
                 Relation hierarchies resemble the schema graphs in the
                 entity-relationship model.\par

                 An approach to the design of relation hierarchies is
                 outlined in the context of data dependencies and
                 relational decomposition. The approach consists of two
                 steps; each is described as an algorithm. Algorithm DEC
                 decomposes a given universal relation according to a
                 given set of data dependencies and produces a set of
                 nondecomposable relation schemes. This algorithm
                 differs from its predecessors in that it produces no
                 redundant relation schemes. Algorithm RH further
                 structures the relation schemes produced by Algorithm
                 DEC into a hierarchical schema. These algorithms can be
                 useful software tools for database designers.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database design; lossless join;
                 multivalued dependency; relation normalization",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Schema and subschema}",
}

@Article{Chamberlin:1981:SRT,
  author =       "D. D. Chamberlin and M. M. Astrahan and W. F. King and
                 R. A. Lorie and J. W. Mehl and T. G. Price and M.
                 Schkolnick and P. Griffiths Selinger and D. R. Slutz
                 and B. W. Wade and R. A. Yost",
  title =        "Support for Repetitive Transactions and Ad Hoc Queries
                 in {System R}",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "70--94",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: IBM Research Report
                 RJ2551(33151), May. 1979.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p70-chamberlin/p70-chamberlin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p70-chamberlin/",
  abstract =     "System R supports a high-level relational user
                 language called SQL which may be used by ad hoc users
                 at terminals or as an embedded data sublanguage in PL/I
                 or COBOL. Host-language programs with embedded SQL
                 statements are processed by the System R precompiler
                 which replaces the SQL statements by calls to a
                 machine-language access module. The precompilation
                 approach removes much of the work of parsing, name
                 binding, and access path selection from the path of a
                 running program, enabling highly efficient support for
                 repetitive transactions. Ad hoc queries are processed
                 by a similar approach of name binding and access path
                 selection which takes place on-line when the query is
                 specified. By providing a flexible spectrum of binding
                 times, System R permits transaction-oriented programs
                 and ad hoc query users to share a database without loss
                 of efficiency.\par

                 System R is an experimental database management system
                 designed and built by members of the IBM San Jose
                 Research Laboratory as part of a research program on
                 the relational model of data. This paper describes the
                 architecture of System R, and gives some preliminary
                 measurements of system performance in both the ad hoc
                 query and the ``canned program'' environments.",
  acknowledgement = ack-nhfb,
  annote =       "Embedded SQL statements are processed by the System R
                 precompiler enabling highly efficient support for
                 repetitive transactions. Ad hoc query is specified. By
                 providing a flexible spectrum of binding times. System
                 R permits transaction-oriented programs and ad hoc
                 query users to share a database without loss of
                 efficiency.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compilation; data base systems, TODS ad-hoc relation
                 database IBM San Jose; performance measurements; query
                 languages; relational database systems; transaction
                 processing",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf System R}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Schlorer:1981:SSD,
  author =       "Jan Schl{\"o}rer",
  title =        "Security of Statistical Databases: Multidimensional
                 Transformation",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "95--112",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15",
  MRnumber =     "82b:68018",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p95-schlorer/p95-schlorer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p95-schlorer/",
  abstract =     "The concept of multidimensional transformation of
                 statistical databases is described. A given set of
                 statistical output may be compatible with more than one
                 statistical database. A transformed database $ D' $ is
                 a database which (1) differs from the original database
                 $D$ in its record content, for (2) produces, within
                 certain limits, the same statistical output as the
                 original database. For a transformable database $D$
                 there are two options: One may physically transform $D$
                 into a suitable database $ D' $, or one may release
                 only that output which will not permit the users to
                 decide whether it comes from $D$ or $ D' $. The second
                 way is, of course, the easier one. Basic structural
                 requirements for transformable statistical databases
                 are investigated. Advantages, drawbacks, and open
                 questions are discussed.",
  acknowledgement = ack-nhfb,
  annote =       "A transformed database differs from the original
                 database in its record content but produces within
                 certain limits the same statistical output.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "confidentiality; data base systems; data processing
                 --- security of data; database; database security;
                 matrices; security; statistical database",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}",
}

@Article{Chin:1981:SDD,
  author =       "Francis Y. Chin and Gultekin {\"O}zsoyo{\u{g}}lu",
  title =        "Statistical Database Design",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "113--139",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p113-chin/p113-chin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p113-chin/",
  abstract =     "The security problem of a statistical database is to
                 limit the use of the database so that no sequence of
                 statistical queries is sufficient to deduce
                 confidential or private information. In this paper it
                 is suggested that the problem be investigated at the
                 conceptual data model level. The design of a
                 statistical database should utilize a statistical
                 security management facility to enforce the security
                 constraints at the conceptual model level. Information
                 revealed to users is well defined in the sense that it
                 can at most be reduced to nondecomposable information
                 involving a group of individuals. In addition, the
                 design also takes into consideration means of storing
                 the query information for auditing purposes, changes in
                 the database, users' knowledge, and some security
                 measures.",
  acknowledgement = ack-nhfb,
  annote =       "Limit the use of the database so that no sequence of
                 statistical queries is sufficient to deduce
                 confidential information at the conceptual data model
                 level.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromisability; conceptual databases model; data
                 base systems; data processing --- security of data;
                 database design; protection; security; statistical
                 database",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}",
}

@Article{Shipman:1981:FDM,
  author =       "David W. Shipman",
  title =        "The Functional Data Model and the Data Language
                 {DAPLEX}",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "140--173",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/bibdb.bib;
                 Database/Graefe.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/Functional.bib; Misc/is.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p140-shipman/p140-shipman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p140-shipman/",
  abstract =     "DAPLEX is a database language which incorporates:
                 \par

                 a formulation of data in terms of entities;\par

                 a functional representation for both actual and virtual
                 data relationships;\par

                 a rich collection of language constructs for expressing
                 entity selection criteria;\par

                 a notion of subtype/supertype relationships among
                 entity types.\par

                 This paper presents and motivates the DAPLEX language
                 and the underlying data model on which it is based.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data base systems;
                 database; functional data model; language",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf DAPLEX}; Information Systems
                 --- Database Management --- Languages (H.2.3)",
}

@Article{Rosenberg:1981:TSO,
  author =       "Arnold L. Rosenberg and Lawrence Snyder",
  title =        "Time- and Space-Optimality in {B-Trees}",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "174--193",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15 (68E10)",
  MRnumber =     "82m:68048",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p174-rosenberg/p174-rosenberg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p174-rosenberg/",
  abstract =     "A B-tree is {\em compact\/} if it is minimal in number
                 of nodes, hence has optimal space utilization, among
                 equally capacious B-trees of the same order. The space
                 utilization of compact B-trees is analyzed and compared
                 with that of noncompact B-trees and with
                 (node)-visit-optimal B-trees, which minimize the
                 expected number of nodes visited per key access.
                 Compact B-trees can be as much as a {\em factor\/} of
                 2.5 more space efficient than visit-optimal B-trees;
                 and the node-visit cost of a compact tree is never more
                 than 1 + the node-visit cost of an optimal tree. The
                 utility of initializing a B-tree to be compact (which
                 initialization can be done in time linear in the number
                 of keys if the keys are presorted) is demonstrated by
                 comparing the space utilization of a compact tree that
                 has been augmented by random insertions with that of a
                 tree that has been grown entirely by random insertions.
                 Even after increasing the number of keys by a modest
                 amount, the effects of compact initialization are still
                 felt. Once the tree has grown so large that these
                 effects are no longer discernible, the tree can be
                 expeditiously compacted in place using an algorithm
                 presented here; and the benefits of compactness
                 resume.",
  acknowledgement = ack-nhfb,
  annote =       "A Btree is compact if it is minimal in number of
                 nodes. Compact Btree initialization can be done in time
                 linear in the number of keys if the keys are presorted.
                 Study indicates that space-optimal trees are nearly
                 time optimal, but time-optimal trees are nearly space
                 pessimal.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "2,3-tree; B-tree; bushy B-tree; compact B-tree; data
                 processing; node-visit cost; space utilization",
  subject =      "Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Trees}",
}

@Article{Scholl:1981:NFO,
  author =       "Michel Scholl",
  title =        "New File Organizations Based on Dynamic Hashing",
  journal =      j-TODS,
  volume =       "6",
  number =       "1",
  pages =        "194--211",
  month =        mar,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15",
  MRnumber =     "82c:68016",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-1/p194-scholl/p194-scholl.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-1/p194-scholl/",
  abstract =     "New file organizations based on hashing and suitable
                 for data whose volume may vary rapidly recently
                 appeared in the literature. In the three schemes which
                 have been independently proposed, rehashing is avoided,
                 storage space is dynamically adjusted to the number of
                 records actually stored, and there are no overflow
                 records. Two of these techniques employ an index to the
                 data file. Retrieval is fast and storage utilization is
                 low.\par

                 In order to increase storage utilization, we introduce
                 two schemes based on a similar idea and analyze the
                 performance of the second scheme. Both techniques use
                 an index of much smaller size. In both schemes,
                 overflow records are accepted. The price which has to
                 be paid for the improvement in storage utilization is a
                 slight access cost degradation.",
  acknowledgement = ack-nhfb,
  annote =       "In the three schemes which proposed, rehashing is
                 avoided, storage space is dynamically adjusted to the
                 number of records actually stored, and there are no
                 overflow records. Two of these techniques employ an
                 index to the data file.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; data structure; dynamic hashing; file
                 organization; hashing; linear splitting",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Kung:1981:OMC,
  author =       "H. T. Kung and John T. Robinson",
  title =        "On Optimistic Methods for Concurrency Control",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "213--226",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Compiler/garbage.collection.bib;
                 Compiler/Heaps.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/misc.1.bib; Misc/real.time.bib;
                 Object/Nierstrasz.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p213-kung/p213-kung.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p213-kung/",
  abstract =     "Most current approaches to concurrency control in
                 database systems rely on locking of data objects as a
                 control mechanism. In this paper, two families of
                 nonlocking concurrency controls are presented. The
                 methods used are ``optimistic'' in the sense that they
                 rely mainly on transaction backup as a control
                 mechanism, ``hoping'' that conflicts between
                 transactions will not occur. Applications for which
                 these methods should be more efficient than locking are
                 discussed.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency controls; data base systems, concurrency
                 other; databases; transaction processing",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}",
}

@Article{Boral:1981:PAS,
  author =       "Haran Boral and David J. DeWitt",
  title =        "Processor Allocation Strategies for Multiprocessor
                 Database Machines",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "227--254",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p227-boral/p227-boral.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p227-boral/",
  abstract =     "In this paper four alternative strategies for
                 assigning processors to queries in multiprocessor
                 database machines are described and evaluated. The
                 results demonstrate that SIMD database machines are
                 indeed a poor design when their performance is compared
                 with that of the three MIMD strategies presented.
                 \par

                 Also introduced is the application of data-flow machine
                 techniques to the processing of relational algebra
                 queries. A strategy that employs data-flow techniques
                 is shown to be superior to the other strategies
                 described by several experiments. Furthermore, if the
                 data-flow query processing strategy is employed, the
                 results indicate that a two-level storage hierarchy (in
                 which relations are paged between a shared data cache
                 and mass storage) does not have a significant impact on
                 performance.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative processors; back-end computers; computer
                 architecture; data base systems, Direct TODS; data-flow
                 computers; database machines; database management;
                 parallel processors; processor scheduling",
  subject =      "Information Systems --- Database Management ---
                 Database Machines (H.2.6); Information Systems ---
                 Database Management (H.2)",
}

@Article{Su:1981:TDT,
  author =       "Stanley Y. W. Su and Herman Lam and Der Her Lo",
  title =        "Transformation of Data Traversals and Operations in
                 Application Programs to Account for Semantic Changes of
                 Databases",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "255--294",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p255-su/p255-su.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p255-su/",
  abstract =     "This paper addresses the problem of application
                 program conversion to account for changes in database
                 semantics that result in changes in the schema and
                 database contents. With the observation that the
                 existing data models can be viewed as alternative ways
                 of modeling the same database semantics, a methodology
                 of application program analysis and conversion based on
                 an existing-DBMS-model-and schema-independent
                 representation of both the database and programs is
                 presented. In this methodology, the source and target
                 databases are described in terms of the association
                 types of a semantic association model. The structural
                 properties, the integrity constraints, and the
                 operational characteristics (storage operation
                 behaviors) of the association types are more explicitly
                 defined to reveal the semantics that is generally
                 hidden in application programs. The explicit
                 descriptions of the source and target databases are
                 used as the basis for program analysis and conversion.
                 Application programs are described in terms of a small
                 number of ``access patterns'' which define the data
                 traversals and operations of the programs. In addition
                 to the methodology, this paper (1) describes a model of
                 a generalized application program conversion system
                 that serves as a framework for research, (2) presents
                 an analysis of access patterns that serve as the
                 primitives for program description, (3) delineates some
                 meaningful semantic changes to databases and their
                 corresponding transformation rules for program
                 conversion, (4) illustrates the application of these
                 rules to two different approaches to program conversion
                 problems, and (5) reports on the development effort
                 undertaken at the University of Florida.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access pattern; application program conversion; data
                 base systems; database changes; semantic data model;
                 transformation rules",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8); Information Systems ---
                 Database Management --- Physical Design (H.2.2): {\bf
                 Access methods}; Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Data
                 models}",
}

@Article{Clemons:1981:DES,
  author =       "Eric K. Clemons",
  title =        "Design of an External Schema Facility to Define and
                 Process Recursive Structures",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "295--311",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p295-clemons/p295-clemons.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p295-clemons/",
  abstract =     "The role of the external schema is to support user
                 views of data and thus to provide programmers with
                 easier data access. This author believes that an
                 external schema facility is best based on hierarchies,
                 both simple and recursive. After a brief introduction
                 to an external schema facility to support simple
                 hierarchical user views, the requirements for a
                 facility for recursive hierarchies are listed and the
                 necessary extensions to the external schema definition
                 language are offered.\par

                 Functions that must be provided for generality in
                 definition are node specification and node control.
                 Tree traversal functions must be provided for
                 processing. Definitions of each and examples of use are
                 presented.",
  acknowledgement = ack-nhfb,
  annote =       "[Ahad,Yao,Choi87] A.2.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "ANSI SPARC architectures; data base systems; external
                 schemata; recursive data structures; user views",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Schema and subschema}",
}

@Article{Davida:1981:DES,
  author =       "George I. Davida and David L. Wells and John B. Kam",
  title =        "A Database Encryption System with Subkeys",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "312--328",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15",
  MRnumber =     "82f:68020",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p312-davida/p312-davida.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p312-davida/",
  abstract =     "A new cryptosystem that is suitable for database
                 encryption is presented. The system has the important
                 property of having subkeys that allow the encryption
                 and decryption of fields within a record. The system is
                 based on the Chinese Remainder Theorem.",
  acknowledgement = ack-nhfb,
  annote =       "Subkeys allow the encryption and decryption of fields
                 within a record. The system is based on the Chinese
                 Remainder Theorem.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "codes, symbolic; data base systems; data security;
                 databases; decryption; encryption; subkeys",
  subject =      "Data --- Data Encryption (E.3)",
}

@Article{Ling:1981:ITN,
  author =       "Tok Wang Ling and Frank W. Tompa and Tiko Kameda",
  title =        "An Improved Third Normal Form for Relational
                 Databases",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "329--346",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15",
  MRnumber =     "82f:68024",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p329-ling/p329-ling.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p329-ling/",
  abstract =     "In this paper, we show that some Codd third normal
                 form relations may contain ``superfluous'' attributes
                 because the definitions of transitive dependency and
                 prime attribute are inadequate when applied to sets of
                 relations. To correct this, an improved third normal
                 form is defined and an algorithm is given to construct
                 a set of relations from a given set of functional
                 dependencies in such a way that the superfluous
                 attributes are guaranteed to be removed. This new
                 normal form is compared with other existing definitions
                 of third normal form, and the deletion normalization
                 method proposed is shown to subsume the decomposition
                 method of normalization.",
  acknowledgement = ack-nhfb,
  annote =       "An improved third normal form is defined and an
                 algorithm is given to construct a set of relations from
                 a given set of functional dependencies in such a way
                 that the superfluous attributes are guaranteed to be
                 removed.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "covering; data base systems; database design;
                 functional dependency; normalization; prime attribute;
                 reconstructibility; relational schema; third normal
                 form; transitive dependency",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Schema and subschema}",
}

@Article{McLean:1981:CSC,
  author =       "Gordon {McLean, Jr.}",
  title =        "Comments on {SDD-1} Concurrency Control Mechanisms",
  journal =      j-TODS,
  volume =       "6",
  number =       "2",
  pages =        "347--350",
  month =        jun,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-2/p347-mclean/p347-mclean.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-2/p347-mclean/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}",
}

@Article{Hammer:1981:DDS,
  author =       "Michael Hammer and Dennis Mc Leod",
  title =        "Database Description with {SDM}: a Semantic Database
                 Model",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "351--386",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/bibdb.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Reprinted in \cite{Stonebraker:1988:RDS}. Also
                 published in \cite{Zdonik:1990:ROO}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p351-hammer/p351-hammer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p351-hammer/",
  abstract =     "SDM is a high-level semantics-based database
                 description and structuring formalism (database model)
                 for databases. This database model is designed to
                 capture more of the meaning of an application
                 environment than is possible with contemporary database
                 models. An SDM specification describes a database in
                 terms of the kinds of entities that exist in the
                 application environment, the classifications and
                 groupings of those entities, and the structural
                 interconnections among them. SDM provides a collection
                 of high-level modeling primitives to capture the
                 semantics of an application environment. By
                 accommodating derived information in a database
                 structural specification, SDM allows the same
                 information to be viewed in several ways; this makes it
                 possible to directly accommodate the variety of needs
                 and processing requirements typically present in
                 database applications. The design of the present SDM is
                 based on our experience in using a preliminary version
                 of it.\par

                 SDM is designed to enhance the effectiveness and
                 usability of database systems. An SDM database
                 description can serve as a formal specification and
                 documentation tool for a database; it can provide a
                 basis for supporting a variety of powerful user
                 interface facilities, it can serve as a conceptual
                 database model in the database design process; and, it
                 can be used as the database model for a new kind of
                 database management system.",
  acknowledgement = ack-nhfb,
  annote =       "SDM is a high-level semantics-based database model, to
                 capture the meaning of an application environment. One
                 of the papers usually referred to when discussing
                 semantic data models. Describes a model which permits a
                 lot of flexibility and expressiveness, and is
                 consequently difficult to implement. Advantage is that
                 it can be used as a specification and documentation
                 tool. Good introduction, giving an overview of (some?,
                 most?, all?) problems in semantic data models. The
                 section describing SDM DDL is a bit too detailed (one
                 needs to pick up the essential concepts like
                 subclassing, and redundancy in model (which may be
                 necessary to make the model easier to use)). Some
                 discussion of inheritance is also present. Nothing much
                 is said in the final discussion. Reasonable paper. To
                 benefit, one needs to be careful not to get lost in the
                 details. A detailed description of the semantic data
                 model.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database definition; database
                 management; database modeling; database models;
                 database semantics; logical database design",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}",
}

@Article{Fagin:1981:NFR,
  author =       "Ronald Fagin",
  title =        "A Normal Form for Relational Databases That is Based
                 on Domains and Keys",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "387--415",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Compiler/prog.lang.theory.bib;
                 Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p387-fagin/p387-fagin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p387-fagin/",
  abstract =     "The new normal form for relational databases, called
                 domain-key normal form (DK\slash NF), is defined. Also,
                 formal definitions of insertion anomaly and deletion
                 anomaly are presented. It is shown that a schema is in
                 DK\slash NF if and only if it has no insertion or
                 deletion anomalies. Unlike previously defined normal
                 forms, DK\slash NF is not defined in terms of
                 traditional dependencies (functional, multivalued, or
                 join). Instead, it is defined in terms of the more
                 primitive concepts of domain and key, along with the
                 general concept of a ``constraint''. It is considered
                 how the definitions of traditional normal forms might
                 be modified by taking into consideration, for the first
                 time, the combinatorial consequences of bounded domain
                 sizes. It is shown that after this modification, these
                 traditional normal forms are all implied by DK\slash
                 NF. In particular, if all domains are infinite, then
                 these traditional normal forms are all implied by
                 DK\slash NF.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "anomaly; complexity; data base systems; database
                 design; DK/NF; domain-key normal form; functional
                 dependency; join dependency; multivalued dependency;
                 normalization; relational database",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Normal forms}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Relational databases}",
}

@Article{Hong:1981:AHS,
  author =       "Y. C. Hong and Stanley Y. W. Su",
  title =        "Associative Hardware and Software Techniques for
                 Integrity Control",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "416--440",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p416-hong/p416-hong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p416-hong/",
  abstract =     "This paper presents the integrity control mechanism of
                 the associative processing system, CASSM. The mechanism
                 takes advantage of the associative techniques, such as
                 content and context addressing, tagging and marking
                 data, parallel processing, automatic triggering of
                 integrity control procedures, etc., for integrity
                 control and as a result offers three significant
                 advantages: (1) The problem of staging data in a main
                 memory for integrity checking can be eliminated because
                 database storage operations are verified at the place
                 where the data are stored. (2) The backout or merging
                 procedures are relatively easy and inexpensive in the
                 associative system because modified copies can be
                 substituted for the originals or may be discarded by
                 merely changing their associated tags. (3) The database
                 management system software is simplified because
                 database integrity functions are handled by the
                 associative processing system to which a mainframe
                 computer is a front-end computer.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "assertion and trigger; associative techniques;
                 cellular-logic devices; data base systems; database
                 integrity; database management; integrity control;
                 integrity control, SYWSu hardware support relational
                 database machine TODS",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0): {\bf Security, integrity, and
                 protection**}",
}

@Article{March:1981:FMS,
  author =       "Salvatore T. March and Dennis G. Severance and Michael
                 Wilens",
  title =        "Frame Memory: a Storage Architecture to Support Rapid
                 Design and Implementation of Efficient Databases",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "441--463",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p441-march/p441-march.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p441-march/",
  abstract =     "Frame memory is a virtual view of secondary storage
                 that can be implemented with reasonable overhead to
                 support database record storage and accessing
                 requirements. Frame memory is designed so that its
                 operating characteristics can be easily manipulated by
                 either designers or design algorithms, while
                 performance effects of such changes can be accurately
                 predicted. Automated design procedures exist to
                 generate and evaluate alternative database designs
                 built upon frame memory, and the existence of these
                 procedures establishes frames as an attractive memory
                 management architecture for future database management
                 systems.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "analytic modeling; data base systems; database design
                 system; database machine; hardware support; TODS;
                 virtual secondary storage",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2); Software ---
                 Operating Systems --- Storage Management (D.4.2): {\bf
                 Secondary storage}",
}

@Article{vandeRiet:1981:HLP,
  author =       "Reind P. {van de Riet} and Anthony I. Wasserman and
                 Martin L. Kersten and Wiebren {de Jonge}",
  title =        "High-Level Programming Features for Improving the
                 Efficiency of a Relational Database System",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "464--485",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: UCSF, Lab. of Med. Inf. Science,
                 Tech. Rpt. 44, Feb. 1980.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p464-van_de_riet/p464-van_de_riet.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p464-van_de_riet/",
  abstract =     "This paper discusses some high-level language
                 programming constructs that can be used to manipulate
                 the relations of a relational database system
                 efficiently. Three different constructs are described:
                 (1) tuple identifiers that directly reference tuples of
                 a relation; (2) cursors that may iterate over the
                 tuples of a relation; and (3) markings, a form of
                 temporary relation consisting of a set of tuple
                 identifiers. In each case, attention is given to
                 syntactic, semantic, and implementation considerations.
                 \par

                 The use of these features is first presented within the
                 context of the programming language PLAIN, and it is
                 then shown how these features could be used more
                 generally to provide database manipulation capabilities
                 in a high-level programming language. Consideration is
                 also given to issues of programming methodology, with
                 an important goal being the achievement of a balance
                 between the enforcement of good programming practices
                 and the ability to write efficient programs.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "markings; PLAIN; programming languages; programming
                 methodology; relational algebra; relational database
                 management",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Database (persistent)
                 programming languages}",
}

@Article{Culik:1981:DMT,
  author =       "K. {Culik II} and Th. Ottmann and D. Wood",
  title =        "Dense multiway trees",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "486--512",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15 (05C05)",
  MRnumber =     "82m:68038",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p486-culic/",
  abstract =     "B-trees of order $m$ are a ``balanced'' class of
                 $m$-ary trees, which have applications in the areas of
                 file organization. In fact, they have been the only
                 choice when balanced multiway trees are required.
                 Although they have very simple insertion and deletion
                 algorithms, their storage utilization, that is, the
                 number of keys per page or node, is at worst 50
                 percent. In the present paper we investigate a new
                 class of balanced $m$-ary trees, the dense multiway
                 trees, and compare their storage utilization with that
                 of B-trees of order $m$. \par

                 Surprisingly, we are able to demonstrate that weakly
                 dense multiway trees have an $ (l o g_2 N) $ insertion
                 algorithm. We also show that inserting $ m h - 1 $ keys
                 in ascending order into an initially empty dense
                 multiway tree yields the complete $m$-ary tree of
                 height $h$, and that at intermediate steps in the
                 insertion sequence the intermediate trees can also be
                 considered to be as dense as possible. Furthermore, an
                 analysis of the limiting dynamic behavior of the dense
                 $m$-ary trees under insertion shows that the average
                 storage utilization tends to 1; that is, the trees
                 become as dense as possible. This motivates the use of
                 the term ``dense.''",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "B-trees; balanced trees; dense trees; multiway trees;
                 search trees; storage utilization",
  subject =      "Data --- Data Structures (E.1): {\bf Trees}",
}

@Article{Comer:1981:AHF,
  author =       "Douglas Comer",
  title =        "Analysis of a Heuristic for Full Trie Minimization",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "513--537",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-3/p513-comer/p513-comer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-3/p513-comer/",
  abstract =     "A trie is a distributed-key search tree in which
                 records from a file correspond to leaves in the tree.
                 Retrieval consists of following a path from one root to
                 a leaf, where the choice of edge at each node is
                 determined by attribute values of the key. For full
                 tries, those in which all leaves lie at the same depth,
                 the problem of finding an ordering of attributes which
                 yields a minimum size trie is NP-complete.\par

                 This paper considers a ``greedy'' heuristic for
                 constructing low-cost tries. It presents simulation
                 experiments which show that the greedy method tends to
                 produce tries with small size, and analysis leading to
                 a worst case bound on approximations produced by the
                 heuristic. It also shows a class of files for which the
                 greedy method may perform badly, producing tries of
                 high cost.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; heuristic; trie index; trie size",
  subject =      "Computing Methodologies --- Artificial Intelligence
                 --- Problem Solving, Control Methods, and Search
                 (I.2.8): {\bf Heuristic methods}",
}

@Article{Kent:1981:CAU,
  author =       "W. Kent",
  title =        "Consequences of Assuming a Universal Relation",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "539--556",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See remark \cite{Ullman:1983:KCA}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p539-kent/p539-kent.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p539-kent/",
  abstract =     "Although central to the current direction of
                 dependency theory, the assumption of a universal
                 relation is incompatible with some aspects of
                 relational database theory and practice. Furthermore,
                 the universal relation is itself ill defined in some
                 important ways. And, under the universal relation
                 assumption, the decomposition approach to database
                 design becomes virtually indistinguishable from the
                 synthetic approach.",
  acknowledgement = ack-nhfb,
  annote =       "The assumption of a universal relation is incompatible
                 with some aspects of relational database theory and
                 practice. Under the universal relation assumption, the
                 decomposition approach to database design becomes
                 virtually indistinguishable from the synthetic
                 approach.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database design; dependency theory;
                 rational database; relational theory; universal
                 relation",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Models and Principles ---
                 Systems and Information Theory (H.1.1); Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1)",
}

@Article{Bancilhon:1981:USR,
  author =       "F. B. Bancilhon and N. Spyratos",
  title =        "Update Semantics of Relational Views",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "557--575",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See comment \cite{Keller:1987:CBS}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p557-bancilhon/p557-bancilhon.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p557-bancilhon/",
  abstract =     "A database view is a portion of the data structured in
                 a way suitable to a specific application. Updates on
                 views must be translated into updates on the underlying
                 database. This paper studies the translation process in
                 the relational model.\par

                 The procedure is as follows: first, a ``complete'' set
                 of updates is defined such that\par

                 together with every update the set contains a
                 ``return'' update, that is, one that brings the view
                 back to the original state;\par

                 given two updates in the set, their composition is also
                 in the set.\par

                 To translate a complete set, we define a mapping called
                 a ``translator,'' that associates with each view update
                 a unique database update called a ``translation.'' The
                 constraint on a translation is to take the database to
                 a state mapping onto the updated view. The constraint
                 on the translator is to be a morphism.\par

                 We propose a method for defining translators. Together
                 with the user-defined view, we define a
                 ``complementary'' view such that the database could be
                 computed from the view and its complement. We show that
                 a view can have many different complements and that the
                 choice of a complement determines an update policy.
                 Thus, we fix a view complement and we define the
                 translation of a given view update in such a way that
                 the complement remains invariant (``translation under
                 constant complement''). The main result of the paper
                 states that, given a complete set $U$ of view updates,
                 $U$ has a translator if and only if $U$ is translatable
                 under constant complement.",
  acknowledgement = ack-nhfb,
  annote =       "A mapping called a ``translator'', associates with
                 each view update a unique database update. A method for
                 defining translators with the user-defined view, define
                 a ``complementary'' view such that the database could
                 be computed from the view and its complement. We define
                 the translation of a given view update in such a way
                 that the complement remains invariant. Aplies to
                 Universal relations.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual model; data base systems; data model; data
                 semantics; database view; relation; relational model
                 database; update translation; view updating",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Theory of
                 Computation --- Logics and Meanings of Programs ---
                 Semantics of Programming Languages (F.3.2)",
}

@Article{Baroody:1981:OOA,
  author =       "A. James {Baroody, Jr.} and David J. DeWitt",
  title =        "An Object-Oriented Approach to Database System
                 Implementation",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "576--601",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Object/Nierstrasz.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p576-baroody/p576-baroody.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p576-baroody/",
  abstract =     "This paper examines object-oriented programming as an
                 implementation technique for database systems. The
                 object-oriented approach encapsulates the
                 representations of database entities and relationships
                 with the procedures that manipulate them. To achieve
                 this, we first define abstractions of the modeling
                 constructs of the data model that describe their common
                 properties and behavior. Then we represent the entity
                 types and relationship types in the conceptual schema
                 and the internal schema by objects that are instances
                 of these abstractions. The generic procedures (data
                 manipulation routines) that comprise the user interface
                 can now be implemented as calls to the procedures
                 associated with these objects.\par

                 A generic procedure model of database implementation
                 techniques is presented and discussed. Several current
                 database system implementation techniques are
                 illustrated as examples of this model, followed by a
                 critical analysis of our implementation technique based
                 on the use of objects. We demonstrate that the
                 object-oriented approach has advantages of data
                 independence, run-time efficiency due to eliminating
                 access to system descriptors, and support for low-level
                 views.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming, olit-db casais; data base
                 systems; data independence; data manipulation routines;
                 database systems; high-level languages; object-oriented
                 programming; procedural binding",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Object-oriented databases};
                 Computer Systems Organization --- Computer System
                 Implementation (C.5); Information Systems --- Database
                 Management --- Languages (H.2.3)",
}

@Article{Bernstein:1981:QPS,
  author =       "Philip A. Bernstein and Nathan Goodman and Eugene Wong
                 and Christopher L. Reeve and James B. {Rothnie, Jr.}",
  title =        "Query Processing in a System for Distributed Databases
                 ({SDD-1})",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "602--625",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p602-bernstein/p602-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p602-bernstein/",
  abstract =     "This paper describes the techniques used to optimize
                 relational queries in the SDD-1 distributed database
                 system. Queries are submitted to SDD-1 in a high-level
                 procedural language called Datalanguage. Optimization
                 begins by translating each Datalanguage query into a
                 relational calculus form called an {\em envelope},
                 which is essentially an aggregate-free QUEL query. This
                 paper is primarily concerned with the optimization of
                 envelopes.\par

                 Envelopes are processed in two phases. The first phase
                 executes relational operations at various sites of the
                 distributed database in order to delimit a subset of
                 the database that contains all data relevant to the
                 envelope. This subset is called a {\em reduction\/} of
                 the database. The second phase transmits the reduction
                 to one designated site, and the query is executed
                 locally at that site.\par

                 The critical optimization problem is to perform the
                 reduction phase efficiently. Success depends on
                 designing a good repertoire of operators to use during
                 this phase, and an effective algorithm for deciding
                 which of these operators to use in processing a given
                 envelope against a given database. The principal
                 reduction operator that we employ is called a {\em
                 semijoin}. In this paper we define the semijoin
                 operator, explain why semijoin is an effective
                 reduction operator, and present an algorithm that
                 constructs a cost-effective program of semijoins, given
                 an envelope and a database.",
  acknowledgement = ack-nhfb,
  annote =       "Techniques to optimize relational queries in the SDD-1
                 distributed database system. First phase executes
                 relational operations at various sites to delimit a
                 subset called a reduction. The second phase transmits
                 the reduction to one designated site. The principal
                 reduction operator, introduced here, is called a
                 semijoin.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- subroutines; data base
                 systems; distributed databases; query optimization;
                 query processing; query processing, TODS semijoins
                 semi-join join; relational databases; semijoins",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Distributed databases}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Welty:1981:HFC,
  author =       "Charles Welty and David W. Stemple",
  title =        "Human Factors Comparison of a Procedural and a
                 Nonprocedural Query Language",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "626--649",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p626-welty/p626-welty.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p626-welty/",
  abstract =     "Two experiments testing the ability of subjects to
                 write queries in two different query languages were
                 run. The two languages, SQL and TABLET, differ
                 primarily in their procedurality; both languages use
                 the relational data model, and their Halstead levels
                 are similar. Constructs in the languages which do not
                 affect their procedurality are identical. The two
                 languages were learned by the experimental subjects
                 almost exclusively from manuals presenting the same
                 examples and problems ordered identically for both
                 languages. The results of the experiments show that
                 subjects using the more procedural language wrote
                 difficult queries better than subjects using the less
                 procedural language. The results of the experiments are
                 also used to compare corresponding constructs in the
                 two languages and to recommend improvements for these
                 constructs.",
  acknowledgement = ack-nhfb,
  annote =       "SQL and TABLET. The results show that subjects using
                 the more procedural language wrote difficult queries
                 better than subjects using the less procedural
                 language.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; database systems; human factors;
                 procedural and nonprocedural languages; query
                 languages",
  subject =      "Information Systems --- Models and Principles ---
                 User/Machine Systems (H.1.2): {\bf Human factors};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Systems (H.2.4)",
}

@Article{Lehman:1981:ELC,
  author =       "Philip L. Lehman and S. Bing Yao",
  title =        "Efficient Locking for Concurrent Operations on
                 {B-Trees}",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "650--670",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p650-lehman/p650-lehman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p650-lehman/",
  abstract =     "The B-tree and its variants have been found to be
                 highly useful (both theoretically and in practice) for
                 storing large amounts of information, especially on
                 secondary storage devices. We examine the problem of
                 overcoming the inherent difficulty of concurrent
                 operations on such structures, using a practical
                 storage model. A single additional ``link'' pointer in
                 each node allows a process to easily recover from tree
                 modifications performed by other concurrent processes.
                 Our solution compares favorably with earlier solutions
                 in that the locking scheme is simpler (no read-locks
                 are used) and only a (small) constant number of nodes
                 are locked by any update process at any given time. An
                 informal correctness proof for our system is given.",
  acknowledgement = ack-nhfb,
  annote =       "A single additional `link' pointer in each node allows
                 a process to easily recover from tree modifications
                 performed by other concurrent processes. No read-locks
                 are used only a (small) constant number of nodes are
                 locked by any update process at any given time.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "B-tree; concurrenct algorithms; concurrency controls;
                 consistency; correctness; data processing; data
                 structures; database; index organizations; locking
                 protocols; multiway search trees",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}; Mathematics of
                 Computing --- Discrete Mathematics --- Graph Theory
                 (G.2.2): {\bf Trees}",
}

@Article{Larson:1981:AIS,
  author =       "Per-{\AA}ke Larson",
  title =        "Analysis of Index-Sequential Files with Overflow
                 Chaining",
  journal =      j-TODS,
  volume =       "6",
  number =       "4",
  pages =        "671--680",
  month =        dec,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68B15 (68H05)",
  MRnumber =     "82m:68044",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1981-6-4/p671-larson/p671-larson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1981-6-4/p671-larson/",
  abstract =     "The gradual performance deterioration caused by
                 deletions from and insertions into an index-sequential
                 file after loading is analyzed. The model developed
                 assumes that overflow records are handled by chaining.
                 Formulas for computing the expected number of overflow
                 records and the expected number of additional accesses
                 caused by the overflow records for both successful and
                 unsuccessful searches are derived.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "analysis of algorithms; analytic model; data
                 processing, TODS ISAM; file organization; file
                 structure; index sequential files; indexed sequential
                 access method; ISAM; overflow; overflow chaining;
                 overflow handling; performance analysis",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Information Systems --- Database
                 Management --- Physical Design (H.2.2): {\bf Access
                 methods}",
}

@Article{Comer:1981:EKD,
  author =       "D. Comer",
  title =        "Extended {K-d} Tree Database Organization: a Dynamic
                 Multiattribute File Corresponds to Leaves in the Tree",
  journal =      j-TODS,
  volume =       "6",
  number =       "3",
  pages =        "??--??",
  month =        sep,
  year =         "1981",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:49:00 1996",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  annote =       "This paper considers a `greedy' heuristic for
                 constructing low-cost trees.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Zaniolo:1982:DRN,
  author =       "C. Zaniolo",
  title =        "Database Relations with Null Values",
  journal =      j-TODS,
  volume =       "1",
  number =       "1",
  pages =        "??--??",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:48:57 1996",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  annote =       "a three-valued logic: TRUE, FALSE, UNKNOWN",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Katz:1982:DCD,
  author =       "R. H. Katz and E. Wong",
  title =        "Decompiling {CODASYL DML} into Relational Queries",
  journal =      j-TODS,
  volume =       "7",
  number =       "1",
  pages =        "1--23",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-1/p1-katz/p1-katz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-1/p1-katz/",
  abstract =     "A ``decompilation'' algorithm is developed to
                 transform a program written with the procedural
                 operations of CODASYL DML into one which interacts with
                 a relational system via a nonprocedural query
                 specification. An Access Path Model is introduced to
                 interpret the semantic accesses performed by the
                 program. Data flow analysis is used to determine how
                 FIND operations implement semantic accesses. A sequence
                 of these is mapped into a relational query and embedded
                 into the original program. The class of programs for
                 which the algorithm succeeds is characterized.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems; decompilation; semantic data
                 models",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Relational databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data manipulation languages
                 (DML)}; Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Heterogeneous
                 Databases (H.2.5): {\bf Program translation**}",
}

@Article{Zaniolo:1982:FAD,
  author =       "Carlo Zaniolo and Michel A. Melkanoff",
  title =        "A Formal Approach to the Definition and the Design of
                 Conceptual Schemata for Database Systems",
  journal =      j-TODS,
  volume =       "7",
  number =       "1",
  pages =        "24--59",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-1/p24-zaniolo/p24-zaniolo.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-1/p24-zaniolo/",
  abstract =     "A formal approach is proposed to the definition and
                 the design of conceptual database diagrams to be used
                 as conceptual schemata in a system featuring a
                 multilevel schema architecture, and as an aid for the
                 design of other forms of schemata. We consider E-R
                 (entity-relationship) diagrams, and we introduce a new
                 representation called {\em CAZ\/}-graphs. A rigorous
                 connection is established between these diagrams and
                 some formal constraints used to describe relationships
                 in the framework of the relational data model. These
                 include functional and multivalued dependencies of
                 database relations. The basis for our schemata is a
                 combined representation for two fundamental structures
                 underlying every relation: the first defined by its
                 minimal atomic decompositions, the second by its
                 elementary functional dependencies.\par

                 The interaction between these two structures is
                 explored, and we show that, jointly, they can represent
                 a wide spectrum of database relationships, of which the
                 well-known one-to-one, one-to-many, and many-to-many
                 associations constitute only a small subset. It is
                 suggested that a main objective in conceptual schema
                 design is to ensure a complete representation of these
                 two structures. A procedure is presented to design
                 schemata which obtain this objective while eliminating
                 redundancy. A simple correspondence between the
                 topological properties of these schemata and the
                 structure of multivalued dependencies of the original
                 relation is established. Various applications are
                 discussed and a number of illustrative examples are
                 given.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems, logical design TODS",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Data
                 models}; Information Systems --- Database Management
                 --- Logical Design (H.2.1): {\bf Schema and
                 subschema}",
}

@Article{Batory:1982:OFD,
  author =       "D. S. Batory",
  title =        "Optimal File Designs and Reorganization Points",
  journal =      j-TODS,
  volume =       "7",
  number =       "1",
  pages =        "60--81",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: University of Toronto,
                 TR-CSRG-110, 1980.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-1/p60-batory/p60-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-1/p60-batory/",
  abstract =     "A model for studying the combined problems of file
                 design and file reorganization is presented. New
                 modeling techniques for predicting the performance
                 evolution of files and for finding optimal
                 reorganization points for files are introduced.
                 Applications of the model to hash-based and
                 indexed-sequential files reveal important relationships
                 between initial loading factors and reorganization
                 frequency. A practical file design strategy, based on
                 these relationships, is proposed.",
  acknowledgement = ack-nhfb,
  annote =       "Applications of the model to hash-based and
                 indexed-sequential files reveal important relationships
                 between initial loading factors and reorganization
                 frequency.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; file design; file reorganization",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Physical Design (H.2.2)",
}

@Article{Du:1982:DAC,
  author =       "H. C. Du and J. S. Sobolewski",
  title =        "Disk Allocation for {Cartesian} Product Files on
                 Multiple-Disk Systems",
  journal =      j-TODS,
  volume =       "7",
  number =       "1",
  pages =        "82--101",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-1/p82-du/p82-du.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-1/p82-du/",
  abstract =     "Cartesian product files have recently been shown to
                 exhibit attractive properties for partial match
                 queries. This paper considers the file allocation
                 problem for Cartesian product files, which can be
                 stated as follows: Given a $k$-attribute Cartesian
                 product file and an $m$-disk system, allocate buckets
                 among the $m$ disks in such a way that, for all
                 possible partial match queries, the concurrency of disk
                 accesses is maximized. The Disk Modulo (DM) allocation
                 method is described first, and it is shown to be strict
                 optimal under many conditions commonly occurring in
                 practice, including all possible partial match queries
                 when the number of disks is 2 or 3. It is also shown
                 that although it has good performance, the DM
                 allocation method is not strict optimal for all
                 possible partial match queries when the number of disks
                 is greater than 3. The General Disk Modulo (GDM)
                 allocation method is then described, and a sufficient
                 but not necessary condition for strict optimality of
                 the GDM method for all partial match queries and any
                 number of disks is then derived. Simulation studies
                 comparing the DM and random allocation methods in terms
                 of the average number of disk accesses, in response to
                 various classes of partial match queries, show the
                 former to be significantly more effective even when the
                 number of disks is greater than 3, that is, even in
                 cases where the DM method is not strict optimal. The
                 results that have been derived formally and shown by
                 simulation can be used for more effective design of
                 optimal file systems for partial match queries. When
                 considering multiple-disk systems with independent
                 access paths, it is important to ensure that similar
                 records are clustered into the same or similar buckets,
                 while similar buckets should be dispersed uniformly
                 among the disks.",
  acknowledgement = ack-nhfb,
  annote =       "For partial match queries. Allocate buckets among the
                 m disks in such a way that, for all possible partial
                 match queries, the concurrency of disk accesses is
                 maximized.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Cartesian product files; data processing",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2); Information Systems ---
                 Information Storage and Retrieval --- Information
                 Storage (H.3.2): {\bf File organization}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3): {\bf Search
                 process}; Information Systems --- Database Management
                 --- Systems (H.2.4)",
}

@Article{Dahl:1982:DSD,
  author =       "Ver{\'o}nica Dahl",
  title =        "On Database Systems Development through Logic",
  journal =      j-TODS,
  volume =       "7",
  number =       "1",
  pages =        "102--123",
  month =        mar,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05 (03B99)",
  MRnumber =     "83f:68112",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-1/p102-dahl/p102-dahl.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-1/p102-dahl/",
  abstract =     "The use of logic as a single tool for formalizing and
                 implementing different aspects of database systems in a
                 uniform manner is discussed. The discussion focuses on
                 relational databases with deductive capabilities and
                 very high-level querying and defining features. The
                 computational interpretation of logic is briefly
                 reviewed, and then several pros and cons concerning the
                 description of data, programs, queries, and language
                 parser in terms of logic programs are examined. The
                 inadequacies are discussed, and it is shown that they
                 can be overcome by the introduction of convenient
                 extensions into logic programming. Finally, an
                 experimental database query system with a natural
                 language front end, implemented in PROLOG, is presented
                 as an illustration of these concepts. A description of
                 the latter from the user's point of view and a sample
                 consultation session in Spanish are included.",
  acknowledgement = ack-nhfb,
  annote =       "The use of logic as a single tool for relational
                 databases with deductive capabilities and very
                 high-level querying and defining features. Inadequacies
                 are discussed, and overcome by extensions into logic
                 programming. An experimental database query system with
                 a natural language front end, implemented in PROLOG, is
                 presented.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems, TODS relational database; rational
                 database",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Theory of Computation --- Mathematical
                 Logic and Formal Languages --- Mathematical Logic
                 (F.4.1): {\bf Logic and constraint programming};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Relational databases}; Information Systems ---
                 Database Management --- Languages (H.2.3): {\bf
                 Prolog}",
}

@Article{Addis:1982:RBL,
  author =       "T. R. Addis",
  title =        "A Relation-Based Language Interpreter for a Content
                 Addressable File Store",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "125--163",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p125-addis/p125-addis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p125-addis/",
  abstract =     "The combination of the Content Addressable File Store
                 (CAFS \footnote{CAFS is a registered trademark of
                 International Computers Limited}) and an extension of
                 relational analysis is described. This combination
                 allows a simple and compact implementation of a
                 database query and update language (FIDL). The language
                 has one of the important properties of a ``natural''
                 language interface by using a ``world model'' derived
                 from the relational analysis. The interpreter (FLIN)
                 takes full advantage of the CAFS by employing a unique
                 database storage technique which results in a fast
                 response to both queries and updates.",
  acknowledgement = ack-nhfb,
  annote =       "ICL CAFS is used.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer operating systems --- program Interpreters,
                 hardware support database machine CAFS TODS; content
                 addressing; data base systems",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- Database Machines
                 (H.2.6)",
}

@Article{Buneman:1982:ITD,
  author =       "Peter Buneman and Robert E. Frankel and Rishiyur
                 Nikhil",
  title =        "An Implementation Technique for Database Query
                 Languages",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "164--186",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/Functional.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p164-buneman/p164-buneman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p164-buneman/",
  abstract =     "Structured query languages, such as those available
                 for relational databases, are becoming increasingly
                 desirable for all database management systems. Such
                 languages are applicative: there is no need for an
                 assignment or update statement. A new technique is
                 described that allows for the implementation of
                 applicative query languages against most commonly used
                 database systems. The technique involves ``lazy''
                 evaluation and has a number of advantages over existing
                 methods: it allows queries and functions of arbitrary
                 complexity to be constructed; it reduces the use of
                 secondary storage; it provides a simple control
                 structure through which interfaces to other programs
                 may be constructed; and the implementation, including
                 the database interface, is quite compact. Although the
                 technique is presented for a specific functional
                 programming system and for a CODASYL DBMS, it is
                 general and may be used for other query languages and
                 database systems.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "applicative programming; coroutines; database
                 interfaces; functional, data base systems; lazy
                 evaluation; query languages; TODS functional FQL
                 applicative programming lazy evaluation",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Software --- Programming Languages ---
                 Language Classifications (D.3.2): {\bf Applicative
                 (functional) languages}; Information Systems ---
                 Database Management --- Languages (H.2.3): {\bf Query
                 languages}",
}

@Article{Obermarck:1982:DDD,
  author =       "Ron Obermarck",
  title =        "Distributed Deadlock Detection Algorithm",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "187--208",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/misc.1.bib; Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p187-obermarck/p187-obermarck.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p187-obermarck/",
  abstract =     "We propose an algorithm for detecting deadlocks among
                 transactions running concurrently in a distributed
                 processing network (i.e., a distributed database
                 system). The proposed algorithm is a distributed
                 deadlock detection algorithm. A proof of the
                 correctness of the distributed portion of the algorithm
                 is given, followed by an example of the algorithm in
                 operation. The performance characteristics of the
                 algorithm are also presented.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming; data base systems; deadlock
                 detection",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Computer Systems Organization ---
                 Computer-Communication Networks --- Distributed Systems
                 (C.2.4): {\bf Distributed databases}; Software ---
                 Operating Systems --- Process Management (D.4.1): {\bf
                 Deadlocks}; Software --- Operating Systems ---
                 Organization and Design (D.4.7): {\bf Distributed
                 systems}",
}

@Article{Garcia-Molina:1982:ROT,
  author =       "H{\'e}ctor Garc{\'\i}a-Molina and Gio Wiederhold",
  title =        "Read-Only Transactions in a Distributed Database",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "209--234",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p209-garcia-molina/p209-garcia-molina.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p209-garcia-molina/",
  abstract =     "A read-only transaction or query is a transaction
                 which does not modify any data. Read-only transactions
                 could be processed with general transaction processing
                 algorithms, but in many cases it is more efficient to
                 process read-only transactions with special algorithms
                 which take advantage of the knowledge that the
                 transaction only reads. This paper defines the various
                 consistency and currency requirements that read-only
                 transactions may have. The processing of the different
                 classes of read-only transactions in a distributed
                 database is discussed. The concept of {$R$} insularity
                 is introduced to characterize both the read-only and
                 update algorithms. Several simple update and read-only
                 transaction processing algorithms are presented to
                 illustrate how the query requirements and the update
                 algorithms affect the read-only transaction processing
                 algorithms.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; consistency; currency; data base
                 systems, TODS R insularity; query; R insularity;
                 read-only transaction; schedule; serializability;
                 transaction; transaction processing algorithm",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Query processing};
                 Computer Systems Organization ---
                 Computer-Communication Networks --- Distributed Systems
                 (C.2.4): {\bf Distributed databases}",
}

@Article{Shneiderman:1982:AAR,
  author =       "Ben Shneiderman and Glenn Thomas",
  title =        "An Architecture for Automatic Relational Database
                 System Conversion",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "235--257",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p235-shneiderman/p235-shneiderman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p235-shneiderman/",
  abstract =     "Changes in requirements for database systems
                 necessitate schema restructuring, database translation,
                 and application or query program conversion. An
                 alternative to the lengthy manual revision process is
                 proposed by offering a set of 15 transformations keyed
                 to the relational model of data and the relational
                 algebra. Motivations, examples, and detailed
                 descriptions are provided.",
  acknowledgement = ack-nhfb,
  annote =       "Alterations to the logical structure of a DB may
                 necessitate changes at three levels: (1) stored
                 database, (2) schema definition, and (3) application
                 programs or queries. Each transformation is assessed on
                 three features: (1) information preservation (data are
                 not destroyed, only their logical format is altered);
                 (2) data dependence (a data dependent transformation is
                 one in which the stored DB must be checked to determine
                 whether it is consistent with the logical format of the
                 target system); and (3) program dependence (a program
                 dependent transformation is one in which the
                 application programs must be checked to determine
                 whether the transformation is permissible). At every
                 stage the DB is kept in fourth normal form. The 15
                 transformations are divided into five groups. The first
                 group includes simple alterations, such as changing the
                 name of an attribute or relation (CHANGE NAME), or
                 adding or deleting attributes or relations (ADD/DELETE
                 ATTRIBUTES, INTRODUCE/SEPARATE). The role played by
                 keys in the relational model is clearly critical, and
                 particular care must be taken when transformations
                 involving these keys are being carried out. The second
                 group of transformations concerns the effect of adding
                 attributes to or deleting attributes from keys
                 (PROMOTE/DEMOTE). The third and fourth sets of
                 transformations are provided for the combining and
                 dividing of relations. (COMPOSE/DECOMPOSE,
                 PARTITION/MERGE). The final group of transformations is
                 concerned with functional dependencies",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "automatic conversion; data base systems; database
                 systems; relational model; transformations",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Schema and subschema};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data manipulation languages
                 (DML)}; Information Systems --- Database Management ---
                 Heterogeneous Databases (H.2.5): {\bf Program
                 translation**}",
}

@Article{Roussopoulos:1982:VIR,
  author =       "Nicholas Roussopoulos",
  title =        "View Indexing in Relational Databases",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "258--290",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p258-roussopoulos/p258-roussopoulos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p258-roussopoulos/",
  abstract =     "The design and maintenance of a useful database system
                 require efficient optimization of the logical access
                 paths which demonstrate repetitive usage patterns.
                 Views (classes of queries given by a query model) are
                 an appropriate intermediate logical representation for
                 databases. Frequently accessed views of databases need
                 to be supported by indexing to enhance retrieval. This
                 paper investigates the problem of selecting an optimal
                 index set of views and describes an efficient algorithm
                 for this selection.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data base systems, views precomputation index
                 selection TODS index selection; index selection",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2); Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf Query
                 processing}",
}

@Article{Jacobs:1982:IRL,
  author =       "Barry E. Jacobs and Alan R. Aronson and Anthony C.
                 Klug",
  title =        "On Interpretations of Relational Languages and
                 Solutions to the Implied Constraint Problem",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "291--315",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-2/p291-jacobs/p291-jacobs.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-2/p291-jacobs/",
  abstract =     "The interconnection between conceptual and external
                 levels of a relational database is made precise in
                 terms of the notion of ``interpretation'' between
                 first-order languages. This is then used to obtain a
                 methodology for discovering constraints at the external
                 level that are ``implied'' by constraints at the
                 conceptual level and by conceptual-to-external
                 mappings. It is also seen that these concepts are
                 important in other database issues, namely, automatic
                 program conversion, database design, and compile-time
                 error checking of embedded database languages. Although
                 this study deals exclusively with the relational
                 approach, it also discusses how these ideas can be
                 extended to hierarchical and network databases.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "constraints; data base systems; program conversion;
                 relational database",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Schema and subschema}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Data manipulation languages (DML)}",
}

@Article{Chamberlin:1982:HFC,
  author =       "Donald D. Chamberlin",
  title =        "On ``Human Factors Comparison of a Procedural and a
                 Nonprocedural Query Language''",
  journal =      j-TODS,
  volume =       "7",
  number =       "2",
  pages =        "316--317",
  month =        jun,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:45:59 1996",
  bibsource =    "Database/Graefe.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "TODS technical correspondence",
}

@Article{Traiger:1982:TCD,
  author =       "Irving L. Traiger and Jim Gray and Cesare A. Galtieri
                 and Bruce G. Lindsay",
  title =        "Transactions and Consistency in Distributed Database
                 Systems",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "323--342",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p323-traiger/p323-traiger.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p323-traiger/",
  abstract =     "The concepts of transaction and of data consistency
                 are defined for a distributed system. The cases of
                 partitioned data, where fragments of a file are stored
                 at multiple nodes, and replicated data, where a file is
                 replicated at several nodes, are discussed. It is
                 argued that the distribution and replication of data
                 should be transparent to the programs which use the
                 data. That is, the programming interface should provide
                 location transparency, replica transparency,
                 concurrency transparency, and failure transparency.
                 Techniques for providing such transparencies are
                 abstracted and discussed.\par

                 By extending the notions of system schedule and system
                 clock to handle multiple nodes, it is shown that a
                 distributed system can be modeled as a single
                 sequential execution sequence. This model is then used
                 to discuss simple techniques for implementing the
                 various forms of transparency.",
  acknowledgement = ack-nhfb,
  annote =       "This paper is a easy-to-read introduction to required
                 transparency in distributed database systems. 4
                 transparencies are chosen and explained here, namely
                 location transparency, replication transparency,
                 concurrency transparency, and failure transparency. The
                 transaction model adapted by the paper is fully
                 synchronous and 2 phase protocol is used to implement
                 concurrency transparency. The paper proves that if all
                 transaction executions are two-phase, any legal
                 execution of the transactions by a distributed system
                 will be equivalent to some serial execution of the
                 transactions by a system consisting of a single node
                 under the assumption that updates are synchronous. The
                 paper introduces special node-associated clock to prove
                 it. The paper also gives simple explanation about a
                 protocol to implement failure transparency using logs
                 and two-phase commit protocol.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; data partitioning; data
                 replication; recovery; TODS data replication, data
                 partitioning",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}",
}

@Article{Fagin:1982:SUR,
  author =       "Ronald Fagin and Alberto O. Mendelzon and Jeffrey D.
                 Ullman",
  title =        "A Simplified Universal Relation Assumption and its
                 Properties",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "343--360",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05 (68B15)",
  MRnumber =     "83k:68100",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/bibdb.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p343-fagin/p343-fagin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p343-fagin/",
  abstract =     "One problem concerning the universal relation
                 assumption is the inability of known methods to obtain
                 a database scheme design in the general case, where the
                 real-world constraints are given by a set of
                 dependencies that includes embedded multivalued
                 dependencies. We propose a simpler method of describing
                 the real world, where constraints are given by
                 functional dependencies and a single join dependency.
                 The relationship between this method of defining the
                 real world and the classical methods is exposed. We
                 characterize in terms of hypergraphs those multivalued
                 dependencies that are the consequence of a given join
                 dependency. Also characterized in terms of hypergraphs
                 are those join dependencies that are equivalent to a
                 set of multivalued dependencies.",
  acknowledgement = ack-nhfb,
  annote =       "Constraints are functional dependencies and a single
                 join dependency.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "acyclic; database scheme; hypergraph; join dependency;
                 multivalued dependency; relational database",
  subject =      "Theory of Computation --- Mathematical Logic and
                 Formal Languages --- Mathematical Logic (F.4.1);
                 Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Graph algorithms};
                 Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Trees}; Information Systems
                 --- Database Management --- Logical Design (H.2.1):
                 {\bf Normal forms}; Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Schema and
                 subschema}; Information Systems --- Information Storage
                 and Retrieval --- Information Search and Retrieval
                 (H.3.3): {\bf Query formulation}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Relational databases}",
}

@Article{Klug:1982:DVD,
  author =       "Anthony Klug and Rod Price",
  title =        "Determining {View} dependencies using tableaux",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "361--380",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05",
  MRnumber =     "83k:68103",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p361-klug/p361-klug.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p361-klug/",
  abstract =     "A relational database models some part of the real
                 world by a set of relations and a set of constraints.
                 The constraints model properties of the stored
                 information and must be maintained true at all times.
                 For views defined over physically stored (base)
                 relations, this is done by determining whether the view
                 constraints are logical consequences of base relation
                 constraints. A technique for determining such valid
                 view constraints is presented in this paper. A
                 generalization of the tableau chase is used. The idea
                 of the method is to generate a tableau for the
                 expression whose summary violates the test constraints
                 in a ``canonical'' way. The chase then tries to remove
                 this violation.\par

                 It is also shown how this method has applications to
                 schema design. Relations not in normal form or having
                 other deficiencies can be replaced by normal form
                 projections without losing the ability to represent all
                 constraint information.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "chase; dependencies; rational algebra; relational
                 model; tableaux; views TODS",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Schema and subschema}",
}

@Article{Dayal:1982:CTU,
  author =       "Umeshwar Dayal and Philip A. Bernstein",
  title =        "On the Correct Translation of Update Operations on
                 Relational Views",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "381--416",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05",
  MRnumber =     "83k:68099",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p381-dayal/p381-dayal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p381-dayal/",
  abstract =     "Most relational database systems provide a facility
                 for supporting user views. Permitting this level of
                 abstraction has the danger, however, that update
                 requests issued by a user within the context of his
                 view may not translate correctly into equivalent
                 updates on the underlying database. The purpose of this
                 paper is to formalize the notion of update translation
                 and derive conditions under which translation
                 procedures will produce correct translations of view
                 updates.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "external schemata; relational databases; schema
                 mapping; update translation; user views",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Schema and subschema};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data manipulation languages
                 (DML)}; Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Database
                 Management --- Heterogeneous Databases (H.2.5): {\bf
                 Program translation**}; Computing Methodologies ---
                 Artificial Intelligence --- Automatic Programming
                 (I.2.2): {\bf Program transformation}",
}

@Article{Griffith:1982:TPR,
  author =       "Robert L. Griffith",
  title =        "Three Principles of Representation for Semantic
                 Networks",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "417--442",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p417-griffith/p417-griffith.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p417-griffith/",
  abstract =     "Semantic networks are so intuitive and easy to use
                 that they are often employed without much thought as to
                 the phenomenon of semantic nets themselves. Since they
                 are becoming more and more a tool of artificial
                 intelligence and now database technology, it is
                 appropriate to focus on the principles of semantic
                 nets. Such focus finds a harmonious and consistent base
                 which can increase the semantic quality and usefulness
                 of such nets. Three rules of representation are
                 presented which achieve greater conceptual simplicity
                 for users, simplifications in semantic net
                 implementations and maintenance, and greater
                 consistency across semantic net applications. These
                 rules, applied to elements of the net itself, reveal
                 how fundamental structures should be organized, and
                 show that the common labeled-edge semantic net can be
                 derived from a more primitive structure involving only
                 nodes and membership relationships (and special nodes
                 which represent names). Also, the correlation between
                 binary and $n$-ary relations is presented.",
  acknowledgement = ack-nhfb,
  annote =       "Semantic networks are employed without much thought.
                 They are becoming a tool of artificial intelligences
                 and database technology, principles of semantic nets.
                 Three rules of representation are presented. The common
                 labeled-edge semantic net can be derived from a more
                 primitive structure involving only nodes and membership
                 relationships.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "Data --- Data Structures (E.1): {\bf Graphs and
                 networks}; Computing Methodologies --- Artificial
                 Intelligence --- Knowledge Representation Formalisms
                 and Methods (I.2.4): {\bf Semantic networks}",
}

@Article{Kim:1982:OSL,
  author =       "Won Kim",
  title =        "On Optimizing an {SQL-like} Nested Query",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "443--469",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p443-kim/p443-kim.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p443-kim/",
  abstract =     "SQL is a high-level nonprocedural data language which
                 has received wide recognition in relational databases.
                 One of the most interesting features of SQL is the
                 nesting of query blocks to an arbitrary depth. An
                 SQL-like query nested to an arbitrary depth is shown to
                 be composed of five basic types of nesting. Four of
                 them have not been well understood and more work needs
                 to be done to improve their execution efficiency.
                 Algorithms are developed that transform queries
                 involving these basic types of nesting into
                 semantically equivalent queries that are amenable to
                 efficient processing by existing query-processing
                 subsystems. These algorithms are then combined into a
                 coherent strategy for processing a general nested query
                 of arbitrary complexity.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregate function; divide; join; nested query;
                 predicate; relational database; SQL queries TODS",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}",
}

@Article{Wong:1982:SAI,
  author =       "Eugene Wong",
  title =        "A Statistical Approach to Incomplete Information in
                 Database Systems",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "470--488",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05",
  MRnumber =     "83k:68108",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/bibdb.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p470-wong/p470-wong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p470-wong/",
  abstract =     "There are numerous situations in which a database
                 cannot provide a precise answer to some of the
                 questions that are posed. Sources of imprecision vary
                 and include examples such as recording errors,
                 incompatible scaling, and obsolete data. In many such
                 situations, considerable prior information concerning
                 the imprecision exists and can be exploited to provide
                 valuable information for queries to which no exact
                 answer can be given. The objective of this paper is to
                 provide a framework for doing so.",
  acknowledgement = ack-nhfb,
  annote =       "Sources of imprecision include recording errors,
                 incompatible scaling, and obsolete data. In many
                 situations considerable prior information concerning
                 the imprecision exists and can be exploited. This paper
                 provides a framework. Null values.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "incomplete information; missing values; null values;
                 TODS null values",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Schema and subschema}",
}

@Article{Zaniolo:1982:NNF,
  author =       "Carlo Zaniolo",
  title =        "A New Normal Form for the Design of Relational
                 Database Schemata",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "489--499",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05 (68B15)",
  MRnumber =     "83k:68109",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p489-zaniolo/p489-zaniolo.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p489-zaniolo/",
  abstract =     "This paper addresses the problem of database schema
                 design in the framework of the relational data model
                 and functional dependencies. It suggests that both
                 Third Normal Form (3NF) and Boyce-Codd Normal Form
                 (BCNF) supply an inadequate basis for relational schema
                 design. The main problem with 3NF is that it is too
                 forgiving and does not enforce the separation principle
                 as strictly as it should. On the other hand, BCNF is
                 incompatible with the principle of representation and
                 prone to computational complexity. Thus a new normal
                 form, which lies between these two and captures the
                 salient qualities of both is proposed. The new normal
                 form is stricter than 3NF, but it is still compatible
                 with the representation principle. First a simpler
                 definition of 3NF is derived, and the analogy of this
                 new definition to the definition of BCNF is noted. This
                 analogy is used to derive the new normal form. Finally,
                 it is proved that Bernstein's algorithm for schema
                 design synthesizes schemata that are already in the new
                 normal form.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database schema; functional dependencies; relational
                 model",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Normal forms}",
}

@Article{Lam:1982:CSA,
  author =       "K. Lam and C. T. Yu",
  title =        "A Clustered Search Algorithm Incorporating Arbitrary
                 Term Dependencies",
  journal =      j-TODS,
  volume =       "7",
  number =       "3",
  pages =        "500--508",
  month =        sep,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68H05",
  MRnumber =     "83k:68104",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/bibdb.bib; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-3/p500-lam/p500-lam.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-3/p500-lam/",
  abstract =     "The documents in a database are organized into
                 clusters, where each cluster contains similar documents
                 and a representative of these documents. A user query
                 is compared with all the representatives of the
                 clusters, and on the basis of such comparisons, those
                 clusters having many {\em close neighbors\/} with
                 respect to the query are selected for searching. This
                 paper presents an estimation of the number of close
                 neighbors in a cluster in relation to the given query.
                 The estimation takes into consideration the
                 dependencies between terms. It is demonstrated by
                 experiments that the estimate is accurate and the time
                 to generate the estimate is small.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Bahadur-Lazarsfeld expansion; clustered search; CTYU
                 TODS; generating polynomial; term dependencies",
  subject =      "Theory of Computation --- Analysis of Algorithms and
                 Problem Complexity --- General (F.2.0); Mathematics of
                 Computing --- Discrete Mathematics --- Combinatorics
                 (G.2.1): {\bf Combinatorial algorithms}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Storage (H.3.2): {\bf File organization};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Clustering}; Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3): {\bf Retrieval models}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3): {\bf Search
                 process}",
}

@Article{Batory:1982:UMP,
  author =       "D. S. Batory and C. C. Gotlieb",
  title =        "A Unifying Model of Physical Databases",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "509--539",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p509-batory/p509-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p509-batory/",
  abstract =     "A unifying model for the study of database performance
                 is proposed. Applications of the model are shown to
                 relate and extend important work concerning batched
                 searching, transposed files, index selection, dynamic
                 hash-based files, generalized access path structures,
                 differential files, network databases, and multifile
                 query processing.",
  acknowledgement = ack-nhfb,
  annote =       "See also \cite{Piwowarski:1985:CBS}.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, TODS decomposition; decomposition;
                 linksets; simple files; unifying model",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2)",
}

@Article{Aghili:1982:PGD,
  author =       "Houtan Aghili and Dennis G. Severance",
  title =        "Practical Guide to the Design of Differential Files
                 for Recovery of On-Line Databases",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "540--565",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p540-aghili/p540-aghili.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p540-aghili/",
  abstract =     "The concept of a differential file has previously been
                 proposed as an efficient means of collecting database
                 updates for on-line systems. This paper studies the
                 problem of database backup and recovery for such
                 systems, and presents an analytic model of their
                 operation. Five key design decisions are identified and
                 an optimization procedure for each is developed. A
                 design algorithm that quickly provides parameters for a
                 near-optimal differential file architecture is
                 provided.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backup and recovery; data processing; database
                 maintenance; database systems; differential files;
                 hashing functions; numerical methods; optimization;
                 reorganization",
  subject =      "Data --- Data Storage Representations (E.2);
                 Mathematics of Computing --- Numerical Analysis (G.1);
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2); Information Systems ---
                 Database Management --- Database Administration
                 (H.2.7)",
}

@Article{Larson:1982:PAL,
  author =       "Per-{\AA}ke Larson",
  title =        "Performance Analysis of Linear Hashing with Partial
                 Expansions",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "566--587",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p566-larson/p566-larson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p566-larson/",
  abstract =     "Linear hashing with partial expansions is a new file
                 organization primarily intended for files which grow
                 and shrink dynamically. This paper presents a
                 mathematical analysis of the expected performance of
                 the new scheme. The following performance measures are
                 considered: length of successful and unsuccessful
                 searches, accesses required to insert or delete a
                 record, and the size of the overflow area. The
                 performance is cyclical. For all performance measures,
                 the necessary formulas are derived for computing the
                 expected performance at any point of a cycle and the
                 average over a cycle. Furthermore, the expected worst
                 case in connection with searching is analyzed. The
                 overall performance depends on several file parameters.
                 The numerical results show that for many realistic
                 parameter combinations the performance is expected to
                 be extremely good. Even the longest search is expected
                 to be of quite reasonable length.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "dynamic hashing schemes; extendible hashing; hashing;
                 linear hashing; TODS dynamic hashing, extendible
                 hashing, TODS dynamic hashing, extendible hashing, data
                 processing",
  subject =      "Theory of Computation --- Analysis of Algorithms and
                 Problem Complexity --- Nonnumerical Algorithms and
                 Problems (F.2.2): {\bf Sorting and searching};
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Babb:1982:JNF,
  author =       "E. Babb",
  title =        "Joined Normal Form: a Storage Encoding for Relational
                 Databases",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "588--614",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p588-babb/p588-babb.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p588-babb/",
  abstract =     "A new on-line query language and storage structure for
                 a database machine is presented. By including a
                 mathematical model in the interpreter the query
                 language has been substantially simplified so that no
                 reference to relation names is necessary. By storing
                 the model as a single joined normal form (JNF) file, it
                 has been possible to exploit the powerful search
                 capability of the Content Addressable File Store (CAFS;
                 CAFS is a registered trademark of International
                 Computers Limited) database machine.",
  acknowledgement = ack-nhfb,
  annote =       "prejoining for CAFS.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CAFS; content addressing hardware; database systems;
                 functional dependencies; implication network; joined
                 normal form; joins; mathematical model; network;
                 queries; relational database; storage encoding tags;
                 storage encoding, TODS CAFS, third normal form; third
                 normal form; updates",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Query languages}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Search process}",
}

@Article{Heyman:1982:MMD,
  author =       "Daniel P. Heyman",
  title =        "Mathematical Models of Database Degradation",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "615--631",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p615-heyman/p615-heyman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p615-heyman/",
  abstract =     "As data are updated, the initial physical structure of
                 a database is changed and retrieval of specific pieces
                 of data becomes more time consuming. This phenomenon is
                 called database degradation. In this paper two models
                 of database degradation are described. Each model
                 refers to a different aspect of the problem.\par

                 It is assumed that transactions are statistically
                 independent and either add, delete, or update data. The
                 first model examines the time during which a block of
                 data is filling up. The second model examines the
                 overflows from a block of data, which essentially
                 describes the buildup of disorganization. Analytical
                 results are obtained for both models. In addition,
                 several numerical examples are presented which show
                 that the mean number of overflows grows approximately
                 linearly with time. This approximation is used to
                 devise a simple formula for the optimal time to
                 reorganize a stochastically growing database.",
  acknowledgement = ack-nhfb,
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data overflows; database degradation, TODS data
                 overflows; database systems; file organization;
                 mathematical models",
  subject =      "Computer Systems Organization --- Performance of
                 Systems (C.4): {\bf Modeling techniques}; Information
                 Systems --- Database Management --- Database
                 Administration (H.2.7): {\bf Logging and recovery}",
}

@Article{Korth:1982:DFU,
  author =       "Henry F. Korth",
  title =        "Deadlock Freedom Using Edge Locks",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "632--652",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p632-korth/p632-korth.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p632-korth/",
  abstract =     "We define a series of locking protocols for database
                 systems that all have three main features: freedom from
                 deadlock, multiple granularity, and support for general
                 collections of locking primitives. A rooted directed
                 acyclic graph is used to represent multiple
                 granularities, as in System R. Deadlock freedom is
                 guaranteed by extending the System R protocol to
                 require locks on edges of the graph in addition to the
                 locks required on nodes.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; database systems; locking;
                 serializability",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Transaction processing}",
}

@Article{Goodman:1982:TQS,
  author =       "Nathan Goodman and Oded Shmueli",
  title =        "Tree Queries: a Simple Class of Relational Queries",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "653--677",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p653-goodman/p653-goodman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p653-goodman/",
  abstract =     "One can partition the class of relational database
                 schemas into tree schemas and cyclic schemas. (These
                 are called acyclic hypergraphs and cyclic hypergraphs
                 elsewhere in the literature.) This partition has
                 interesting implications in query processing,
                 dependency theory, and graph theory.\par

                 The tree/cyclic partitioning of database schemas
                 originated with a similar partition of equijoin
                 queries. Given an arbitrary equijoin query one can
                 obtain an equivalent query that calculates the natural
                 join of all relations in (an efficiently) derived
                 database; such a query is called a natural join (NJ)
                 query. If the derived database is a tree schema the
                 original query is said to be a tree query, and
                 otherwise a cyclic query.\par

                 In this paper we analyze query processing consequences
                 of the tree/cyclic partitioning. We are able to argue,
                 qualitatively, that queries which imply a tree schema
                 are easier to process than those implying a cyclic
                 schema. Our results also extend the study of the
                 semijoin operator.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "acyclic schemes; cyclic schemas; database systems;
                 join; semijoin; tree queries; tree schemas",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}",
}

@Article{Kerschberg:1982:QOS,
  author =       "Larry Kerschberg and Peter D. Ting and S. Bing Yao",
  title =        "Query Optimization in Star Computer Networks",
  journal =      j-TODS,
  volume =       "7",
  number =       "4",
  pages =        "678--711",
  month =        dec,
  year =         "1982",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1982-7-4/p678-kerschberg/p678-kerschberg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1982-7-4/p678-kerschberg/",
  abstract =     "Query processing is investigated for relational
                 databases distributed over several computers organized
                 in a star network. Minimal response-time processing
                 strategies are presented for queries involving the
                 select, project, and join commands. These strategies
                 depend on system parameters such as communication costs
                 and different machine processing speeds; database
                 parameters such as relation cardinality and file size;
                 and query parameters such as estimates of the size and
                 number of tuples in the result relation. The optimal
                 strategies specify relation preparation processes, the
                 shipping strategy, serial or parallel processing, and,
                 where applicable, the site of join filtering and
                 merging. Strategies for optimizing select and join
                 queries have been implemented and tested.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer networks; database systems; query
                 optimization; relational database system; star computer
                 network",
  subject =      "Computer Systems Organization ---
                 Computer-Communication Networks --- Distributed Systems
                 (C.2.4): {\bf Distributed applications}; Computer
                 Systems Organization --- Computer-Communication
                 Networks --- Distributed Systems (C.2.4): {\bf
                 Distributed databases}; Computer Systems Organization
                 --- Performance of Systems (C.4): {\bf Design studies};
                 Computer Systems Organization --- Performance of
                 Systems (C.4): {\bf Modeling techniques}; Software ---
                 Operating Systems --- File Systems Management (D.4.3):
                 {\bf Distributed file systems}; Software --- Operating
                 Systems --- Organization and Design (D.4.7): {\bf
                 Distributed systems}; Information Systems --- Database
                 Management --- Physical Design (H.2.2): {\bf Access
                 methods}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Distributed databases}",
}

@Article{Maier:1983:MOS,
  author =       "David Maier and Jeffrey D. Ullman",
  title =        "Maximal Objects and the Semantics of Universal
                 Relation Databases",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "1--14",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  note =         "Also published in/as: SUNY, Stony Brook, CS, TR
                 80/016, 1980.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p1-maier/p1-maier.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p1-maier/",
  abstract =     "The universal relation concept is intended to provide
                 the database user with a simplified model in which he
                 can compose queries without regard to the underlying
                 structure of the relations in the database. Frequently,
                 the lossless join criterion provides the query
                 interpreter with the clue needed to interpret the query
                 as the user intended. However, some examples exist
                 where interpretation by the lossless-join rule runs
                 contrary to our intuition. To handle some of these
                 cases, we propose a concept called {\em maximal
                 objects}, which modifies the universal relation concept
                 in exactly those situations where it appears to go awry
                 --- when the underlying relational structure has
                 ``cycles.'' We offer examples of how the maximal object
                 concept provides intuitively correct interpretations.
                 We also consider how one might construct maximal
                 objects mechanically from purely syntactic structural
                 information --- the relation schemes and functional
                 dependencies --- about the database.",
  acknowledgement = ack-nhfb,
  annote =       "A universal relation is represented by a hypergraph.
                 If the hypergraph is cyclic, some queries can be
                 evaluated in different ways; restricting navigation to
                 few acyclic components (maximal objects) gives
                 intuitively correct answers.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "acyclic hypergraph; database systems; relational
                 database; universal relation",
  subject =      "Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Path and circuit problems};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Data models}; Information Systems
                 --- Database Management --- Languages (H.2.3): {\bf
                 Data description languages (DDL)}; Information Systems
                 --- Database Management --- Languages (H.2.3): {\bf
                 Data manipulation languages (DML)}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf Query
                 processing}",
}

@Article{Haskin:1983:OCH,
  author =       "Roger L. Haskin and Lee A. Hollaar",
  title =        "Operational Characteristics of a Hardware-Based
                 Pattern Matcher",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "15--40",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p15-haskin/p15-haskin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p15-haskin/",
  abstract =     "The design and operation of a new class of
                 hardware-based pattern matchers, such as would be used
                 in a backended database processor in a full-text or
                 other retrieval system, is presented. This recognizer
                 is based on a unique implementation technique for
                 finite state automata consisting of partitioning the
                 state table among a number of simple digital machines.
                 It avoids the problems generally associated with
                 implementing finite state machines, such as large state
                 table memories, complex control mechanisms, and state
                 encodings. Because it consists primarily of memory,
                 with its high regularity and density, needs only
                 limited static interconnections, and operates at a
                 relatively low speed, it can be easily constructed
                 using integrated circuit techniques.\par

                 After a brief discussion of other pattern-matching
                 hardware, the structure and operation of the
                 partitioned finite state automaton is given, along with
                 a simplified discussion of how the state tables are
                 partitioned. The expected performance of the resulting
                 system and the state table partitioning programs is
                 then discussed.",
  acknowledgement = ack-nhfb,
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backend processors; computer system architecture;
                 database systems; finite state automata; full text
                 retrieval systems; information science; text
                 searching",
  subject =      "Hardware --- Logic Design --- Design Styles (B.6.1):
                 {\bf Cellular arrays and automata}; Hardware ---
                 Integrated Circuits --- Types and Design Styles
                 (B.7.1): {\bf Algorithms implemented in hardware};
                 Information Systems --- Database Management ---
                 Database Machines (H.2.6); Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Search process}",
}

@Article{Sicherman:1983:AQR,
  author =       "George L. Sicherman and Wiebren {De Jonge} and Reind
                 P. {Van De Riet}",
  title =        "Answering Queries without Revealing Secrets",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "41--59",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Also published in/as: reprinted in deJonge thesis,
                 Jun. 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p41-sicherman/p41-sicherman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p41-sicherman/",
  abstract =     "Question-answering systems must often keep certain
                 information secret. This can be accomplished, for
                 example, by sometimes refusing to answer a query. Here
                 the danger of revealing a secret by refusing to answer
                 a query is investigated. First several criteria that
                 can be used to decide whether or not to answer a query
                 are developed. First several criteria that can be used
                 to decide whether or not to answer a query are
                 developed. Then it is shown which of these criteria are
                 safe if the questioner knows nothing at all about what
                 is kept secret. Furthermore, it is proved that one of
                 these criteria is safe even if the user of the system
                 knows which information is to be kept secret.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8404-296",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, statistical security TODS; keeping
                 secrets; refusal to answer; strategy",
  subject =      "Information Systems --- Models and Principles ---
                 Systems and Information Theory (H.1.1): {\bf Value of
                 information}; Information Systems --- Models and
                 Principles --- User/Machine Systems (H.1.2);
                 Information Systems --- Information Storage and
                 Retrieval --- Systems and Software (H.3.4): {\bf
                 Current awareness systems (selective dissemination of
                 information--SDI)**}; Information Systems ---
                 Information Storage and Retrieval --- Systems and
                 Software (H.3.4): {\bf Question-answering (fact
                 retrieval) systems**}; Computing Methodologies ---
                 Artificial Intelligence --- Deduction and Theorem
                 Proving (I.2.3): {\bf Answer/reason extraction}",
}

@Article{deJonge:1983:CSD,
  author =       "Wiebren de Jonge",
  title =        "Compromising Statistical Databases Responding to
                 Queries About Means",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "60--80",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Also published in/as: reprinted in Jun. 1985 thesis",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p60-de_jonge/p60-de_jonge.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p60-de_jonge/",
  abstract =     "This paper describes how to compromise a statistical
                 database which only answers queries about arithmetic
                 means for query sets whose cardinality falls in the
                 range $ [k, N - k] $, for some $k$ greater than $0$,
                 where $N$ greater than equivalent to $ 2 k $ is the
                 number of records in the database. The compromise is
                 shown to be easy and to require only a little
                 preknowledge; knowing the cardinality of just one
                 nonempty query set is usually sufficient.\par

                 This means that not only count and sum queries, but
                 also queries for arithmetic means can be extremely
                 dangerous for the security of a statistical database,
                 and that this threat must be taken into account
                 explicitly by protective measures. This seems quite
                 important from a practical standpoint: while arithmetic
                 means were known for some time to be not altogether
                 harmless, the (perhaps surprising) extent of the threat
                 is now shown.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compromise; database security; database systems;
                 security TODS; statistical databases",
  subject =      "Information Systems --- Models and Principles ---
                 Systems and Information Theory (H.1.1): {\bf Value of
                 information}; Information Systems --- Models and
                 Principles --- User/Machine Systems (H.1.2);
                 Information Systems --- Information Storage and
                 Retrieval --- Systems and Software (H.3.4): {\bf
                 Question-answering (fact retrieval) systems**}",
}

@Article{Graham:1983:FD,
  author =       "Marc H. Graham",
  title =        "Functions in Databases",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "81--109",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "85a:68036",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p81-graham/p81-graham.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p81-graham/",
  abstract =     "We discuss the objectives of including functional
                 dependencies in the definition of a relational
                 database. We find two distinct objectives. The
                 appearance of a dependency in the definition of a
                 database indicates that the states of the database are
                 to encode a function. A method based on the chase of
                 calculating the function encoded by a particular state
                 is given and compared to methods utilizing derivations
                 of the dependency. A test for deciding whether the
                 states of a schema may encode a nonempty function is
                 presented as is a characterization of the class of
                 schemas which are capable of encoding nonempty
                 functions for all the dependencies in the definition.
                 This class is the class of dependency preserving
                 schemas as defined by Beeri et al. and is strictly
                 larger than the class presented by Bernstein.\par

                 The second objective of including a functional
                 dependency in the definition of a database is that the
                 dependency be capable of constraining the states of the
                 database; that is, capable of uncovering input errors
                 made by the users. We show that this capability is
                 weaker than the first objective; thus, even
                 dependencies whose functions are everywhere empty may
                 still act as constraints. Bounds on the requirements
                 for a dependency to act as a constraint are derived.
                 \par

                 These results are founded on the notion of a weak
                 instance for a database state, which replaces the
                 universal relation instance assumption and is both
                 intuitively and computationally more nearly
                 acceptable.",
  acknowledgement = ack-nhfb,
  annote =       "Method based on the chase of calculating the function
                 is given; the dependency should constrain the states of
                 the database; many algorithms.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "functional dependencies; tableaux; TODS functional
                 dependencies, tableaux, database systems",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Normal forms}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Schema and subschema}",
}

@Article{Katz:1983:RCG,
  author =       "R. H. Katz and E. Wong",
  title =        "Resolving Conflicts in Global Storage Design Through
                 Replication",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "110--135",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p110-katz/p110-katz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p110-katz/",
  abstract =     "We present a conceptual framework in which a
                 database's intra- and interrecord set access
                 requirements are specified as a constrained assignment
                 of abstract characteristics (``evaluated,''
                 ``indexed,'' ``clustered,'' ``well-placed'') to logical
                 access paths. We derive a physical schema by choosing
                 an available storage structure that most closely
                 provides the desired access characteristics. We use
                 explicit replication of schema objects to reduce the
                 access cost along certain paths, and analyze the
                 trade-offs between increased update overhead and
                 improved retrieval access. Finally, we given an
                 algorithm to select storage structures for a CODASYL 78
                 DBTG schema, given its access requirements
                 specification.",
  acknowledgement = ack-nhfb,
  annote =       "Access path data model deduced from the
                 Entity-Relationship Model. Each function is augmented
                 with access characteristics, evaluated, indexed,
                 clustered, and well-placed.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access path selection; database systems, TODS
                 functional data model; functional data model; storage
                 structure choice",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Physical Design
                 (H.2.2): {\bf Access methods}",
}

@Article{Lomet:1983:BIE,
  author =       "David B. Lomet",
  title =        "Bounded Index Exponential Hashing",
  journal =      j-TODS,
  volume =       "8",
  number =       "1",
  pages =        "136--165",
  month =        mar,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-1/p136-lomet/p136-lomet.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-1/p136-lomet/",
  abstract =     "Bounded index exponential hashing, a new form of
                 extendible hashing, is described. It has the important
                 advantages over most of the other extendible hashing
                 variants of both (i) providing random access to any
                 record of a file in close to one disk access and (ii)
                 having performance which does not vary with file size.
                 It is straightforward to implement and demands only a
                 fixed and specifiable amount of main storage to achieve
                 this performance. Its underlying physical disk storage
                 is readily managed and record overflow is handled so as
                 to insure that unsuccessful searches never take more
                 than two accesses. The method's ability to access data
                 in close to a single disk access makes it possible to
                 organize a database, in which files have a primary key
                 and multiple secondary keys, such that the result is a
                 significant performance advantage over existing
                 organizations.",
  acknowledgement = ack-nhfb,
  classification = "722",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing, TODS extendible hashing; extendible
                 hasing; tree index methods",
  subject =      "Software --- Operating Systems --- File Systems
                 Management (D.4.3): {\bf File organization}; Data ---
                 Data Storage Representations (E.2): {\bf Hash-table
                 representations}; Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2):
                 {\bf File organization}",
}

@Article{Stonebraker:1983:PER,
  author =       "Michael Stonebraker and John Woodfill and Jeff
                 Ranstrom and Marguerite Murphy and Marc Meyer and Eric
                 Allman",
  title =        "Performance Enhancements to a Relational Database
                 System",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "167--185",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-2/p167-stonebraker/p167-stonebraker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-2/p167-stonebraker/",
  abstract =     "In this paper we examine four performance enhancements
                 to a database management system: dynamic compilation,
                 microcoded routines, a special-purpose file system, and
                 a special-purpose operating system. All were examined
                 in the context of the INGRES database management
                 system. Benchmark timings that are included suggest the
                 attractiveness of dynamic compilation and a
                 special-purpose file system. Microcode and a
                 special-purpose operating system are analyzed and
                 appear to be of more limited utility in the INGRES
                 context.",
  acknowledgement = ack-nhfb,
  annote =       "Estimates are given for compilation, micro-coding, a
                 file system which supports locality, and a specialized
                 operating system for INGRES.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "compiled query languages; database performance;
                 database systems, TODS dynamic compilation microcode
                 special purpose file operating system; file systems for
                 databases; microcode",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Query processing};
                 Information Systems --- Database Management ---
                 Database Machines (H.2.6)",
}

@Article{Garcia-Molina:1983:USK,
  author =       "H{\'e}ctor Garc{\'\i}a-Molina",
  title =        "Using Semantic Knowledge for Transaction Processing in
                 a Distributed Database",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "186--213",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-2/p186-garcia-molina/p186-garcia-molina.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-2/p186-garcia-molina/",
  abstract =     "This paper investigates how the semantic knowledge of
                 an application can be used in a distributed database to
                 process transactions efficiently and to avoid some of
                 the delays associated with failures. The main idea is
                 to allow nonserializable schedules which preserve
                 consistency and which are acceptable to the system
                 users. To produce such schedules, the transaction
                 processing mechanism receives semantic information from
                 the users in the form of transaction semantic types, a
                 division of transactions into steps, compatibility
                 sets, and countersteps. Using these notions, we propose
                 a mechanism which allows users to exploit their
                 semantic knowledge in an organized fashion. The
                 strengths and weaknesses of this approach are
                 discussed.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; consistency; database systems;
                 locking; schedule; semantic knowledge;
                 serializability",
  subject =      "Computer Systems Organization ---
                 Computer-Communication Networks --- Distributed Systems
                 (C.2.4): {\bf Distributed databases}; Information
                 Systems --- Database Management --- Systems (H.2.4);
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}",
}

@Article{Clifford:1983:FST,
  author =       "James Clifford and David S. Warren",
  title =        "Formal Semantics for Time in Databases",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "214--254",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/ai.misc.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-2/p214-clifford/p214-clifford.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-2/p214-clifford/",
  abstract =     "The concept of a historical database is introduced as
                 a tool for modeling the dynamic nature of some part of
                 the real world. Just as first-order logic has been
                 shown to be a useful formalism for expressing and
                 understanding the underlying semantics of the
                 relational database model, intensional logic is
                 presented as an analogous formalism for expressing and
                 understanding the temporal semantics involved in a
                 historical database. The various components of the
                 relational model, as extended to include historical
                 relations, are discussed in terms of the model theory
                 for the logic IL//s, a variation of the logic IL
                 formulated by Richard Montague. The modal concepts of
                 intensional and extensional data constraints and
                 queries are introduced and contrasted. Finally, the
                 potential application of these ideas to the problem of
                 natural language database querying is discussed.",
  acknowledgement = ack-nhfb,
  annote =       "all timeslices are represented. No inference needed.
                 Storage could be huge. Uses term historical db versus
                 temporal db. Intensional Montague logic. Two
                 timestamps: `state' and `exist'",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; entity-relationship model;
                 historical databases; intensional logic; relational
                 database; temporal semantics",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}",
}

@Article{Bitton:1983:DRE,
  author =       "Dina Bitton and David J. DeWitt",
  title =        "Duplicate Record Elimination in Large Data Files",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "255--265",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-2/p255-bitton/p255-bitton.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-2/p255-bitton/",
  abstract =     "The issue of duplicate elimination for large data
                 files in which many occurrences of the same record may
                 appear is addressed. A comprehensive cost analysis of
                 the duplicate elimination operation is presented. This
                 analysis is based on a combinatorial model developed
                 for estimating the size of intermediate runs produced
                 by a modified merge-sort procedure. The performance of
                 this modified merge-sort procedure is demonstrated to
                 be significantly superior to the standard duplicate
                 elimination technique of sorting followed by a
                 sequential pass to locate duplicate records. The
                 results can also be used to provide critical input to a
                 query optimizer in a relational database system.",
  acknowledgement = ack-nhfb,
  annote =       "use a modified sort-merge.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, early aggregation TODS; duplicate
                 elimination; projection operator; sorting",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}",
}

@Article{Sagiv:1983:CGC,
  author =       "Yehoshua Sagiv",
  title =        "A Characterization of Globally Consistent Databases
                 and Their Correct Access Paths",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "266--286",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/prolog.1.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-2/p266-sagiv/p266-sagiv.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-2/p266-sagiv/",
  abstract =     "The representative instance is proposed as a
                 representation of the data stored in a database whose
                 relations are not the projections of a universal
                 instance. Database schemes are characterized for which
                 local consistency implies global consistency. (Local
                 consistency means that each relation satisfies its own
                 functional dependencies; global consistency means that
                 the representative instance satisfies all the
                 functional dependencies). A method of efficiently
                 computing projections of the representative instance is
                 given, provided that local consistency implies global
                 consistency. Throughout, it is assumed that a cover of
                 the functional dependencies is embodied in the database
                 scheme in the form of keys.",
  acknowledgement = ack-nhfb,
  annote =       "Inter-relational consistency based on FD's",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "chase; database systems, TODS chase, universal
                 relation scheme, extension join, relational algebra;
                 extension join; functional dependency; null value;
                 prolog; relational algebra; relational database;
                 representative instance; universal relation scheme",
  subject =      "Theory of Computation --- Mathematical Logic and
                 Formal Languages --- Mathematical Logic (F.4.1);
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Normal forms}; Information Systems
                 --- Database Management --- Logical Design (H.2.1):
                 {\bf Schema and subschema}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Query formulation}",
}

@Article{Ullman:1983:CTJ,
  author =       "Jeffrey D. Ullman",
  title =        "Corrigendum: The Theory of Joins in Relational
                 Databases",
  journal =      j-TODS,
  volume =       "8",
  number =       "2",
  pages =        "287--287",
  month =        jun,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Graefe.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Aho:1979:TJR}.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cardenas:1983:PRA,
  author =       "Alfonso F. Cardenas and Farid Alavian and Algirdas
                 Avizienis",
  title =        "Performance of Recovery Architectures in Parallel
                 Associative Database Processors",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "291--323",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p291-cardenas/p291-cardenas.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p291-cardenas/",
  abstract =     "The need for robust recovery facilities in modern
                 database management systems is quite well known.
                 Various authors have addressed recovery facilities and
                 specific techniques, but none have delved into the
                 problem of recovery in database machines. In this
                 paper, the types of undesirable events that occur in a
                 database environment are classified and the necessary
                 recovery information, with subsequent actions to
                 recover the correct state of the database, is
                 summarized. A model of the ``processor-per-track''
                 class of parallel associative database processor is
                 presented. Three different types of recovery mechanisms
                 that may be considered for parallel associative
                 database processors are identified. For each
                 architecture, both the workload imposed by the recovery
                 mechanisms on the execution of database operations
                 (i.e., retrieve, modify, delete, and insert) and the
                 workload involved in the recovery actions (i.e.,
                 rollback, restart, restore, and reconstruct) are
                 analyzed. The performance of the three architectures is
                 quantitatively compared. This comparison is made in
                 terms of the number of extra revolutions of the
                 database area required to process a transaction versus
                 the number of records affected by a transaction. A
                 variety of different design parameters of the database
                 processor, of the database, and of a mix of transaction
                 types (modify, insert, and delete) are considered. A
                 large number of combinations is selected and the
                 effects of the parameters on the extra processing time
                 are identified.",
  acknowledgement = ack-nhfb,
  annote =       "three methods of recovery in logic-per-track
                 processors are analyzed.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative database processors; database systems,
                 hardware support machine TODS",
  subject =      "Computer Systems Organization --- Performance of
                 Systems (C.4): {\bf Performance attributes};
                 Information Systems --- Database Management ---
                 Database Machines (H.2.6); Information Systems ---
                 Database Management --- Database Administration
                 (H.2.7): {\bf Logging and recovery}",
}

@Article{Bitton:1983:PAE,
  author =       "Dina Bitton and Haran Boral and David J. DeWitt and W.
                 Kevin Wilkinson",
  title =        "Parallel Algorithms for the Execution of Relational
                 Database Operations",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "324--353",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p324-bitton/p324-bitton.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p324-bitton/",
  abstract =     "This paper presents and analyzes algorithms for
                 parallel processing of relational database operations
                 in a general multiprocessor framework. To analyze
                 alternative algorithms, we introduce an analysis
                 methodology which incorporates I/O, CPU, and message
                 costs and which can be adjusted to fit different
                 multiprocessor architectures. Algorithms are presented
                 and analyzed for sorting, projection, and join
                 operations. While some of these algorithms have been
                 presented and analyzed previously, we have generalized
                 each in order to handle the case where the number of
                 pages is significantly larger than the number of
                 processors. In addition, we present and analyze
                 algorithms for the parallel execution of update and
                 aggregate operations.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregate operations; database machines; database
                 systems; join operation; parallel processing;
                 projection operator; sorting",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- Database Machines
                 (H.2.6)",
}

@Article{Eager:1983:ARD,
  author =       "Derek L. Eager and Kenneth C. Sevcik",
  title =        "Achieving Robustness in Distributed Database Systems",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "354--381",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p354-eager/p354-eager.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p354-eager/",
  abstract =     "The problem of concurrency control in distributed
                 database systems in which site and communication link
                 failures may occur is considered. The possible range of
                 failures is not restricted; in particular, failures may
                 induce an arbitrary network partitioning. It is
                 desirable to attain a high ``level of robustness'' in
                 such a system; that is, these failures should have only
                 a small impact on system operation.\par

                 A level of robustness termed {\em maximal partial
                 operability\/} is identified. Under our models of
                 concurrency control and robustness, this robustness
                 level is the highest level attainable without
                 significantly degrading performance.\par

                 A basis for the implementation of maximal partial
                 operability is presented. To illustrate its use, it is
                 applied to a distributed locking concurrency control
                 method and to a method that utilizes timestamps. When
                 no failures are present, the robustness modifications
                 for these methods induce no significant additional
                 overhead.",
  acknowledgement = ack-nhfb,
  annote =       "Three phases: 1. read and write to a private
                 workspace, 2. indicate intention to commit, restart, or
                 abort, 3. if verified complete actual transaction.
                 Intention to update from phase 2 is withdrawn if abort
                 or restart is the end result of phase 2. Two versions
                 are presented, either the data are transmitted in phase
                 3, or, if handled as phase 2 of two-phase commit
                 protocol, the data are held in secure storage from
                 phase 2 to phase 3. In case of partitioning, voting
                 (ref. Gifford) is used, but to prevent broad lockouts,
                 those transactions which cannot update all copies must
                 post this failure at a quorum of nodes, so that
                 successor transactions can test that they do not
                 conflict, and do not enter phase 3. Dangling precommits
                 may be resolved by checking other nodes in the
                 partition. On re-establishing the net, acyclicness is
                 created by possibly restarting transactions which led
                 to a cycle in the combined schedule.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; database systems; network
                 partitioning; robustness; serializability",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2); Information Systems ---
                 Database Management --- Systems (H.2.4); Information
                 Systems --- Database Management --- Database
                 Administration (H.2.7)",
}

@Article{Trueblood:1983:MMM,
  author =       "Robert P. Trueblood and H. Rex Hartson and Johannes J.
                 Martin",
  title =        "{MULTISAFE} --- {A} Modular Multiprocessing Approach
                 to Secure Database Management",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "382--409",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p382-trueblood/p382-trueblood.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p382-trueblood/",
  abstract =     "This paper describes the configuration and intermodule
                 communication of a MULTImodule system for supporting
                 Secure Authorization with Full Enforcement (MULTISAFE)
                 for database management. A modular architecture is
                 described which provides secure, controlled access to
                 shared data in a multiuser environment, with low
                 performance penalties, even for complex protection
                 policies. The primary mechanisms are structured and
                 verifiable. The entire approach is immediately
                 extendible to distributed protection of distributed
                 data. The system includes a user and applications
                 module (UAM), a data storage and retrieval module
                 (SRM), and a protection and security module (PSM). The
                 control of intermodule communication is based on a data
                 abstraction approach, initially described in terms of
                 function invocations. An implementation within a formal
                 message system is then described. The discussion of
                 function invocations begins with the single terminal
                 case and extends to the multiterminal case. Some
                 physical implementation aspects are also discussed, and
                 some examples of message sequences are given.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data types; access control; back-end
                 database; database systems; intermodule communication;
                 secure database; TODS",
  subject =      "Computer Systems Organization --- Processor
                 Architectures --- Multiple Data Stream Architectures
                 (Multiprocessors) (C.1.2); Software --- Operating
                 Systems --- Storage Management (D.4.2); Software ---
                 Operating Systems --- Security and Protection (D.4.6);
                 Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management --- Systems
                 (H.2.4); Information Systems --- Database Management
                 --- Database Machines (H.2.6)",
}

@Article{Ito:1983:HFO,
  author =       "Tetsuro Ito and Makoto Kizawa",
  title =        "Hierarchical File Organization and its Application to
                 Similar-String Matching",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "410--433",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/spell.bib;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 https://www.math.utah.edu/pub/tex/bib/unix.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p410-ito/p410-ito.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p410-ito/",
  abstract =     "The automatic correction of misspelled inputs is
                 discussed from a viewpoint of similar-string matching.
                 First a hierarchical file organization based on a
                 linear ordering of records is presented for retrieving
                 records highly similar to any input query. Then the
                 spelling problem is attacked by constructing a
                 hierarchical file for a set of strings in a dictionary
                 of English words. The spelling correction steps proceed
                 as follows: (1) find one of the best-match strings
                 which are most similar to a query, (2) expand the
                 search area for obtaining the good-match strings, and
                 (3) interrupt the file search as soon as the required
                 string is displayed. Computational experiments verify
                 the performance of the proposed methods for
                 similar-string matching under the UNIX time-sharing
                 system.",
  acknowledgement = ack-nhfb,
  annote =       "A spelling checker to provide possible correct
                 spellings for all possible words. Results are quite
                 sketchy",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "best match; data processing, algorithms;
                 experimentation; file organization; good match;
                 hierarchical clustering; linear ordering; measurement;
                 office automation; performance; similar-string;
                 similarity; spelling correction; text editor; theory;
                 verification",
  review =       "ACM CR 8408-0665",
  subject =      "I.2 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Natural Language Processing \\ I.5.4 Computing
                 Methodologies, PATTERN RECOGNITION, Applications, Text
                 processing \\ E.5 Data, FILES, Organization/structure
                 \\ H.3.2 Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Storage, File organization \\
                 H.3.3 Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Search
                 process \\ H.3.3 Information Systems, INFORMATION
                 STORAGE AND RETRIEVAL, Information Search and
                 Retrieval, Selection process \\ H.4 Information
                 Systems, INFORMATION SYSTEMS APPLICATIONS, Office
                 Automation",
}

@Article{Kolodner:1983:IRS,
  author =       "Janet L. Kolodner",
  title =        "Indexing and Retrieval Strategies for Natural Language
                 Fact Retrieval",
  journal =      j-TODS,
  volume =       "8",
  number =       "3",
  pages =        "434--464",
  month =        sep,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/ai.misc.bib; Compendex database;
                 Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-3/p434-kolodner/p434-kolodner.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-3/p434-kolodner/",
  abstract =     "Researchers in artificial intelligence have recently
                 become interested in natural language fact retrieval;
                 currently, their research is at a point where it can
                 begin contributing to the field of Information
                 Retrieval. In this paper, strategies for a natural
                 language fact retrieval system are mapped out, and
                 approaches to many of the organization and retrieval
                 problems are presented. The CYRUS system, which keeps
                 track of important people and is queried in English, is
                 presented and used to illustrate those solutions.",
  acknowledgement = ack-nhfb,
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "artificial intelligence; conceptual memory; database
                 retrieval; fact retrieval; information science; natural
                 language processing; question answering",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Content Analysis and Indexing (H.3.1);
                 Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3);
                 Computing Methodologies --- Artificial Intelligence ---
                 Knowledge Representation Formalisms and Methods
                 (I.2.4)",
}

@Article{Bernstein:1983:MCC,
  author =       "Philip A. Bernstein and Nathan Goodman",
  title =        "Multiversion Concurrency Control --- Theory and
                 Algorithms",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "465--483",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86m:68025",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/Discrete.event.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p465-bernstein/p465-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p465-bernstein/",
  abstract =     "Concurrency control is the activity of synchronizing
                 operations issued by concurrently executing programs on
                 a shared database. The goal is to produce an execution
                 that has the same effect as a serial (noninterleaved)
                 one. In a multiversion database system, each write on a
                 data item produces a new copy (or {\em version\/}) of
                 that data item. This paper presents a theory for
                 analyzing the correctness of concurrency control
                 algorithms for multiversion database systems. We use
                 the theory to analyze some new algorithms and some
                 previously published ones.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- Algorithms; database systems;
                 transaction processing",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4)",
}

@Article{Lynch:1983:MAN,
  author =       "Nancy A. Lynch",
  title =        "Multilevel Atomicity --- {A} New Correctness Criterion
                 for Database Concurrency Control",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "484--502",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86j:68022",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p484-lynch/p484-lynch.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p484-lynch/",
  abstract =     "{\em Multilevel atomicity}, a new correctness criteria
                 for database concurrency control, is defined. It
                 weakens the usual notion of serializability by
                 permitting controlled interleaving among transactions.
                 It appears to be especially suitable for applications
                 in which the set of transactions has a natural
                 hierarchical structure based on the hierarchical
                 structure of an organization. A characterization for
                 multilevel atomicity, in terms of the absence of cycles
                 in a dependency relation among transaction steps, is
                 given. Some remarks are made concerning
                 implementation.",
  acknowledgement = ack-nhfb,
  annote =       "A weaker level of concurrency control than transaction
                 serializability, a generalization of Garc{\'\i}a-Molina
                 [1981].",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "atomicity; breakpoint; database systems; transaction",
  subject =      "Software --- Programming Languages --- Language
                 Constructs and Features (D.3.3): {\bf Concurrent
                 programming structures}",
}

@Article{Hecht:1983:SMF,
  author =       "Matthew S. Hecht and John D. Gabbe",
  title =        "Shadowed Management of Free Disk Pages with a Linked
                 List",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "503--514",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p503-hecht/p503-hecht.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p503-hecht/",
  abstract =     "We describe and prove correct a programming technique
                 using a linked list of pages for managing the free disk
                 pages of a file system where shadowing is the recovery
                 technique. Our technique requires a window of only two
                 pages of main memory for accessing and maintaining the
                 free list, and avoids wholesale copying of free-list
                 pages during a {\em checkpoint\/} or {\em recover\/}
                 operation.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "checkpoint; computer operating systems; computer
                 programming; database systems; dynamic storage
                 allocation; file system; recovery; shadowing; storage
                 management",
  subject =      "Software --- Operating Systems --- Storage Management
                 (D.4.2): {\bf Allocation/deallocation strategies};
                 Software --- Operating Systems --- Reliability (D.4.5):
                 {\bf Checkpoint/restart}",
}

@Article{Malhotra:1983:EIA,
  author =       "A. Malhotra and H. M. Markowitz and D. P. Pazel",
  title =        "{EAS-E}: An Integrated Approach to Application
                 Development",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "515--542",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p515-malhotra/p515-malhotra.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p515-malhotra/",
  abstract =     "{\em EAS-E\/} (pronounced EASY) is an experimental
                 programming language integrated with a database
                 management system now running on VM/370 at the IBM
                 Thomas J. Watson Research Center. The EAS-E programming
                 language is built around the entity, attribute, and set
                 ({\em EAS\/}) view of application development. It
                 provides a means for translating operations on EAS
                 structures directly into executable code. EAS-E
                 commands have an English-like syntax, and thus EAS-E
                 programs are easy to read and understand. EAS-E
                 programs are also more compact than equivalent programs
                 in other database languages.\par

                 The EAS-E database management system allows many users
                 simultaneous access to the database. It supports
                 locking and deadlock detection and is capable of
                 efficiently supporting network databases of various
                 sizes including very large databases, consisting of
                 several millions of entities stored on multiple DASD
                 extends. Also available is a nonprocedural facility
                 that allows a user to browse and update the database
                 without writing programs.",
  acknowledgement = ack-nhfb,
  annote =       "ER model based tool",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database systems, TODS
                 E/R model; entity relationship model",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data description languages
                 (DDL)}; Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data manipulation languages
                 (DML)}; Information Systems --- Database Management ---
                 Systems (H.2.4); Software --- Software Engineering ---
                 Programming Environments (D.2.6)",
}

@Article{Moran:1983:CDO,
  author =       "Shlomo Moran",
  title =        "On the Complexity of Designing Optimal Partial-Match
                 Retrieval Systems",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "543--551",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (68P10)",
  MRnumber =     "86j:68024",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Graphics/siggraph/83.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p543-moran/p543-moran.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p543-moran/",
  abstract =     "We consider the problem of designing an information
                 retrieval system on which partial match queries have to
                 be answered. Each record in the system consists of a
                 list of {\em attributes}, and a partial match query
                 specifies the values of some of the attributes. The
                 records are stored in {\em buckets\/} in a secondary
                 memory, and in order to answer a partial match query
                 all the buckets that may contain a record satisfying
                 the specifications of that query must be retrieved. The
                 bucket in which a given record is stored is found by a
                 multiple key hashing function, which maps each
                 attribute to a string of a fixed number of bits. The
                 address of that bucket is then represented by the
                 string obtained by concatenating the strings on which
                 the various attributes were mapped. A partial match
                 query may specify only part of the bits in the string
                 representing the address, and the larger the number of
                 bits specified, the smaller the number of buckets that
                 have to be retrieved in order to answer the query.
                 \par

                 The optimization problem considered in this paper is
                 that of deciding to how many bits each attribute should
                 be mapped by the bashing function above, so that the
                 expected number of buckets retrieved per query is
                 minimized. Efficient solutions for special cases of
                 this problem have been obtained in [1], [12], and [14].
                 It is shown that in general the problem is NP-hard, and
                 that if $P$ NP, it is also not fully approximable. Two
                 heuristic algorithms for the problem are also given and
                 compared.",
  acknowledgement = ack-nhfb,
  annote =       "Optimal variable bit lengths of hashstrings, it is
                 NP-hard.",
  classification = "723; 901; 922",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximation algorithms; computer programming ---
                 algorithms; file organization; hashing; information
                 science; NP-hard problems; optimization, TODS hashing,
                 searching; partial match retrieval; searching",
  oldlabel =     "geom-947",
  review =       "ACM CR 8411-0954",
  subject =      "Theory of Computation --- Analysis of Algorithms and
                 Problem Complexity --- Nonnumerical Algorithms and
                 Problems (F.2.2); Information Systems --- Information
                 Storage and Retrieval --- Information Search and
                 Retrieval (H.3.3)",
}

@Article{Ramamohanarao:1983:PMR,
  author =       "K. Ramamohanarao and John W. Lloyd and James A. Thom",
  title =        "Partial-Match Retrieval using Hashing and
                 Descriptors",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "552--576",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "794 538",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Graphics/siggraph/83.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p552-ramamohanarao/p552-ramamohanarao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p552-ramamohanarao/",
  abstract =     "This paper studies a partial-match retrieval scheme
                 based on hash functions and descriptors. The emphasis
                 is placed on showing how the use of a descriptor file
                 can improve the performance of the scheme. Records in
                 the file are given addresses according to hash
                 functions for each field in the record. Furthermore,
                 each page of the file has associated with it a
                 descriptor, which is a fixed-length bit string,
                 determined by the records actually present in the page.
                 Before a page is accessed to see if it contains records
                 in the answer to a query, the descriptor for the page
                 is checked. This check may show that no relevant
                 records are on the page and, hence, that the page does
                 not have to be accessed. The method is shown to have a
                 very substantial performance advantage over pure
                 hashing schemes, when some fields in the records have
                 large key spaces. A mathematical model of the scheme,
                 plus an algorithm for optimizing performance, is
                 given.",
  acknowledgement = ack-nhfb,
  classification = "723; 901; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- Algorithms; DATA PROCESSING
                 --- File Organization; descriptors; dynamic file;
                 hashing; information science; MATHEMATICAL MODELS;
                 OPTIMIZATION; optimization; partial-match retrieval",
  oldlabel =     "geom-948",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Access methods};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}",
}

@Article{Ceri:1983:CQE,
  author =       "S. Ceri and G. Pelagatti",
  title =        "Correctness of Query Execution Strategies in
                 Distributed Databases",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "577--607",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p577-ceri/p577-ceri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p577-ceri/",
  abstract =     "A major requirement of a Distributed DataBase
                 Management System (DDBMS) is to enable users to write
                 queries as though the database were not distributed
                 (distribution transparency). The DDBMS transforms the
                 user's queries into execution strategies, that is,
                 sequences of operations on the various nodes of the
                 network and of transmissions between them. An execution
                 strategy on a distributed database is correct if it
                 returns the same result as if the query were applied to
                 a nondistributed database.\par

                 This paper analyzes the correctness problem for query
                 execution strategies. A formal model, called
                 Multirelational Algebra, is used as a unifying
                 framework for this purpose. The problem of proving the
                 correctness of execution strategies is reduced to the
                 problem of proving the equivalence of two expressions
                 of Multirelational Algebra. A set of theorems on
                 equivalence is given in order to facilitate this task.
                 \par

                 The proposed approach can be used also for the
                 generation of correct execution strategies, because it
                 defines the rules which allow the transformation of a
                 correct strategy into an equivalent one. This paper
                 does not deal with the problem of evaluating equivalent
                 strategies, and therefore is not in itself a proposal
                 for a query optimizer for distributed databases.
                 However, it constitutes a theoretical foundation for
                 the design of such optimizers.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "correctness of database access; database systems;
                 distributed database access; read-only transactions;
                 relational algebra",
  subject =      "Computer Systems Organization ---
                 Computer-Communication Networks --- Distributed Systems
                 (C.2.4): {\bf Distributed databases}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Query languages}; Information Systems --- Database
                 Management --- Systems (H.2.4); Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf Query
                 processing}",
}

@Article{Biskup:1983:FCR,
  author =       "Joachim Biskup",
  title =        "A Foundation of {Codd}'s Relational Maybe Operators",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "608--636",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (03B70)",
  MRnumber =     "86j:68019",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p608-biskup/p608-biskup.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p608-biskup/",
  abstract =     "Database relations which possibly contain maybe-tuples
                 and null values of type ``value at present unknown''
                 are studied. Maybe-tuples and null values are formally
                 interpreted by our notion of {\em representation},
                 which uses classical notions of predicate logic,
                 elaborates Codd's proposal of maybe-tuples, and adopts
                 Reiter's concept of a closed world. Precise notions of
                 {\em information content\/} and {\em redundancy},
                 associated with our notion of representation, are
                 investigated. {\em Extensions of the relational
                 algebra\/} to relations with maybe-tuples and null
                 values are proposed. Our extensions are essentially
                 Codd's, with some modifications. It is proved that
                 these extensions have natural properties which are
                 formally stated as being {\em adequate\/} and {\em
                 restricted}.\par

                 By the treatment of difference and division, our formal
                 framework can be used even for operations that require
                 ``negative information.'' Finally, extensions of {\em
                 update operations\/} are discussed.",
  acknowledgement = ack-nhfb,
  annote =       "Join, Project, Select, Union, Difference is defined
                 for nulls which are not labeled, results include Maybe.
                 Division does not work. Update is discussed.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "closed world assumption; database systems; information
                 content; maybe-tuple; negative information; null value;
                 open word assumption; redundancy; relational algebra;
                 relational database; representation",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Query processing}; Computing Methodologies ---
                 Artificial Intelligence --- Knowledge Representation
                 Formalisms and Methods (I.2.4): {\bf Predicate logic}",
}

@Article{Ullman:1983:KCA,
  author =       "Jeffrey D. Ullman",
  title =        "On {Kent}'s {``Consequences of assuming a universal
                 relation''} ({Technical} correspondence)",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "637--643",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/database.bib; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Kent:1981:CAU}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p637-ullman/p637-ullman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p637-ullman/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "TODS technical correspondence",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0)",
}

@Article{Kent:1983:URR,
  author =       "William Kent",
  title =        "The universal relation revisited (technical
                 correspondence)",
  journal =      j-TODS,
  volume =       "8",
  number =       "4",
  pages =        "644--648",
  month =        dec,
  year =         "1983",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/database.bib; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1983-8-4/p644-kent/p644-kent.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1983-8-4/p644-kent/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "TODS technical correspondence",
  subject =      "Information Systems --- Database Management ---
                 General (H.2.0)",
}

@Article{Kaplan:1984:DPN,
  author =       "S. Jerrold Kaplan",
  title =        "Designing a Portable Natural Language Database Query
                 System",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "1--19",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/ai.misc.bib; Compendex database;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p1-kaplan/p1-kaplan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p1-kaplan/",
  abstract =     "One barrier to the acceptance of natural language
                 database query systems is the substantial installation
                 effort required for each new database. Much of this
                 effort involves the encoding of semantic knowledge for
                 the domain of discourse, necessary to correctly
                 interpret and respond to natural language questions.
                 For such systems to be practical, techniques must be
                 developed to increase their portability to new domains.
                 \par

                 This paper discusses several issues involving the
                 portability of natural language interfaces to database
                 systems, and presents the approach taken in {\em
                 CO-OP\/} -- a natural language database query system
                 that provides cooperative responses to English
                 questions and operates with a typical CODA-SYL database
                 system. {\em CO-OP\/} derives its domain-specific
                 knowledge from a {\em lexicon\/} (the list of words
                 known to the system) and the information already
                 present in the structure and content of the underlying
                 database. Experience with the implementation suggests
                 that strategies that are not directly derivative of
                 cognitive or linguistic models may nonetheless play an
                 important role in the development of practical natural
                 language systems.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems",
}

@Article{Reiss:1984:PDS,
  author =       "Steven P. Reiss",
  title =        "Practical Data-Swapping: The First Steps",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "20--37",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p20-reiss/p20-reiss.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p20-reiss/",
  abstract =     "The problem of statistical database confidentiality in
                 releasing microdata is addressed through the use of
                 approximate data-swapping. Here, a portion of the
                 microdata is replaced with a database that has been
                 selected with approximately the same statistics. The
                 result guarantees the confidentiality of the original
                 data, while providing microdata with accurate
                 statistics. Methods for achieving such transformations
                 are considered and analyzed through simulation.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, TODS statistical databases",
  subject =      "Data --- Coding and Information Theory (E.4): {\bf
                 Nonsecret encoding schemes**}; Information Systems ---
                 Models and Principles --- Systems and Information
                 Theory (H.1.1): {\bf Value of information}; Information
                 Systems --- Information Storage and Retrieval ---
                 Online Information Services (H.3.5): {\bf Data
                 sharing}",
}

@Article{Nievergelt:1984:GFA,
  author =       "J. Nievergelt and Hans Hinterberger and Kenneth C.
                 Sevcik",
  title =        "The Grid File: An Adaptable, Symmetric Multikey File
                 Structure",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "38--71",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p38-nievergelt/p38-nievergelt.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p38-nievergelt/",
  abstract =     "Traditional file structures that provide multikey
                 access to records, for example, inverted files, are
                 extensions of file structures originally designed for
                 single-key access. They manifest various deficiencies
                 in particular for multikey access to highly dynamic
                 files. We study the dynamic aspects of file structures
                 that treat all keys symmetrically, that is, file
                 structures which avoid the distinction between primary
                 and secondary keys. We start from a bitmap approach and
                 treat the problem of file design as one of data
                 compression of a large sparse matrix. This leads to the
                 notions of a {\em grid partition\/} of the search space
                 and of a {\em grid directory}, which are the keys to a
                 dynamic file structure called the {\em grid file}. This
                 file system adapts gracefully to its contents under
                 insertions and deletions, and thus achieves an upper
                 bound of two disk accesses for single record retrieval;
                 it also handles range queries and partially specified
                 queries efficiently. We discuss in detail the design
                 decisions that led to the grid file, present simulation
                 results of its behavior, and compare it to other
                 multikey access file structures.",
  acknowledgement = ack-nhfb,
  annote =       "Grid files use a vector of hash-keys, partition the
                 result into clusters, and store the clusters into
                 blocks. Two accesses are used for retrieval. Update may
                 be more costly. Access structures fit in core?",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing, TODS multidimensional hashing,
                 multidimensional hashing",
  review =       "ACM CR 8411-0931",
}

@Article{Buchanan:1984:DMS,
  author =       "Jack R. Buchanan and Richard D. Fennell and Hanan
                 Samet",
  title =        "A Database Management System for the {Federal
                 Courts}",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "72--88",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p72-buchanan/p72-buchanan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p72-buchanan/",
  abstract =     "A judicial systems laboratory has been established and
                 several large-scale information management systems
                 projects have been undertaken within the Federal
                 Judicial Center in Washington, D.C. The newness of the
                 court application area, together with the experimental
                 nature of the initial prototypes, required that the
                 system building tools be as flexible and efficient as
                 possible for effective software design and development.
                 The size of the databases, the expected transaction
                 volumes, and the long-term value of the court records
                 required a data manipulation system capable of
                 providing high performance and integrity. The resulting
                 design criteria, the programming capabilities
                 developed, and their use in system construction are
                 described herein. This database programming facility
                 has been especially designed as a technical management
                 tool for the database administrator, while providing
                 the applications programmer with a flexible database
                 software interface for high productivity.
                 \par

                 Specifically, a network-type database management system
                 using SAIL as the data manipulation host language is
                 described. Generic data manipulation verb formats using
                 SAIL's macro facilities and dynamic data structuring
                 facilities allowing in-core database representations
                 have been developed to achieve a level of flexibility
                 not usually attained in conventional database
                 systems.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, TODS SAIL, network model",
}

@Article{Papadimitriou:1984:CCM,
  author =       "Christos H. Papadimitriou and Paris C. Kanellakis",
  title =        "On Concurrency Control by Multiple Versions",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "89--99",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p89-papadimitriou/p89-papadimitriou.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p89-papadimitriou/",
  abstract =     "We examine the problem of concurrency control when the
                 database management system supports multiple versions
                 of the data. We characterize the limit of the
                 parallelism achievable by the multiversion approach and
                 demonstrate the resulting space-parallelism
                 trade-off.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems",
}

@Article{Shultz:1984:RTA,
  author =       "Roger K. Shultz and Roy J. Zingg",
  title =        "Response Time Analysis of Multiprocessor Computers for
                 Database Support",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "100--132",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p100-shultz/p100-shultz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p100-shultz/",
  abstract =     "Comparison of three multiprocessor computer
                 architectures for database support is made possible
                 through evaluation of response time expressions. These
                 expressions are derived by parameterizing algorithms
                 performed by each machine to execute a relational
                 algebra query. Parameters represent properties of the
                 database and components of the machines. Studies of
                 particular parameter values exhibit response times for
                 conventional machine technology, for low selectivity,
                 high duplicate occurrence, and parallel disk access,
                 increasing the number of processors, and improving
                 communication and processing technology.",
  acknowledgement = ack-nhfb,
  annote =       "analyzes DIRECT, HYPERTREE, and REPT, their own
                 proposal.",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "COMPUTER SYSTEMS, DIGITAL --- multiprocessing, TODS
                 relational model, relational queries, direct,
                 hypertree, rept; database systems",
}

@Article{Valduriez:1984:JSA,
  author =       "Patrick Valduriez and Georges Gardarin",
  title =        "Join and Semijoin Algorithms for a Multiprocessor
                 Database Machine",
  journal =      j-TODS,
  volume =       "9",
  number =       "1",
  pages =        "133--161",
  month =        mar,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-1/p133-valduriez/p133-valduriez.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-1/p133-valduriez/",
  abstract =     "This paper presents and analyzes algorithms for
                 computing joins and semijoins of relations in a
                 multiprocessor database machine. First, a model of the
                 multiprocessor architecture is described, incorporating
                 parameters defining I/O, CPU, and message transmission
                 times that permit calculation of the execution times of
                 these algorithms. Then, three join algorithms are
                 presented and compared. It is shown that, for a given
                 configuration, each algorithm has an application domain
                 defined by the characteristics of the operand and
                 result relations. Since a semijoin operator is useful
                 for decreasing I/O and transmission times in a
                 multiprocessor system, we present and compare two
                 equi-semijoin algorithms and one non-equi-semijoin
                 algorithm. The execution times of these algorithms are
                 generally linearly proportional to the size of the
                 operand and result relations, and inversely
                 proportional to the number of processors. We then
                 compare a method which consists of joining two
                 relations to a method whereby one joins their
                 semijoins. Finally, it is shown that the latter method,
                 using semijoins, is generally better. The various
                 algorithms presented are implemented in the SABRE
                 database system; an evaluation model selects the best
                 algorithm for performing a join according to the
                 results presented here. A first version of the SABRE
                 system is currently operational at INRIA.",
  acknowledgement = ack-nhfb,
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "COMPUTER SYSTEMS, DIGITAL --- Multiprocessing;
                 database systems",
}

@Article{Christodoulakis:1984:ICA,
  author =       "S. Christodoulakis",
  title =        "Implications of Certain Assumptions in Database
                 Performance Evaluation",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "163--186",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86k:68011",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p163-christodoulakis/p163-christodoulakis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p163-christodoulakis/",
  abstract =     "The assumptions of uniformity and independence of
                 attribute values in a file, uniformity of queries,
                 constant number of records per block, and random
                 placement of qualifying records among the blocks of a
                 file are frequently used in database performance
                 evaluation studies. In this paper we show that these
                 assumptions often result in predicting only an upper
                 bound of the expected system cost. We then discuss the
                 implications of nonrandom placement, nonuniformity, and
                 dependencies of attribute values on database design and
                 database performance evaluation.",
  acknowledgement = ack-nhfb,
  annote =       "After a somewhat cursory reading of the paper --- A
                 few comments: A. Uniform distributions, particularly
                 for parallel machines, may imply uniform distribution
                 of work over the machines --- thus giving an upper
                 bound for speedup --- thus a uniform distribution is an
                 optimistic assumption. B. For uniprocessor systems --
                 the assumption of a uniform dist. is optimistic when:
                 1. hashing --- fewer collisions, shorter lists at
                 collisions should be expected from a uniform
                 distribution. 2. trees --- more balancing costs may be
                 incurred for non-uniform distributions. 3. searching
                 --- for example, a binary search on a non-uniform could
                 cost significantly more. 4. sorting --- I suspect that
                 uniform distributions are optimal for some sorting
                 methods, although I haven't looked at this in any
                 detail.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, selectivity cost estimation approx
                 TODS",
}

@Article{Effelsberg:1984:LIP,
  author =       "Wolfgang Effelsberg and Mary E. S. Loomis",
  title =        "Logical, Internal, and Physical Reference Behavior in
                 {CODASYL} Database Systems",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "187--213",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p187-effelsberg/p187-effelsberg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p187-effelsberg/",
  abstract =     "This work investigates one aspect of the performance
                 of CODASYL database systems: the data reference
                 behavior. We introduce a model of database traversals
                 at three levels: the logical, internal, and physical
                 levels. The mapping between the logical and internal
                 levels is defined by the internal schema, whereas the
                 mapping between the internal and the physical levels
                 depends on cluster properties of the database. Our
                 model explains the physical reference behavior for a
                 given sequence of DML statements at the logical level.
                 \par

                 Software has been implemented to monitor references in
                 two selected CODASYL DBMS applications. In a series of
                 experiments the physical reference behavior was
                 observed for varying internal schemas and cluster
                 properties of the database. The measurements were
                 limited to retrieval transactions, so that a variety of
                 queries could be analyzed for the same well-known state
                 of the database. Also, all databases were relatively
                 small in order to allow fast reloading with varying
                 internal schema parameters. In all cases, the database
                 transactions showed less locality of reference than do
                 programs under virtual memory operating systems; some
                 databases showed no locality at all. No evidence of
                 physical sequentiality was found. This suggests that
                 standard page replacement strategies are not optimal
                 for CODASYL database buffer management; instead,
                 replacement decisions in a database buffer should be
                 based on specific knowledge available from higher
                 system layers.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8506 0534",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems, TODS buffer management",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Computer Systems Organization ---
                 Performance of Systems (C.4): {\bf Measurement
                 techniques}; Software --- Operating Systems --- Storage
                 Management (D.4.2): {\bf Storage hierarchies}; Software
                 --- Operating Systems --- Performance (D.4.8): {\bf
                 Measurements}; Software --- Operating Systems ---
                 Performance (D.4.8): {\bf Modeling and prediction};
                 Information Systems --- Database Management ---
                 Physical Design (H.2.2); Information Systems ---
                 Database Management --- Logical Design (H.2.1)",
}

@Article{Kim:1984:PPR,
  author =       "Won Kim and Daniel Gajski and David J. Kuck",
  title =        "A Parallel Pipelined Relational Query Processor",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "214--242",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p214-kim/p214-kim.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p214-kim/",
  abstract =     "This paper presents the design of a relational query
                 processor. The query processor consists of only four
                 processing PIPEs and a number of random-access memory
                 modules. Each PIPE processes tuples of relations in a
                 bit-serial, tuple-parallel manner for each of the
                 primitive database operations which comprise a complex
                 relational query. The design of the query processor
                 meets three major objectives: the query processor must
                 be manufacturable using existing and near-term LSI
                 (VLSI) technology; it must support in a uniform manner
                 both the numeric and nonnumeric processing requirements
                 a high-level user interface like SQL presents; and it
                 must support the query-processing strategy derived in
                 the query optimizer to satisfy certain system-wide
                 performance optimality criteria.",
  acknowledgement = ack-nhfb,
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer architecture; database systems; pipeline
                 processing; relational query processor, hardware
                 support database machine TODS",
  subject =      "Computer Systems Organization --- Processor
                 Architectures --- Other Architecture Styles (C.1.3):
                 {\bf High-level language architectures**}; Information
                 Systems --- Database Management --- Database Machines
                 (H.2.6)",
}

@Article{Al-Suwaiyel:1984:ATC,
  author =       "M. Al-Suwaiyel and E. Horowitz",
  title =        "Algorithms for Trie Compaction",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "243--263",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P05",
  MRnumber =     "794 541",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p243-ai-suwaiyel/",
  abstract =     "The trie data structure has many properties which make
                 it especially attractive for representing large files
                 of data. These properties include fast retrieval time,
                 quick unsuccessful search determination, and finding
                 the longest match to a given identifier. The main
                 drawback is the space requirement. In this paper the
                 concept of trie compaction is formalized. An exact
                 algorithm for optimal trie compaction and three
                 algorithms for approximate trie compaction are given,
                 and an analysis of the three algorithms is done. The
                 analyses indicate that for actual tries, reductions of
                 around 70 percent in the space required by the
                 uncompacted trie can be expected. The quality of the
                 compaction is shown to be insensitive to the number of
                 nodes, while a more relevant parameter is the alphabet
                 size of the key.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing",
  subject =      "Data --- Data Structures (E.1); Data --- Data Storage
                 Representations (E.2); Data --- Coding and Information
                 Theory (E.4): {\bf Data compaction and compression};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2); Theory of
                 Computation --- Analysis of Algorithms and Problem
                 Complexity --- Nonnumerical Algorithms and Problems
                 (F.2.2)",
}

@Article{Mendelzon:1984:DST,
  author =       "Alberto O. Mendelzon",
  title =        "Database States and Their Tableaux",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "264--282",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "794 542",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p264-mendelzon/p264-mendelzon.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p264-mendelzon/",
  abstract =     "Recent work considers a database state to satisfy a
                 set of dependencies if there exists a satisfying
                 universal relation whose projections contain each of
                 the relations in the state. Such relations are called
                 {\em weak instances\/} for the state. We propose the
                 set of all weak instances for a state as an embodiment
                 of the information represented by the state. We
                 characterize states that have the same set of weak
                 instances by the equivalence of their associated
                 tableaux. We apply this notion to the comparison of
                 database schemes and characterize all pairs of schemes
                 such that for every legal state of one of them there
                 exists an equivalent legal state of the other one. We
                 use this approach to provide a new characterization of
                 Boyce-Codd Normal Form relation schemes.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "TODS weak instance assumption, database systems",
}

@Article{Maier:1984:FUR,
  author =       "David Maier and Jeffrey D. Ullman and Moshe Y. Vardi",
  title =        "On the Foundations of the Universal Relation Model",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "283--308",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86m:68031",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p283-maier/p283-maier.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p283-maier/",
  abstract =     "Two fundamentally different approaches to the
                 universal relation model have been taken. According to
                 the first approach, the user's view of the database is
                 a universal relation or many universal relations, about
                 which the user poses queries. The second approach sees
                 the model as having query-processing capabilities that
                 relieve the user of the need to specify the logical
                 access path. Thus, while the first approach gives a
                 denotational semantics to query answering, the second
                 approach gives it an operational semantics. The authors
                 investigate the relationship between these two
                 approaches.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Query languages}; Computing Methodologies ---
                 Artificial Intelligence --- Deduction and Theorem
                 Proving (I.2.3); Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Normal
                 forms}; Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}",
}

@Article{Katsuno:1984:ECF,
  author =       "Hirofumi Katsuno",
  title =        "An Extension of Conflict-free Multi-valued Dependency
                 Sets",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "309--326",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86m:68029",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-2/p309-katsuno/p309-katsuno.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-2/p309-katsuno/",
  abstract =     "Several researchers (Beeri, Bernstein, Chiu, Fagin,
                 Goodman, Maier, Mendelzon, Ullman, and Yannakakis) have
                 introduced a special class of database schemes, called
                 {\em acyclic\/} or {\em tree\/} schemes. Beeri et al.
                 have shown that an acyclic join dependency, naturally
                 defined by an acyclic database scheme, has several
                 desirable properties, and that an acyclic join
                 dependency is equivalent to a conflict-free set of
                 multivalued dependencies. However, since their results
                 are confined to multivalued and join dependencies, it
                 is not clear whether we can handle functional
                 dependencies independently of other dependencies.
                 \par

                 In the present paper we define an extension of a
                 conflict-free set, called an {\em extended
                 conflict-free set}, including multivalued dependencies
                 and functional dependencies, and show the following two
                 properties of an extended conflict-free set:\par

                 There are three equivalent definitions of an extended
                 conflict-free set. One of them is defined as a set
                 including an acyclic joint dependency and a set of
                 functional dependencies such that the left and right
                 sides of each functional dependency are included in one
                 of the attribute sets that construct the acyclic join
                 dependency.\par

                 For a relation scheme with an extended conflict-free
                 set, there is a decomposition into third normal form
                 with a lossless join and preservation of
                 dependencies.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Schema and subschema};
                 Information Systems --- Database Management --- Logical
                 Design (H.2.1): {\bf Normal forms}",
}

@Article{Korth:1984:SUD,
  author =       "Henry F. Korth and Gabriel M. Kuper and Joan
                 Feigenbaum and Allen {Van Gelder} and Jeffrey D.
                 Ullman",
  title =        "{System/U}: a Database System Based on the Universal
                 Relation Assumption",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "331--347",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p331-korth/p331-korth.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p331-korth/",
  abstract =     "System/U is a universal relation database system under
                 development at Stanford University which uses the
                 language C on UNIX. The system is intended to test the
                 use of the universal view, in which the entire database
                 is seen as one relation. This paper describes the
                 theory behind System/U, in particular the theory of
                 maximal objects and the connection between a set of
                 attributes. We also describe the implementation of the
                 DDL (Data Description Language) and the DML (Data
                 Manipulation Language), and discuss in detail how the
                 DDL finds maximal objects and how the DML determines
                 the connection between the attributes that appear in a
                 query.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database systems;
                 RELATIONAL DATABASE; SYSTEM/U; UNIVERSAL RELATION",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Data description languages (DDL)}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Data manipulation languages (DML)}; Information
                 Systems --- Database Management --- Physical Design
                 (H.2.2): {\bf Access methods}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf Query
                 processing}; Mathematics of Computing --- Discrete
                 Mathematics --- Graph Theory (G.2.2): {\bf Path and
                 circuit problems}",
}

@Article{Wald:1984:RQI,
  author =       "Joseph A. Wald and Paul G. Sorenson",
  title =        "Resolving the Query Inference Problem Using {Steiner}
                 Trees",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "348--368",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Graphics/siggraph/86.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p348-wald/p348-wald.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p348-wald/",
  abstract =     "The query inference problem is to translate a sentence
                 of a query language into an unambiguous representation
                 of a query. A query is represented as an expression
                 over a set of query trees. A metric is introduced for
                 measuring the complexity of a query and also a proposal
                 that a sentence be translated into the least complex
                 query which `satisfies' the sentence. This method of
                 query inference can be used to resolve ambiguous
                 sentences and leads to easier formulation of
                 sentences.",
  acknowledgement = ack-nhfb,
  annote =       "MDCST resolves queries over attributes using a schema
                 tree.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; query inference problem; Steiner
                 trees, TODS E/R model",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Transaction processing}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Query formulation};
                 Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Graph algorithms};
                 Mathematics of Computing --- Discrete Mathematics ---
                 Graph Theory (G.2.2): {\bf Trees}; Information Systems
                 --- Database Management --- Languages (H.2.3): {\bf
                 Query languages}",
}

@Article{Ramamohanarao:1984:RLH,
  author =       "K. Ramamohanarao and R. Sacks-Davis",
  title =        "Recursive Linear Hashing",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "369--391",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "794 545",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p369-ramamohanarao/p369-ramamohanarao.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p369-ramamohanarao/",
  abstract =     "A modification of linear hashing is proposed for which
                 the conventional use of overflow records is avoided.
                 Furthermore, an implementation of linear hashing is
                 presented for which the amount of physical storage
                 claimed is only fractionally more than the minimum
                 required. This implementation uses a fixed amount of
                 in-core space. Simulation results are given which
                 indicate that even for storage utilizations approaching
                 95 percent, the average successful search cost for this
                 method is close to one disk access.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "COMPUTER SIMULATION; data processing; DYNAMIC FILES;
                 LINEAR HASHING",
  subject =      "Data --- Data Storage Representations (E.2): {\bf
                 Hash-table representations}; Software --- Operating
                 Systems --- File Systems Management (D.4.3): {\bf File
                 organization}; Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2):
                 {\bf File organization}; Computing Methodologies ---
                 Simulation and Modeling --- Applications (I.6.3);
                 Software --- Operating Systems --- Storage Management
                 (D.4.2): {\bf Secondary storage}",
}

@Article{Cooper:1984:ATU,
  author =       "Robert B. Cooper and Martin K. Solomon",
  title =        "The Average Time Until Bucket Overflow",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "392--408",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p392-cooper/p392-cooper.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p392-cooper/",
  abstract =     "It is common for file structures to be divided into
                 equal-length partitions, called buckets, into which
                 records arrive for insertion and from which records are
                 physically deleted. We give a simple algorithm which
                 permits calculation of the average time until overflow
                 for a bucket of capacity $n$ records, assuming that
                 record insertions and deletions can be modeled as a
                 stochastic process in the usual manner of queueing
                 theory. We present some numerical examples, from which
                 we make some general observations about the
                 relationships among insertion and deletion rates,
                 bucket capacity, initial fill, and average time until
                 overflow. In particular, we observe that it makes sense
                 to define the {\em stable point\/} as the product of
                 the arrival rate and the average residence time of the
                 records; then a bucket tends to fill up to its stable
                 point quickly, in an amount of time almost independent
                 of the stable point, but the average time until
                 overflow increases rapidly with the difference between
                 the bucket capacity and the stable point.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "BUCKET OVERFLOW; data processing",
}

@Article{March:1984:SER,
  author =       "Salvatore T. March and Gary D. Scudder",
  title =        "On the Selection of Efficient Record Segmentations and
                 Backup Strategies for Large Shared Databases",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "409--438",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p409-march/p409-march.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p409-march/",
  abstract =     "In recent years the information processing
                 requirements of business organizations have expanded
                 tremendously. With this expansion, the design of
                 databases to efficiently manage and protect business
                 information has become critical. We analyze the impacts
                 of {\em record segmentation\/} (the assignment of data
                 items to segments defining subfiles), an
                 efficiency-oriented design technique, and of {\em
                 backup and recovery strategies}, a data protection
                 technique, on the overall process of database design. A
                 combined record segmentation/backup and recovery
                 procedure is presented and an application of the
                 procedure is discussed. Results in which problem
                 characteristics are varied along three dimensions:
                 update frequencies, available types of access paths,
                 and the predominant type of data retrieval that must be
                 supported by the database, are presented.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "backup strategies; database systems; large shared
                 databases; record segmentations",
  subject =      "Data --- Files (E.5): {\bf Organization/structure};
                 Information Systems --- Information Storage and
                 Retrieval --- Information Storage (H.3.2): {\bf File
                 organization}; Software --- Operating Systems ---
                 Reliability (D.4.5): {\bf Backup procedures}; Software
                 --- Operating Systems --- File Systems Management
                 (D.4.3): {\bf File organization}; Data --- Files (E.5):
                 {\bf Backup/recovery}; Information Systems --- Database
                 Management --- Database Administration (H.2.7): {\bf
                 Logging and recovery}",
}

@Article{Manber:1984:CCD,
  author =       "Udi Manber and Richard E. Ladner",
  title =        "Concurrency Control in a Dynamic Search Structure",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "439--455",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68N25 (68P15)",
  MRnumber =     "794 546",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: ACM Proc. on Database Systems,
                 Boston, Apr. 1982, pp. 268--282.",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p439-manbar/",
  abstract =     "A design of a data structure and efficient algorithms
                 for concurrent manipulations of a dynamic search
                 structure by independent user processes is presented in
                 this paper. The algorithms include updating data,
                 inserting new elements, and deleting elements. The
                 algorithms support a high level of concurrency. Each of
                 the operations listed above requires only constant
                 amount of locking. In order to make the system even
                 more efficient for the user processes, maintenance
                 processes are introduced. The maintenance processes
                 operate independently in the background to reorganize
                 the data structure and ``clean up'' after the (more
                 urgent) user processes. A proof of correctness of the
                 algorithms is given and some experimental results and
                 extensions are examined.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- Algorithms; concurrency
                 control; data processing; dynamic search structure",
}

@Article{Davidson:1984:OCP,
  author =       "Susan B. Davidson",
  title =        "Optimism and Consistency in Partitioned Distributed
                 Database Systems",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "456--481",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "794 547",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p456-davidson/p456-davidson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p456-davidson/",
  abstract =     "A protocol for transaction processing during partition
                 failures is presented which guarantees mutual
                 consistency between copies of data-items after repair
                 is completed. The protocol is ``optimistic'' in that
                 transactions are processed without restrictions during
                 failure; conflicts are then detected at repair time
                 using a {\em precedence graph}, and are resolved by
                 backing out transactions according to some {\em backout
                 strategy}. The resulting database state then
                 corresponds to a serial execution of some subset of
                 transactions run during the failure. Results from
                 simulation and probabilistic modeling show that the
                 optimistic protocol is a reasonable alternative in many
                 cases. Conditions under which the protocol performs
                 well are noted, and suggestions are made as to how
                 performance can be improved. In particular, a backout
                 strategy is presented which takes into account
                 individual transaction costs and attempts to minimize
                 total backout cost. Although the problem of choosing
                 transactions to minimize total backout cost is, in
                 general, NP-complete, the backout strategy is efficient
                 and produces very good results.",
  acknowledgement = ack-nhfb,
  annote =       "Counter example to uniform is last hypothesis of
                 Christodoulakis.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; DISTRIBUTED database systems",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}; Information
                 Systems --- Database Management --- Physical Design
                 (H.2.2): {\bf Recovery and restart}; Mathematics of
                 Computing --- Discrete Mathematics --- Graph Theory
                 (G.2.2)",
}

@Article{Ibaraki:1984:ONO,
  author =       "Toshihide Ibaraki and Tiko Kameda",
  title =        "On the Optimal Nesting Order for Computing $ {N}
                 $-Relational Joins",
  journal =      j-TODS,
  volume =       "9",
  number =       "3",
  pages =        "482--502",
  month =        sep,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "794 548",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-3/p482-ibaraki/p482-ibaraki.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-3/p482-ibaraki/",
  abstract =     "Using the nested loops method, this paper addresses
                 the problem of minimizing the number of page fetches
                 necessary to evaluate a given query to a relational
                 database. We first propose a data structure whereby the
                 number of page fetches required for query evaluation is
                 substantially reduced and then derive a formula for the
                 expected number of page fetches. An optimal solution to
                 our problem is the nesting order of relations in the
                 evaluation program, which minimizes the number of page
                 fetches. Since the minimization of the formula is
                 NP-hard, as shown in the Appendix, we propose a
                 heuristic algorithm which produces a good suboptimal
                 solution in polynomial time. For the special case where
                 the input query is a ``tree query,'' we present an
                 efficient algorithm for finding an optimal nesting
                 order.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8506 0535",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "N-relational joins; optimal nesting order; TODS query
                 optimization processing relational model, database
                 systems",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Transaction processing}; Theory
                 of Computation --- Analysis of Algorithms and Problem
                 Complexity --- Nonnumerical Algorithms and Problems
                 (F.2.2); Information Systems --- Database Management
                 --- Logical Design (H.2.1): {\bf Data models};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Query processing}",
}

@Article{Elhardt:1984:DCH,
  author =       "Klaus Elhardt and Rudolf Bayer",
  title =        "A Database Cache for High Performance and Fast Restart
                 in Database Systems",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "503--525",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p503-elhardt/p503-elhardt.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p503-elhardt/",
  abstract =     "Performance in database systems is strongly influenced
                 by buffer management and transaction recovery methods.
                 This paper presents the principles of the database
                 cache, which replaces the traditional buffer. In
                 comparison to buffer management, cache management is
                 more carefully coordinated with transaction management,
                 and integrates transaction recovery. High throughput of
                 \par

                 small- and medium-sized transactions is achieved by
                 fast commit processing and low database traffic. Very
                 fast handling of transaction failures and short restart
                 time after system failure are guaranteed in such an
                 environment. Very long retrieval and update
                 transactions are also supported.",
  acknowledgement = ack-nhfb,
  affiliation =  "Technische Univ, Muenchen, Inst fuer Informatik,
                 Munich, West Ger",
  affiliationaddress = "Technische Univ, Muenchen, Inst fuer Informatik,
                 Munich, West Ger",
  annote =       "The Elhardt-Bayer cache does indeed resemble Alpine in
                 many important respects. The primary difference is that
                 it requires all of a transaction's updates to be
                 written to the log in contiguous log pages, which
                 allows some compact encodings to be used in
                 representing the log, but also means that more work
                 must be done synchronously at commit time. Also, their
                 scheme is not designed to support two-phase commit, and
                 extending it to handle two-phase commit is sure to mess
                 up the pretty log encoding, I think. ---Mark Brown.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer management; crash recovery; data processing;
                 database cache; database systems; fast restart; media
                 failure",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}; Software ---
                 Operating Systems --- Storage Management (D.4.2): {\bf
                 Main memory}",
}

@Article{Reuter:1984:PAR,
  author =       "Andreas Reuter",
  title =        "Performance Analysis of Recovery Techniques",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "526--559",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p526-reuter/p526-reuter.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p526-reuter/",
  abstract =     "Various logging and recovery techniques for
                 centralized transaction-oriented database systems under
                 performance aspects are described and discussed. The
                 classification of functional principles that has been
                 developed in a companion paper is used as a
                 terminological basis. In the main sections, a set of
                 analytic models is introduced and evaluated in order to
                 compare the performance characteristics of nine
                 different recovery techniques with respect to four key
                 parameters and a set of other parameters with less
                 influence. Finally, the results of model evaluation as
                 well as the limitations of the models themselves are
                 discussed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Kaiserslautern, Kaiserslautern, West Ger",
  affiliationaddress = "Univ of Kaiserslautern, Kaiserslautern, West
                 Ger",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; data processing;
                 database systems; logging and recovery; recovery and
                 restart; recovery techniques; Reliability; transaction
                 processing",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart};
                 Software --- Operating Systems --- Reliability (D.4.5):
                 {\bf Fault-tolerance}; Software --- Operating Systems
                 --- Performance (D.4.8): {\bf Modeling and prediction};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}; Information
                 Systems --- Database Management --- Database
                 Administration (H.2.7): {\bf Logging and recovery}",
}

@Article{Effelsberg:1984:PDB,
  author =       "Wolfgang Effelsberg and Theo Haerder",
  title =        "Principles of Database Buffer Management",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "560--595",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p560-effelsberg/p560-effelsberg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p560-effelsberg/",
  abstract =     "This paper discusses the implementation of a database
                 buffer manager as a component of a DBMS. The interface
                 between calling components of higher system layers and
                 the buffer manager is described; the principal
                 differences between virtual memory paging and database
                 buffer management are outlined; the notion of
                 referencing versus addressing of database pages is
                 introduced; and the concept of fixing pages in the
                 buffer to prevent uncontrolled replacement is
                 explained.\par

                 Three basic tasks have to be performed by the buffer
                 manager: buffer search, allocation of frames to
                 concurrent transactions, and page replacement. For each
                 of these tasks, implementation alternatives are
                 discussed and illustrated by examples from a
                 performance evaluation project of a CODASYL DBMS.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM, Scientific Cent, Heidelberg, West Ger",
  affiliationaddress = "IBM, Scientific Cent, Heidelberg, West Ger",
  annote =       "an interface between the buffer manager and the DBMS,
                 choices of page replacement policies; does not cover
                 sequential I/O (read-ahead and write-behind).",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer management; computer programming ---
                 algorithms; data processing; database systems; memory
                 paging; referencing database pages; replacement
                 algorithms",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2); Software --- Operating Systems
                 --- Storage Management (D.4.2): {\bf Storage
                 hierarchies}",
}

@Article{Bernstein:1984:ACC,
  author =       "Philip A. Bernstein and Nathan Goodman",
  title =        "An Algorithm for Concurrency Control and Recovery in
                 Replicated Distributed Databases",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "596--615",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "86k:68010",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 Distributed/fault.tolerant.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p596-bernstein/p596-bernstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p596-bernstein/",
  abstract =     "In a one-copy distributed database, each data item is
                 stored at exactly one site. In a replicated database,
                 some data items may be stored at multiple sites. The
                 main motivation is improved reliability: by storing
                 important data at multiple sites, the DBS can operate
                 even though some sites have failed.\par

                 This paper describes an algorithm for handling
                 replicated data, which allows users to operate on data
                 so long as one copy is ``available.'' A copy is
                 ``available'' when (i) its site is up, and (ii) the
                 copy is not out-of-date because of an earlier crash.
                 \par

                 The algorithm handles clean, detectable site failures,
                 but not Byzantine failures or network partitions.",
  acknowledgement = ack-nhfb,
  affiliation =  "Sequoia Systems Inc, Marlborough, MA, USA",
  affiliationaddress = "Sequoia Systems Inc, Marlborough, MA, USA",
  annote =       "3-phase commit. The first and third phases are
                 identical to the two phases of 2-phase commit. There is
                 a `Precommit' phase after the first phase where the
                 knowledge of the coordinator is replicated elsewhere,
                 thus protecting against a crash of the coordinator
                 (which could result in locks being tied up for long
                 periods).",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Algorithms; computer programming; concurrency control
                 and recovery; continuous operation; database systems;
                 replicated distributed databases; serializability;
                 transaction processing",
  subject =      "Information Systems --- Database Management ---
                 Physical Design (H.2.2): {\bf Recovery and restart};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Distributed databases}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Transaction processing}",
}

@Article{Chen:1984:ANV,
  author =       "Wen Chin Chen and Jeffrey Scott Vitter",
  title =        "Analysis of New Variants of Coalesced Hashing",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "616--645",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "794 550",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p616-chen/p616-chen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p616-chen/",
  abstract =     "The coalesced hashing method has been shown to be very
                 fast for dynamic information storage and retrieval.
                 This paper analyzes in a uniform way the performance of
                 coalesced hashing and its variants, thus settling some
                 open questions in the literature.\par

                 In all the variants, the range of the hash function is
                 called the {\em address region}, and extra space
                 reserved for storing colliders is called the {\em
                 cellar}. We refer to the unmodified method, which was
                 analyzed previously, as {\em late-insertion\/}
                 coalesced hashing. In this paper we analyze late
                 insertion and two new variations called {\em early
                 insertion\/} and {\em varied insertion}. When there is
                 no cellar, the early-insertion method is better than
                 late insertion; however, past experience has indicated
                 that it might be worse when there is a cellar. Our
                 analysis confirms that it is worse. The
                 varied-insertion method was introduced as a means of
                 combining the advantages of late insertion and early
                 insertion. This paper shows that varied insertion
                 requires fewer probes per search, on the average, than
                 do the other variants.\par

                 Each of these three coalesced hashing methods has a
                 parameter that relates the sizes of the address region
                 and the cellar. Techniques in this paper are designed
                 for tuning the parameter in order to achieve optimum
                 search times. We conclude with a list of open
                 problems.",
  acknowledgement = ack-nhfb,
  affiliation =  "Brown Univ, Dep of Computer Science, Providence, RI,
                 USA",
  affiliationaddress = "Brown Univ, Dep of Computer Science, Providence,
                 RI, USA",
  annote =       "Chaining and open addressing. Internal memory is
                 assumed!",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "address region; cellar; coalesced hashing; computer
                 programming --- algorithms; data processing; early
                 insertion; information retrieval systems; late
                 insertion",
  subject =      "Data --- Data Storage Representations (E.2): {\bf
                 Hash-table representations}; Software --- Software
                 Engineering --- Metrics (D.2.8): {\bf Performance
                 measures}; Theory of Computation --- Analysis of
                 Algorithms and Problem Complexity --- Nonnumerical
                 Algorithms and Problems (F.2.2): {\bf Sorting and
                 searching}; Mathematics of Computing --- Discrete
                 Mathematics --- Combinatorics (G.2.1): {\bf Generating
                 functions}; Mathematics of Computing --- Discrete
                 Mathematics --- Combinatorics (G.2.1): {\bf
                 Permutations and combinations}; Mathematics of
                 Computing --- Discrete Mathematics --- Combinatorics
                 (G.2.1): {\bf Recurrences and difference equations};
                 Mathematics of Computing --- Probability and Statistics
                 (G.3): {\bf Random number generation}; Information
                 Systems --- Information Storage and Retrieval ---
                 Information Search and Retrieval (H.3.3): {\bf Search
                 process}",
}

@Article{Deogun:1984:OCF,
  author =       "J. S. Deogun and V. V. Raghavan and T. K. W. Tsou",
  title =        "Organization of Clustered Files for Consecutive
                 Retrieval",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "646--671",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p646-deogun/p646-deogun.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p646-deogun/",
  abstract =     "This paper studies the problem of storing single-level
                 and multilevel clustered files. Necessary and
                 sufficient conditions for a single-level clustered file
                 to have the consecutive retrieval property (CRP) are
                 developed. A linear time algorithm to test the CRP for
                 a given clustered file and to identify the proper
                 arrangement of objects, if CRP exists, is presented.
                 For the single-level clustered files that do not have
                 CRP, it is shown that the problem of identifying a
                 storage organization with minimum redundancy is
                 NP-complete.\par

                 Consequently, an efficient heuristic algorithm to
                 generate a good storage organization for such files is
                 developed. Furthermore, it is shown that, for certain
                 types of multilevel clustered files, there exists a
                 storage organization such that the objects in each
                 cluster, for all clusters in each level of the
                 clustering, appear in consecutive locations.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Nebraska, Lincoln, NE, USA",
  affiliationaddress = "Univ of Nebraska, Lincoln, NE, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CLUSTERED FILES; computer programming --- Algorithms;
                 CONSECUTIVE RETRIEVAL; data processing --- File
                 Organization; FILE ORGANIZATION; information retrieval
                 systems; NP-COMPLETE",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3):
                 {\bf Clustering}; Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2):
                 {\bf File organization}",
}

@Article{Traub:1984:SSS,
  author =       "J. F. Traub and Y. Yemini and H. Wozniakowski",
  title =        "The Statistical Security of a Statistical Database",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "672--679",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Dec 8 08:54:10 MST 1996",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This study proposes a statistical perturbation scheme
                 to protect a statistical database against compromise.
                 The proposed scheme can handle the security of
                 numerical as well as nonnumerical sensitive fields or a
                 combination of fields. Furthermore, knowledge of some
                 records in a database does not help to compromise
                 unknown records. The authors use Chebyshev's inequality
                 to analyze the trade-offs among the magnitude of the
                 perturbations, the error incurred by statistical
                 queries, and the size of the query set to which they
                 apply. They show that if the statistician is given
                 absolute error guarantees, then a compromise is
                 possible, but the cost is made exponential in the size
                 of the database.",
  acknowledgement = ack-nhfb,
  affiliation =  "Columbia Univ, Dep of Computer Science, New York, NY,
                 USA",
  affiliationaddress = "Columbia Univ, Dep of Computer Science, New
                 York, NY, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Chebyshev's inequality; complexity of compromise; data
                 processing --- Security of data; database systems;
                 security; statistical database",
}

@Article{Navathe:1984:VPA,
  author =       "Shamkant Navathe and Stefano Ceri and Gio Wiederhold
                 and Jinglie Dou",
  title =        "Vertical Partitioning Algorithms for Database Design",
  journal =      j-TODS,
  volume =       "9",
  number =       "4",
  pages =        "680--710",
  month =        dec,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  note =         "Also published in/as: Stanford Un., TR-CS-82-957, Jan.
                 1983, revised Aug. 1983.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1984-9-4/p680-navathe/p680-navathe.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1984-9-4/p680-navathe/",
  abstract =     "This paper addresses the vertical partitioning of a
                 set of logical records or a relation into fragments.
                 The rationale behind vertical partitioning is to
                 produce fragments, groups of attribute columns, that
                 ``closely match'' the requirements of transactions.
                 \par

                 Vertical partitioning is applied in three contexts: a
                 database stored on devices of a single type, a database
                 stored in different memory levels, and a distributed
                 database. In a two-level memory hierarchy, most
                 transactions should be processed using the fragments in
                 primary memory. In distributed databases, fragment
                 allocation should maximize the amount of local
                 transaction processing.\par

                 Fragments may be nonoverlapping or overlapping. A
                 two-phase approach for the determination of fragments
                 is proposed; in the first phase, the design is driven
                 by empirical objective functions which do not require
                 specific cost information. The second phase performs
                 cost optimization by incorporating the knowledge of a
                 specific application environment. The algorithms
                 presented in this paper have been implemented, and
                 examples of their actual use are shown.",
  acknowledgement = ack-nhfb,
  affiliation =  "Stanford Univ, Dep of Computer Science, Stanford, CA,
                 USA",
  affiliationaddress = "Stanford Univ, Dep of Computer Science,
                 Stanford, CA, USA",
  annote =       "based on affinity considerations.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Algorithms; CLUSTERS; computer programming; data
                 processing; database systems --- Design; FRAGMENT
                 ALLOCATION; MEMORY LEVELS; VERTICAL PARTITIONING
                 ALGORITHMS",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1); Information Systems ---
                 Database Management --- Physical Design (H.2.2);
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Transaction processing}",
}

@Article{Maier:1984:DFG,
  author =       "D. Maier",
  title =        "Databases in the Fifth Generation Project: Is {Prolog}
                 a Database Language?",
  journal =      j-TODS,
  volume =       "9",
  number =       "2",
  pages =        "??--??",
  month =        jun,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:48:52 1996",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: ACM SIGMOD, 1984.",
  annote =       "very readable discussion, includes links to universal
                 relation research.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Skeen:1984:IAP,
  author =       "D. Skeen and D. D. Wright",
  title =        "Increasing Availability in Partitioned Database
                 Systems",
  journal =      j-TODS,
  volume =       "??",
  number =       "??",
  pages =        "290--299",
  month =        Apr,
  year =         "1984",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:48:54 1996",
  bibsource =    "Distributed/Dist.Sys.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Franaszek:1985:LCT,
  author =       "Peter Franaszek and John T. Robinson",
  title =        "Limitations of Concurrency in Transaction Processing",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "1--28",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/real.time.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p1-franaszek/p1-franaszek.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p1-franaszek/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3160.html",
  abstract =     "Given the pairwise probability of conflict p among
                 transactions in a transaction processing system,
                 together with the total number of concurrent
                 transactions n, the effective level of concurrency
                 E(n,p) is defined as the expected number of the n
                 transactions that can run concurrently and actually do
                 useful work. Using a random graph model of concurrency,
                 we show for three general classes of concurrency
                 control methods, examples of which are (1) standard
                 locking, (2) strict priority scheduling, and (3)
                 optimistic methods, that (1) E(n, p) n(1 - p/2) n-1,
                 (2) E(n, p) (1 - (1 - p) n)/p, and (3) 1 + ((1 -
                 p)/p)ln(p(n - 1) + 1) E(n, p) 1 + (1/p)ln(p(n - 1) +
                 1). Thus, for fixed p, as n ??, (1) E 0 for standard
                 locking methods, (2) E 1/p for strict priority
                 scheduling methods, and (3) E for optimistic methods.
                 Also found are bounds on E in the case where conflicts
                 are analyzed so as to maximize E.\par

                 The predictions of the random graph model are confirmed
                 by simulations of an abstract transaction processing
                 system. In practice, though, there is a price to pay
                 for the increased effective level of concurrency of
                 methods (2) and (3): using these methods there is more
                 wasted work (i.e., more steps executed by transactions
                 that are later aborted). In response to this problem,
                 three new concurrency control methods suggested by the
                 random graph model analysis are developed. Two of
                 these, called (a) running priority and (b) older or
                 running priority, are shown by the simulation results
                 to perform better than the previously known methods
                 (l)-(3) for relatively large n or large p, in terms of
                 achieving a high effective level of concurrency at a
                 comparatively small cost in wasted work.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM, Thomas J. Watson Research Cent, Yorktown Heights,
                 NY, USA",
  affiliationaddress = "IBM, Thomas J. Watson Research Cent, Yorktown
                 Heights, NY, USA",
  annote =       "6 methods, incl. optimistic (best) but not
                 version-ing.",
  classification = "722; 723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer systems, digital --- multiprocessing;
                 concurrency control; database systems; mathematical
                 techniques --- graph theory; performance; theory;
                 transaction processing, algorithms; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems. {\bf
                 D.1.3}: Software, PROGRAMMING TECHNIQUES, Concurrent
                 Programming. {\bf D.4.1}: Software, OPERATING SYSTEMS,
                 Process Management, Concurrency.",
}

@Article{Sacca:1985:DPC,
  author =       "Domenico Sacca and Gio Wiederhold",
  title =        "Database Partitioning in a Cluster of Processors",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "29--56",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Parallel/Multi.bib",
  note =         "Also published in
                 \cite[242--247]{Schkolnick:1983:ICV}, and IBM Research
                 Report No. RJ-4076, 1983.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p29-sacca/p29-sacca.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p29-sacca/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3161.html",
  abstract =     "In a distributed database system the partitioning and
                 allocation of the database over the processor nodes of
                 the network can be a critical aspect of the database
                 design effort. In this paper we develop and evaluate
                 algorithms that perform this task in a computationally
                 feasible manner. The network we consider is
                 characterized by a relatively high communication
                 bandwidth, considering the processing and input output
                 capacities in its processors. Such a balance is typical
                 if the processors are connected via busses or local
                 networks. The common constraint that transactions have
                 a specific root node no longer exists, so that there
                 are more distribution choices. However, a poor
                 distribution leads to less efficient computation,
                 higher costs, and higher loads in the nodes or in the
                 communication network so that the system may not be
                 able to handle the required set of transactions.
                 \par

                 Our approach is to first split the database into
                 fragments which constitute appropriate units for
                 allocation. The fragments to be allocated are selected
                 based on maximal benefit criteria using a greedy
                 heuristic. The assignment to processor nodes uses a
                 first-fit algorithm. The complete algorithm, called
                 GFF, is stated in a procedural form.\par

                 The complexity of the problem and of its candidate
                 solutions are analyzed and several interesting
                 relationships are proven. Alternate benefit metrics are
                 considered, since the execution cost of the allocation
                 procedure varies by orders of magnitude with the
                 alternatives of benefit evaluation. A mixed benefit
                 evaluation strategy is eventually proposed.\par

                 A model for evaluation is presented. Two of the
                 strategies are experimentally evaluated, and the
                 reported results support the discussion. The approach
                 should be suitable for other cases where resources have
                 to be allocated subject to resource constraints.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM, Research Lab, San Jose, CA, USA",
  affiliationaddress = "IBM, Research Lab, San Jose, CA, USA",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer systems, digital --- multiprocessing;
                 database partitioning, parallelism declustering
                 partitioning disk striping TODS, algorithms; database
                 systems; design; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems.",
}

@Article{Pramanik:1985:UGT,
  author =       "Sakti Pramanik and David Ittner",
  title =        "Use of Graph-Theoretic Models for Optimal Relational
                 Database Accesses to Perform Join",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "57--74",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "794 551",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p57-pramanik/p57-pramanik.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p57-pramanik/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3325.html",
  abstract =     "A graph model is presented to analyze the performance
                 of a relational join. The amount of page reaccesses,
                 the page access sequence, and the amount of buffer
                 needed are represented in terms of graph parameters. By
                 using the graph model formed from the index on the join
                 attributes, we determine the relationships between
                 these parameters. Two types of buffer allocation
                 strategies are studied, and the upper bound on the
                 buffer size with no page reaccess is given. This bound
                 is shown to be the maximum cut value of a graph. Hence,
                 the problem of computing this upper bound is NP-hard.
                 We also give algorithms to determine a page access
                 sequence requiring a near optimal buffer size with no
                 page reaccess. The optimal page access sequence for a
                 fixed buffer size has also been considered.",
  acknowledgement = ack-nhfb,
  affiliation =  "Michigan State Univ, Computer Science Dep, East
                 Lansing, MI, USA",
  affiliationaddress = "Michigan State Univ, Computer Science Dep, East
                 Lansing, MI, USA",
  annote =       "buffer management for indexes.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Measurement; Performance;
                 Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; database systems;
                 experimentation; graph-theoretic models, query
                 optimization processing TODS, algorithms; mathematical
                 techniques --- graph theory; measurement; performance;
                 relational database accesses; relational join; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.2.2}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Graph
                 algorithms.",
}

@Article{Larson:1985:LHO,
  author =       "Per-{\AA}ke Larson",
  title =        "Linear Hashing with Overflow-Handling by Linear
                 Probing",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "75--89",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p75-larson/p75-larson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p75-larson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3324.html",
  abstract =     "Linear hashing is a file structure for dynamic files.
                 In this paper, a new, simple method for handling
                 overflow records in connection with linear hashing is
                 proposed. The method is based on linear probing and
                 does not rely on chaining. No dedicated overflow area
                 is required. The expansion sequence of linear hashing
                 is modified to improve the performance, which requires
                 changes in the address computation. A new address
                 computation algorithm and an expansion algorithm are
                 given. The performance of the method is studied by
                 simulation. The algorithms for the basic file
                 operations are very simple, and the overall performance
                 is competitive with that of other variants of linear
                 hashing.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Waterloo, Dep of Computer Science, Waterloo,
                 Ont, Can",
  affiliationaddress = "Univ of Waterloo, Dep of Computer Science,
                 Waterloo, Ont, Can",
  annote =       "New algorithm for files that grow and shrink
                 dynamically; the overflow records of a full page are
                 directed to the next page of a group; the introduction
                 of five groups and the backwards split order makes this
                 algorithm better than previous ones.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; data processing;
                 database systems; design; dynamic hashing; file
                 organization; linear hashing; measurement; open
                 addressing, algorithms; performance",
  review =       "ACM CR 8512-1134",
  subject =      "{\bf E.2}: Data, DATA STORAGE REPRESENTATIONS,
                 Hash-table representations. {\bf D.2.2}: Software,
                 SOFTWARE ENGINEERING, Tools and Techniques, Decision
                 tables. {\bf E.5}: Data, FILES,
                 Organization/structure.",
}

@Article{Veklerov:1985:ADH,
  author =       "Eugene Veklerov",
  title =        "Analysis of Dynamic Hashing with Deferred Splitting",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "90--96",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p90-veklerov/p90-veklerov.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p90-veklerov/",
  abstract =     "Dynamic hashing with deferred splitting is a file
                 organization scheme which increases storage
                 utilization, as compared to `standard' dynamic hashing.
                 In this scheme, splitting of a bucket is deferred if
                 the bucket is full but its brother can accommodate new
                 records. The performance of the scheme is analyzed. In
                 a typical case the expected storage utilization
                 increases from 69 to 76 percent.",
  acknowledgement = ack-nhfb,
  affiliation =  "Lawrence Berkeley Lab, Real Time Systems Group,
                 Berkeley, CA, USA",
  affiliationaddress = "Lawrence Berkeley Lab, Real Time Systems Group,
                 Berkeley, CA, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; database systems; DEFERRED SPLITTING;
                 DYNAMIC HASHING; File Organization; STORAGE
                 UTILIZATION",
}

@Article{Palvia:1985:EBS,
  author =       "Prashant Palvia",
  title =        "Expressions for Batched Searching of Sequential and
                 Hierarchical Files",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "97--106",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p97-palvia/p97-palvia.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p97-palvia/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3326.html",
  abstract =     "Batching yields significant savings in access costs in
                 sequential, tree-structured, and random files. A direct
                 and simple expression is developed for computing the
                 average number of records\slash pages accessed to
                 satisfy a batched query of a sequential file. The
                 advantages of batching for sequential and random files
                 are discussed. A direct equation is provided for the
                 number of nodes accessed in unbatched queries of
                 hierarchical files. An exact recursive expression is
                 developed for node accesses in batched queries of
                 hierarchical files. In addition to the recursive
                 relationship, good, closed-form upper- and lower-bound
                 approximations are provided for the case of batched
                 queries of hierarchical files.",
  acknowledgement = ack-nhfb,
  affiliation =  "Temple Univ, Dep of Computer \& Information Sciences,
                 Philadelphia, PA, USA",
  affiliationaddress = "Temple Univ, Dep of Computer \& Information
                 Sciences, Philadelphia, PA, USA",
  classification = "723; 901",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "batched searching; database systems; design;
                 hierarchical files; information science --- information
                 retrieval; sequential files, performance; theory",
  subject =      "{\bf E.5}: Data, FILES, Organization/structure. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval, Search process.",
}

@Article{Bever:1985:DHS,
  author =       "Martin Bever and Peter C. Lockemann",
  title =        "Database Hosting in Strongly-Typed Programming
                 Languages",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "107--126",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-1/p107-bever/p107-bever.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-1/p107-bever/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3327.html",
  abstract =     "Database system support has become an essential part
                 of many computer applications, which have extended
                 beyond the more traditional commercial applications to,
                 among others, engineering applications.
                 Correspondingly, application programming with the need
                 to access databases has progressively shifted to
                 scientifically oriented languages.\par

                 Modern developments in these languages are
                 characterized by advanced mechanisms for the liberal
                 declaration of data types, for type checking, and
                 facilities for modularization of large programs. The
                 present paper examines how a DBMS can be accessed from
                 such a language in a way that conforms to its syntax
                 and utilizes its type-checking facilities, without
                 modifying the language specification itself, and hence
                 its compilers. The basic idea is to rely on facilities
                 for defining modules as separately compilable units,
                 and to use these to declare user-defined abstract data
                 types.\par

                 The idea is demonstrated by an experiment in which a
                 specific DBMS (ADABAS) is hosted in the programming
                 language (LIS). The paper outlines a number of
                 approaches and their problems, shows how to embed the
                 DML into LIS, and how a more user-oriented DML can be
                 provided in LIS.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8707-597",
  affiliation =  "Univ Karlsruhe, Inst fuer Informatik, Karlsruhe, West
                 Ger",
  affiliationaddress = "Univ Karlsruhe, Inst fuer Informatik, Karlsruhe,
                 West Ger",
  annote =       "ADABAS is the experimental target system and the
                 language is LIS.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database hosting,
                 design; database systems; languages; parameterized data
                 types; schema mapping; strongly-typed programming
                 languages",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Data manipulation languages (DML). {\bf
                 D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Abstract data types. {\bf
                 D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Data types and structures.
                 {\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Data description
                 languages (DDL).",
}

@Article{Chen:1985:AAS,
  author =       "Wen Chin Chen and Jeffrey Scott Vitter",
  title =        "Addendum to: {``Analysis of Some New Variants of
                 Coalesced Hashing''} [{ACM} Trans. Database Systems
                 {\bf 9} (1984), no. 4, 616--645]",
  journal =      j-TODS,
  volume =       "10",
  number =       "1",
  pages =        "127--127",
  month =        mar,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "794 552",
  bibsource =    "Database/Graefe.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Vitter:1985:EIO,
  author =       "Jeffrey Scott Vitter",
  title =        "An Efficient {I/O} Interface for Optical Disks",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "129--162",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p129-vitter/p129-vitter.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p129-vitter/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3862.html",
  abstract =     "We introduce the notion of an I/O interface for
                 optical digital (write-once) disks, which is quite
                 different from earlier research. The purpose of an I/O
                 interface is to allow existing operating systems and
                 application programs that use magnetic disks to use
                 optical disks instead, with minimal change. We define
                 what it means for an I/O interface to be
                 disk-efficient. We demonstrate a practical disk-
                 efficient I/O interface and show that its I/O
                 performance in many cases is optimum, up to a constant
                 factor, among all disk-efficient interfaces. The
                 interface is most effective for applications that are
                 not update-intensive. An additional capability is a
                 built-in history mechanism that provides software
                 support for accessing previous versions of records.
                 Even if not implemented, the I/O interface can be used
                 as a programming tool to develop efficient special
                 purpose applications for use with optical disks.",
  acknowledgement = ack-nhfb,
  affiliation =  "Brown Univ, Dep of Computer Science, Providence, RI,
                 USA",
  affiliationaddress = "Brown Univ, Dep of Computer Science, Providence,
                 RI, USA",
  annote =       "An I/O interface supports basic update operations such
                 as insert write and delete on the block is proposed.
                 Index techniques for erasable media (Btree is assumed
                 in this paper) can be implemented on this interface.
                 Versions of a block is stored as an allocation tree on
                 an optical disk, which is an efficient implementation
                 of the pointer fill-in method. Contents of a version of
                 a block is represented by an offset tree. Theoretical
                 lower bound of these operations is evaluated. This
                 paper assumes that appending into existing block is
                 possible on optical disk.",
  classification = "722; 741",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer interfaces; data storage, optical; design;
                 I/O interface; optical disks, algorithms; performance;
                 theory",
  subject =      "{\bf D.4.2}: Software, OPERATING SYSTEMS, Storage
                 Management, Secondary storage. {\bf D.4.2}: Software,
                 OPERATING SYSTEMS, Storage Management,
                 Allocation/deallocation strategies. {\bf D.4.3}:
                 Software, OPERATING SYSTEMS, File Systems Management,
                 Access methods. {\bf D.4.3}: Software, OPERATING
                 SYSTEMS, File Systems Management, File organization.
                 {\bf E.1}: Data, DATA STRUCTURES, Trees. {\bf E.2}:
                 Data, DATA STORAGE REPRESENTATIONS, Hash-table
                 representations. {\bf E.2}: Data, DATA STORAGE
                 REPRESENTATIONS, Linked representations. {\bf F.2.2}:
                 Theory of Computation, ANALYSIS OF ALGORITHMS AND
                 PROBLEM COMPLEXITY, Nonnumerical Algorithms and
                 Problems, Sorting and searching. {\bf G.2.1}:
                 Mathematics of Computing, DISCRETE MATHEMATICS,
                 Combinatorics, Combinatorial algorithms. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Access methods.",
}

@Article{Schkolnick:1985:ECU,
  author =       "M. Schkolnick and P. Tiberio",
  title =        "Estimating the Cost of Updates in a Relational
                 Database",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "163--179",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p163-schkolnick/p163-schkolnick.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p163-schkolnick/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3863.html",
  abstract =     "In this paper, cost formulas are derived for the
                 updates of data and indexes in a relational database.
                 The costs depend on the data scan type and the
                 predicates involved in the update statements. We show
                 that update costs have a considerable influence, both
                 in the context of the physical database design problem
                 and in access path selection in query optimization for
                 relational DBMSs.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Research Lab, San Jose, CA, USA",
  affiliationaddress = "IBM Research Lab, San Jose, CA, USA",
  annote =       "tradeoff by a given index query cost against update
                 cost.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "costs; database systems; measurement; performance;
                 query optimization; relational databases; update costs,
                 design",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Query processing.",
}

@Article{Yu:1985:ARC,
  author =       "C. T. Yu and Cheing-Mei Suen and K. Lam and M. K.
                 Siu",
  title =        "Adaptive Record Clustering",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "180--204",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p180-yu/p180-yu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p180-yu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3861.html",
  abstract =     "An algorithm for record clustering is presented. It is
                 capable of detecting sudden changes in users' access
                 patterns and then suggesting an appropriate assignment
                 of records to blocks. It is conceptually simple, highly
                 intuitive, does not need to classify queries into
                 types, and avoids collecting individual query
                 statistics. Experimental results indicate that it
                 converges rapidly; its performance is about 50 percent
                 better than that of the total sort method, and about
                 100 percent better than that of randomly assigning
                 records to blocks.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Illinois at Chicago Circle, Dep of Electrical
                 Engineering \& Computer Science, Chicago, IL, USA",
  affiliationaddress = "Univ of Illinois at Chicago Circle, Dep of
                 Electrical Engineering \& Computer Science, Chicago,
                 IL, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Measurement; Performance;
                 Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "adaptive record clustering; algorithms; computer
                 programming --- algorithms; database systems;
                 experimentation; file organization; measurement;
                 performance; physical database design; probabilistic
                 retrieval, CTYU TODS; theory; verification, data
                 processing",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design. {\bf E.5}: Data, FILES,
                 Organization/structure. {\bf H.2.m}: Information
                 Systems, DATABASE MANAGEMENT, Miscellaneous. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods.",
}

@Article{Katoh:1985:CTS,
  author =       "Naoki Katoh and Toshihide Ibaraki and Tiko Kameda",
  title =        "Cautious Transaction Schedulers with Admission
                 Control",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "205--229",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p205-katoh/p205-katoh.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p205-katoh/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3860.html",
  abstract =     "We propose a new class of schedulers, called {\em
                 cautious schedulers}, that grant an input request if it
                 will not necessitate any rollback in the future. In
                 particular, we investigate cautious WRW-schedulers that
                 output schedules in class WRW only. Class WRW consists
                 of all schedules that are serializable, while
                 preserving the write-read and read-write conflict, and
                 is the largest polynomially {\em recognizable\/}
                 subclass of serializable schedules currently known. It
                 is shown, in this paper however, that cautious WRW-
                 {\em scheduling\/} is, in general, NP-complete.
                 Therefore, we introduce a special type ({\em type
                 1R\/}) of transaction, which consists of no more than
                 one read step (an indivisible set of read operations)
                 followed by multiple write steps. It is shown that
                 cautious WRW-scheduling can be performed efficiently if
                 all transactions are of type 1R and if {\em admission
                 control\/} can be exercised. Admission control rejects
                 a transaction unless its first request is immediately
                 grantable.",
  acknowledgement = ack-nhfb,
  affiliation =  "Kobe Univ of Commerce, Dep of Management Science,
                 Kobe, Japan",
  affiliationaddress = "Kobe Univ of Commerce, Dep of Management
                 Science, Kobe, Japan",
  annote =       "serializability control for predefined transaction
                 sequences.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; database systems; scheduling;
                 serializability; transaction scheduler, algorithms",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing.",
}

@Article{Albano:1985:GST,
  author =       "Antonio Albano and Luca Cardelli and Renzo Orsini",
  title =        "{Galileo}: a Strongly-Typed, Interactive Conceptual
                 Language",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "230--260",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/Functional.bib; Object/Nierstrasz.bib",
  note =         "Also published in \cite{Zdonik:1990:ROO}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p230-albano/p230-albano.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p230-albano/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3859.html",
  abstract =     "Galileo, a programming language for database
                 applications, is presented. Galileo is a
                 strongly-typed, interactive programming language
                 designed specifically to support semantic data model
                 features (classification, aggregation, and
                 specialization), as well as the abstraction mechanisms
                 of modern programming languages (types, abstract types,
                 and modularization). The main contributions of Galileo
                 are (a) a flexible type system to model database
                 structure and semantic integrity constraints; (b) the
                 inclusion of type hierarchies to support the
                 specialization abstraction mechanisms of semantic data
                 models; (c) a modularization mechanism to structure
                 data and operations into interrelated units (d) the
                 integration of abstraction mechanisms into an
                 expression-based language that allows interactive use
                 of the database without resorting to a new stand-alone
                 query language.\par

                 Galileo will be used in the immediate future as a tool
                 for database design and, in the long term, as a
                 high-level interface for DBMSs.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ di Pisa, Dipartmento di Informatica, Pisa,
                 Italy",
  affiliationaddress = "Univ di Pisa, Dipartmento di Informatica, Pisa,
                 Italy",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data description
                 languages; data manipulation; database systems;
                 functional abstract data types; Galileo; languages;
                 olit-oopl Galileo; query languages, design",
  subject =      "{\bf D.3.2}: Software, PROGRAMMING LANGUAGES, Language
                 Classifications, GALILEO. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data description languages
                 (DDL). {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data manipulation languages
                 (DML). {\bf D.3.3}: Software, PROGRAMMING LANGUAGES,
                 Language Constructs and Features, Abstract data types.
                 {\bf D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Data types and structures.
                 {\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema.",
}

@Article{Huang:1985:HBT,
  author =       "Shou-Hsuan Stephen Huang",
  title =        "Height-balanced Trees of Order $ (\beta, \gamma,
                 \delta) $",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "261--284",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P05",
  MRnumber =     "801 578",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p261-huang/p261-huang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p261-huang/",
  abstract =     "We study restricted classes of B-trees, called $
                 H(\beta, \gamma, \delta) $ trees. A class is defined by
                 three parameters: $ \beta $, the size of a node; $
                 \gamma $, the minimal number of grandsons a node must
                 have; and $ \delta $, the minimal number of leaves
                 bottom nodes must have. This generalizes the brother
                 condition of 2-3 brother trees in a uniform way to
                 B-trees of higher order. The class of B-trees of order
                 m is obtained by choosing $ \beta = m $, $ \gamma = (m
                 / 2)^2 $, and $ \delta = m / 2 $. An algorithm to
                 construct H-trees for any given number of keys is given
                 in Section 1. Insertion and deletion algorithms are
                 given in Section 2. The costs of these algorithms
                 increase smoothly as the parameters are increased.
                 Furthermore, it is proved that the insertion can be
                 done in time $ O(?? + \log N) $, where $N$ is the
                 number of nodes in the tree. Deletion can also be
                 accomplished without reconstructing the entire tree.
                 Properties of H-trees are given in Section 3. It is
                 shown that the height of H-trees decreases as
                 increases, and the storage utilization increases
                 significantly as increases. Finally, comparisons with
                 other restricted classes of B-trees are given in
                 Section 4 to show the attractiveness of H-trees.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Houston, Dep of Computer Science, Houston, TX,
                 USA",
  affiliationaddress = "Univ of Houston, Dep of Computer Science,
                 Houston, TX, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; b-trees; compact b-trees; computer
                 programming --- algorithms; data processing; data
                 structures; dense multiway trees; height-balanced
                 trees; performance",
  subject =      "{\bf E.1}: Data, DATA STRUCTURES, Trees. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Access methods.",
}

@Article{Piwowarski:1985:CBS,
  author =       "Marek Piwowarski",
  title =        "Comments on Batched Searching of Sequential and
                 Tree-Structured Files",
  journal =      j-TODS,
  volume =       "10",
  number =       "2",
  pages =        "285--287",
  month =        jun,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Shneiderman:1976:BSS,Batory:1982:UMP}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-2/p285-piwowarski/p285-piwowarski.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-2/p285-piwowarski/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214294.html",
  abstract =     "Exact formulas for the expected cost savings from
                 batching requests against two types of j-ary trees are
                 given. Approximate expressions are also presented.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "performance",
  subject =      "{\bf H.3.2}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Storage, File organization.
                 {\bf E.1}: Data, DATA STRUCTURES, Trees. {\bf F.2.2}:
                 Theory of Computation, ANALYSIS OF ALGORITHMS AND
                 PROBLEM COMPLEXITY, Nonnumerical Algorithms and
                 Problems, Sorting and searching.",
}

@Article{Ullman:1985:ILQ,
  author =       "Jeffrey D. Ullman",
  title =        "Implementation of Logical Query Languages for
                 Databases",
  journal =      j-TODS,
  volume =       "10",
  number =       "3",
  pages =        "289--321",
  month =        sep,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Ai/nonmono.bib; Ai/prolog.1.bib; Compendex database;
                 Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Sep., YEAR $=$ 1985",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-3/p289-ullman/p289-ullman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-3/p289-ullman/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/3980.html",
  abstract =     "We examine methods of implementing queries about
                 relational databases in the case where these queries
                 are expressed in first-order logic as a collection of
                 Horn clauses. Because queries may be defined
                 recursively, straightforward methods of query
                 evaluation do not always work, and a variety of
                 strategies have been proposed to handle subsets of
                 recursive queries. We express such query evaluation
                 techniques as ``capture rules'' on a graph representing
                 clauses and predicates. One essential property of
                 capture rules is that they can be applied
                 independently, thus providing a clean interface for
                 query-evaluation systems that use several different
                 strategies in different situations. Another is that
                 there be an efficient test for the applicability of a
                 given rule. We define basic capture rules corresponding
                 to application of operators from relational algebra, a
                 top-down capture rule corresponding to ``backward
                 chaining,'' that is, repeated resolution of goals, a
                 bottom-up rule, corresponding to ``forward chaining,''
                 where we attempt to deduce all true facts in a given
                 class, and a ``sideways'' rule that allows us to pass
                 results from one goal to another.",
  acknowledgement = ack-nhfb,
  affiliation =  "Stanford Univ, Dep of Computer Science, Stanford, CA,
                 USA",
  affiliationaddress = "Stanford Univ, Dep of Computer Science,
                 Stanford, CA, USA",
  classification = "723",
  conference =   "Sel Pap from the 1985 ACM SIGMOD Conf",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database systems; Horn
                 clauses; languages; logical query languages; relational
                 databases, Prolog, algorithms; theory; verification",
  meetingaddress = "Austin, TX, USA",
  meetingdate =  "May 28--31 1985",
  meetingdate2 = "05/28--31/85",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving, Logic programming. {\bf I.2.4}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Knowledge Representation Formalisms and Methods,
                 Predicate logic.",
}

@Article{Anonymous:1985:SPA,
  author =       "Anonymous",
  title =        "Selected Papers from the 1985 {ACM SIGMOD
                 Conference}",
  journal =      j-TODS,
  volume =       "10",
  number =       "3",
  pages =        "289--346",
  month =        sep,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 10 07:59:49 1998",
  bibsource =    "Compendex database; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This issue contains 2 conference papers. The topics
                 covered are: logical query languages for databases; and
                 modeling concepts for VLSI CAD objects.",
  acknowledgement = ack-nhfb,
  classification = "714; 723",
  conference =   "Selected Papers from the 1985 ACM SIGMOD Conference.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  journalabr =   "ACM Transactions on Database Systems",
  keywords =     "CAD; database systems; design automation; integrated
                 circuits, VLSI --- computer aided design; logical query
                 languages; relational databases",
  meetingaddress = "Austin, TX, USA",
  sponsor =      "ACM, Special Interest Group on Management of Data, New
                 York, NY, USA",
}

@Article{Batory:1985:MCV,
  author =       "D. S. Batory and Won Kim",
  title =        "Modeling Concepts for {VLSI CAD} Objects",
  journal =      j-TODS,
  volume =       "10",
  number =       "3",
  pages =        "322--346",
  month =        sep,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: ACM-SIGMOD 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-3/p322-batory/p322-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-3/p322-batory/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/4018.html",
  abstract =     "VLSI CAD applications deal with design objects that
                 have an interface description and an implementation
                 description. Versions of design objects have a common
                 interface but differ in their implementations. A
                 molecular object is a modeling construct which enables
                 a database entity to be represented by two sets of
                 heterogeneous records, one set describes the object's
                 interface and the other describes its implementation.
                 Thus a reasonable starting point for modeling design
                 objects is to begin with the concept of molecular
                 objects.\par

                 In this paper, we identify modeling concepts that are
                 fundamental to capturing the semantics of VLSI CAD
                 design objects and versions in terms of molecular
                 objects. A provisional set of user operations on design
                 objects, consistent with these modeling concepts, is
                 also defined. The modeling framework that we present
                 has been found useful for investigating physical
                 storage techniques and change notification problems in
                 version control.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Texas at Austin, Dep of Computer Sciences,
                 Austin, TX, USA",
  affiliationaddress = "Univ of Texas at Austin, Dep of Computer
                 Sciences, Austin, TX, USA",
  classification = "714; 723",
  conference =   "Sel Pap from the 1985 ACM SIGMOD Conf",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CAD; Computer Aided Design; database systems; design
                 automation; integrated circuits, VLSI; languages;
                 relational databases; storage techniques, design;
                 version control",
  meetingaddress = "Austin, TX, USA",
  meetingdate =  "May 28--31 1985",
  meetingdate2 = "05/28--31/85",
  subject =      "{\bf B.7.1}: Hardware, INTEGRATED CIRCUITS, Types and
                 Design Styles, VLSI (very large scale integration).",
}

@Article{Subieta:1985:SQL,
  author =       "Kazimierz Subieta",
  title =        "Semantics of Query Languages for Network Databases",
  journal =      j-TODS,
  volume =       "10",
  number =       "3",
  pages =        "347--394",
  month =        sep,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-3/p347-subieta/p347-subieta.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-3/p347-subieta/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214293.html",
  abstract =     "Semantics determines the meaning of language
                 constructs; hence it says much more than syntax does
                 about implementing the language. The main purpose of
                 this paper is a formal presentation of the meaning of
                 basic language constructs employed in many database
                 languages (sublanguages). Therefore, stylized query
                 languages SSL (Sample Selection Language) and J (Joins)
                 are introduced, wherein most of the typical entries
                 present in other query languages are collected. The
                 semantics of SSL and J are defined by means of the
                 denotational method and explained informally. In SSL
                 and J, four types of expressions are introduced: a
                 selector (denotes a set of addresses), a term (denotes
                 a set of values), a formula (denotes a truth value),
                 and a join (denotes a set of n-tuples of addresses or
                 values). In many cases alternative semantics are given
                 and discussed. In order to obtain more general
                 properties of the proposed languages, a new database
                 access model is introduced, intended to be a tool for
                 the description of the logical access paths to data. In
                 particular, the access paths of the network and
                 relational models can be described. SSL and J
                 expressions may be addressed to both data structures.
                 In the case of the relational model, expressions of J
                 are similar to SQL or QUEL statements. Thus J may be
                 considered a generalization of relational query
                 languages for the network model. Finally, a programming
                 language, based on SSL and J, is outlined, and the
                 issues of SSL and J implementation are considered.",
  acknowledgement = ack-nhfb,
  affiliation =  "Polish Acad of Sciences, Inst of Computer Science,
                 Warsaw, Pol",
  affiliationaddress = "Polish Acad of Sciences, Inst of Computer
                 Science, Warsaw, Pol",
  classification = "723",
  conference =   "Sel Pap from the 1985 ACM SIGMOD Conf",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data manipulation
                 languages; database systems; denotational semantics;
                 query languages; query optimization, languages;
                 theory",
  meetingaddress = "Austin, TX, USA",
  meetingdate =  "May 28--31 1985",
  meetingdate2 = "05/28--31/85",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML).",
}

@Article{Liew:1985:DDP,
  author =       "Chong K. Liew and Uinam J. Choi and Chung J. Liew",
  title =        "A Data Distortion by Probability Distribution",
  journal =      j-TODS,
  volume =       "10",
  number =       "3",
  pages =        "395--411",
  month =        sep,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-3/p395-liew/p395-liew.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-3/p395-liew/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/4017.html",
  abstract =     "This paper introduces data distortion by probability
                 distribution, a probability distortion that involves
                 three steps. The first step is to identify the
                 underlying density function of the original series and
                 to estimate the parameters of this density function.
                 The second step is to generate a series of data from
                 the estimated density function. And the final step is
                 to map and replace the generated series for the
                 original one. Because it is replaced by the distorted
                 data set, probability distortion guards the privacy of
                 an individual belonging to the original data set. At
                 the same time, the probability distorted series
                 provides asymptotically the same statistical properties
                 as those of the original series, since both are under
                 the same distribution. Unlike conventional point
                 distortion, probability distortion is difficult to
                 compromise by repeated queries, and provides a maximum
                 exposure for statistical analysis.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Oklahoma, Norman, OK, USA",
  affiliationaddress = "Univ of Oklahoma, Norman, OK, USA",
  annote =       "analysis of pollution technique.",
  classification = "723",
  conference =   "Sel Pap from the 1985 ACM SIGMOD Conf",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data distortion; database systems; probability;
                 probability distortion; security; statistical
                 databases, statistical security; TODS, algorithms",
  meetingaddress = "Austin, TX, USA",
  meetingdate =  "May 28--31 1985",
  meetingdate2 = "05/28--31/85",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration. {\bf G.3}: Mathematics of
                 Computing, PROBABILITY AND STATISTICS, Statistical
                 computing. {\bf G.3}: Mathematics of Computing,
                 PROBABILITY AND STATISTICS.",
}

@Article{Tay:1985:LPC,
  author =       "Y. C. Tay and Nathan Goodman and Rajan Suri",
  title =        "Locking Performance in Centralized Databases",
  journal =      j-TODS,
  volume =       "10",
  number =       "4",
  pages =        "415--462",
  month =        dec,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-4/p415-tay/p415-tay.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-4/p415-tay/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/4880.html",
  abstract =     "An analytic model is used to study the performance of
                 dynamic locking. The analysis uses only the
                 steady-state average values of the variables. The
                 solution to the model is given by a cubic, which has
                 exactly one valid root for the range of parametric
                 values that is of interest. The model's predictions
                 agree well with simulation results for transactions
                 that require up to twenty locks. The model separates
                 data contention from resource contention, thus
                 facilitating an analysis of their separate effects and
                 their interaction. It shows that systems with a
                 particular form of nonuniform access, or with shared
                 locks, are equivalent to systems with uniform access
                 and only exclusive locks.\par

                 Blocking due to conflicts is found to impose an upper
                 bound on transaction throughput; this fact leads to a
                 rule of thumb on how much data contention should be
                 permitted in a system. Throughput can exceed this bound
                 if a transaction is restarted whenever it encounters a
                 conflict, provided restart costs and resource
                 contention are low. It can also be exceeded by making
                 transactions predeclare their locks. Raising the
                 multiprogramming level to increase throughput also
                 raises the number of restarts per completion.
                 Transactions should minimize their lock requests,
                 because data contention is proportional to the square
                 of the number of requests. The choice of how much data
                 to lock at a time depends on which part of a general
                 granularity curve the system sees.",
  acknowledgement = ack-nhfb,
  affiliation =  "Natl Univ of Singapore, Singapore, Singapore",
  affiliationaddress = "Natl Univ of Singapore, Singapore, Singapore",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Measurement; Performance; Theory;
                 Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; data contention; database
                 locking; database systems; measurement; performance;
                 resource contention, algorithms; theory; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Deadlock avoidance. {\bf C.4}: Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS, Modeling
                 techniques. {\bf C.2.2}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Network Protocols.
                 {\bf C.2.1}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Network Architecture
                 and Design, Centralized networks.",
}

@Article{Batory:1985:MSA,
  author =       "D. S. Batory",
  title =        "Modeling the Storage Architectures of Commercial
                 Database Systems",
  journal =      j-TODS,
  volume =       "10",
  number =       "4",
  pages =        "463--528",
  month =        dec,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-4/p463-batory/p463-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-4/p463-batory/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5392.html",
  abstract =     "Modeling the storage structures of a DBMS is a
                 prerequisite to understanding and optimizing database
                 performance. Previously, such modeling was very
                 difficult because the fundamental role of
                 conceptual-to-internal mappings in DBMS implementations
                 went unrecognized.\par

                 In this paper we present a model of physical databases,
                 called the transformation model, that makes
                 conceptual-to-internal mappings explicit. By exposing
                 such mappings, we show that it is possible to model the
                 storage architectures (i.e., the storage structures and
                 mappings) of many commercial DBMSs in a precise,
                 systematic, and comprehensible way. Models of the
                 INQUIRE, ADABAS, and SYSTEM 2000 storage architectures
                 are presented as examples of the model's utility.
                 \par

                 We believe the transformation model helps bridge the
                 gap between physical database theory and practice. It
                 also reveals the possibility of a technology to
                 automate the development of physical database
                 software.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Texas at Austin, Austin, TX, USA",
  affiliationaddress = "Univ of Texas at Austin, Austin, TX, USA",
  annote =       "considers ADABAS, INQUIRE, SYSTEM2000 in depth.
                 Classification of linksets. modeling storage methods of
                 Inquire, ADABAS, and System 2000.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Documentation",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing --- data structures; database systems;
                 documentation; storage architectures, design",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf E.5}: Data,
                 FILES, Organization/structure. {\bf H.2.m}: Information
                 Systems, DATABASE MANAGEMENT, Miscellaneous.",
}

@Article{Agrawal:1985:ICC,
  author =       "Rakesh Agrawal and David J. Dewitt",
  title =        "Integrated Concurrency Control and Recovery
                 Mechanisms: Design and Performance Evaluation",
  journal =      j-TODS,
  volume =       "10",
  number =       "4",
  pages =        "529--564",
  month =        dec,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-4/p529-agrawal/p529-agrawal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-4/p529-agrawal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/4958.html",
  abstract =     "In spite of the wide variety of concurrency control
                 and recovery mechanisms proposed during the past
                 decade, the behavior and the performance of various
                 concurrency control and recovery mechanisms remain
                 largely not well understood. In addition, although
                 concurrency control and recovery mechanisms are
                 intimately related, the interaction between them has
                 not been adequately explored. In this paper, we take a
                 unified view of the problems associated with
                 concurrency control and recovery for
                 transaction-oriented multiuser centralized database
                 management systems, and we present several integrated
                 mechanisms. We then develop analytical models to study
                 the behavior and compare the performance of these
                 integrated mechanisms, and we present the results of
                 our performance evaluation.",
  acknowledgement = ack-nhfb,
  affiliation =  "AT\&T Bell Lab, Murray Hill, NJ, USA",
  affiliationaddress = "AT\&T Bell Lab, Murray Hill, NJ, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; database systems; design;
                 measurement; performance; recovery mechanisms;
                 transaction processing, algorithms",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Recovery and restart. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Deadlock avoidance. {\bf D.4.1}:
                 Software, OPERATING SYSTEMS, Process Management,
                 Concurrency. {\bf D.4.3}: Software, OPERATING SYSTEMS,
                 File Systems Management.",
}

@Article{Borgida:1985:LFF,
  author =       "Alexander Borgida",
  title =        "Language Features for Flexible Handling of Exceptions
                 in Information Systems",
  journal =      j-TODS,
  volume =       "10",
  number =       "4",
  pages =        "565--603",
  month =        dec,
  year =         "1985",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Rutgers Un., TR-LCSR-70, rev.
                 Mar. 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1985-10-4/p565-borgida/p565-borgida.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1985-10-4/p565-borgida/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/4995.html",
  abstract =     "An exception-handling facility suitable for languages
                 used to implement database-intensive information
                 systems is presented. Such a mechanism facilitates the
                 development and maintenance of more flexible software
                 systems by supporting the abstraction of details
                 concerning special or abnormal occurrences. The type
                 constraints imposed by the schema as well as various
                 semantic integrity assertions are considered to be
                 normalcy conditions, and the key contribution of this
                 work is to allow exceptions to these constraints to
                 persist. To achieve this, solutions are proposed to a
                 range of problems, including sharing and computing with
                 exceptional information, exception handling by users,
                 the logic of constraints with exceptions, and
                 implementation issues. The use of exception handling in
                 dealing with null values, estimates, and measurement is
                 also illustrated.",
  acknowledgement = ack-nhfb,
  affiliation =  "Rutgers Univ, Dep of Computer Science, New Brunswick,
                 NJ, USA",
  affiliationaddress = "Rutgers Univ, Dep of Computer Science, New
                 Brunswick, NJ, USA",
  annote =       "Adding exception handling to database systems to deal
                 with unusual, unknown, or otherwise exceptional
                 attribute values. A semantic extension that may inspire
                 KBMSers. I have a report in my office on the
                 possibilities of this approach, by Alex Borgida of
                 Rutgers. It's very readable, and it may inspire someone
                 to cook up such a scheme of his or her own for Naxos,
                 thesis, or whatever. -----Marianne W. W.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database systems;
                 exception handling; languages; semantic integrity;
                 theory; type constraints, design; verification",
  subject =      "{\bf D.2.5}: Software, SOFTWARE ENGINEERING, Testing
                 and Debugging, Error handling and recovery. {\bf
                 H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 description languages (DDL). {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML).",
}

@Article{Hagmann:1986:PAS,
  author =       "Robert Brian Hagmann and Domenico Ferrari",
  title =        "Performance Analysis of Several Back-End Database
                 Architectures",
  journal =      j-TODS,
  volume =       "11",
  number =       "1",
  pages =        "1--26",
  month =        mar,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-1/p1-hagmann/p1-hagmann.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-1/p1-hagmann/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5242.html",
  abstract =     "The growing acceptance of database systems makes their
                 performance increasingly more important. One way to
                 gain performance is to off-load some of the functions
                 of the database system to aback-end computer. The
                 problem is what functions should be off-loaded to
                 maximize the benefits of distributed processing.
                 \par

                 Our approach to this problem consisted of constructing
                 several variants of an existing relational database
                 system. INGRES, that partition the database system
                 software into two parts, and assigning these two parts
                 to two computers connected by a local area network. For
                 the purposes of this experiment, six different variants
                 of the database software were constructed to test the
                 sir most interesting functional subdivisions. Each
                 variant was then benchmarked using two different
                 databases and query streams. The communication medium
                 and the communication software were also benchmarked to
                 measure their contribution to the performance of each
                 configuration.\par

                 Combining the database and network measurement results,
                 various conclusions were reached about the viability of
                 the configurations, the desirable properties of the
                 communications mechanisms to he used, the operating
                 system interface and overhead, and the performance of
                 the database system. The variants to be preferred
                 depend on the hardware technology, operating system
                 features, database system internal structure, and
                 network software overhead.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, Berkely, CA, USA",
  affiliationaddress = "Univ of California, Berkely, CA, USA",
  annote =       "an experimental methodology using INGRES.",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Experimentation; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "back-end database architectures; computer networks ---
                 local networks; computer systems, digital ---
                 distributed; database systems; experimentation; Ingres
                 database system; measurement; performance; relational
                 databases, hardware support database machine TODS,
                 design",
  subject =      "{\bf H.2.6}: Information Systems, DATABASE MANAGEMENT,
                 Database Machines. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems. {\bf C.4}: Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS. {\bf H.2.0}:
                 Information Systems, DATABASE MANAGEMENT, General,
                 INGRES.",
}

@Article{Garcia-Molina:1986:ABA,
  author =       "H{\'e}ctor Garc{\'\i}a-Molina and Frank Pittelli and
                 Susan Davidson",
  title =        "Applications of {Byzantine} Agreement in Database
                 Systems",
  journal =      j-TODS,
  volume =       "11",
  number =       "1",
  pages =        "27--47",
  month =        mar,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-1/p27-molina/p27-molina.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-1/p27-molina/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5243.html",
  abstract =     "In this paper we study when and how a Byzantine
                 agreement protocol can be used in general-purpose
                 database management, systems. We present an overview of
                 the failure model used for Bizantine agreement, and of
                 the protocol itself. We then present correctness
                 criteria for database processing in this failure
                 environment and discuss strategies for satisfying them.
                 In doing this, we present new failure models for
                 input\slash output nodes and study ways to distribute
                 input transactions to processing nodes under these
                 models. Finally, we investigate applications of
                 Byzantine agreement protocols in the more common
                 failure environment where processors are assumed to
                 halt after a failure.",
  acknowledgement = ack-nhfb,
  affiliation =  "Princeton Univ, Princeton, NJ, USA",
  affiliationaddress = "Princeton Univ, Princeton, NJ, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; Byzantine agreement protocol; data
                 processing; database systems; distributed; failure
                 models; fault tolerance; reliability",
  subject =      "{\bf D.4.5}: Software, OPERATING SYSTEMS, Reliability,
                 Fault-tolerance. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Distributed systems. {\bf
                 C.2.2}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Network Protocols.
                 {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases.",
}

@Article{Segev:1986:OJO,
  author =       "Arie Segev",
  title =        "Optimization of Join Operations in Horizontally
                 Partitioned Database Systems",
  journal =      j-TODS,
  volume =       "11",
  number =       "1",
  pages =        "48--80",
  month =        mar,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-1/p48-segev/p48-segev.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-1/p48-segev/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5241.html",
  abstract =     "This paper analyzes the problem of joining two
                 horizontally partitioned relations in a distributed
                 database system. Two types of semijoin strategies are
                 introduced, local and remote. Local semijoins are
                 performed at the site of the restricted relation (or
                 fragment), and remote semijoins can be performed at an
                 arbitrary site. A mathematical model of a semijoin
                 strategy for the case of remote semijoins is developed,
                 and lower bounding and heuristic procedures are
                 proposed. The results of computational experiments are
                 reported. The experiments include an analysis of the
                 heuristics' performance relative to the lower bounds,
                 sensitivity analysis, and error analysis. These results
                 reveal a good performance of the heuristic procedures,
                 and demonstrate the benefit of using semijoin
                 operations to reduce the size of fragments prior to
                 their transmission. The algorithms for the case of
                 remote semijoins were found to be superior to the
                 algorithms for the case of local semijoins. In
                 addition, we found that the estimation accuracy of the
                 selectivity factors has a significant effect on the
                 incurred communication cost.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, Berkely, CA, USA",
  affiliationaddress = "Univ of California, Berkely, CA, USA",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; database systems;
                 distributed; horizontally partitioned database systems,
                 query processing optimization tods; join operations;
                 mathematical models; optimization",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.2.2}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Trees.
                 {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Distributed
                 systems. {\bf G.2.1}: Mathematics of Computing,
                 DISCRETE MATHEMATICS, Combinatorics, Combinatorial
                 algorithms.",
}

@Article{Gyssens:1986:CJD,
  author =       "Marc Gyssens",
  title =        "On the Complexity of Join Dependencies",
  journal =      j-TODS,
  volume =       "11",
  number =       "1",
  pages =        "81--108",
  month =        mar,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "87g:68011",
  MRreviewer =   "J. Paredaens",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-1/p81-gyssens/p81-gyssens.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-1/p81-gyssens/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5237.html",
  abstract =     "In [10] a method is proposed for decomposing join
                 dependencies (jds) in a relational database using the
                 notion of a hinge. This method was subsequently studied
                 in [11] and [12]. We show how the technique of
                 decomposition can be used to make integrity checking
                 more efficient. It turns out that it is important to
                 find a decomposition that minimizes the number of edges
                 of its largest element. We show that the decompositions
                 obtained with the method described in [10] are optimal
                 in this respect. This minimality criterion leads to the
                 definition of the {\em degree of cyclicity}, which
                 allows us to classify jds and leads to the notion of
                 {\em n-cyclicity}, of which acyclicity is a special
                 case for n = 2. We then show that, for a fixed value of
                 n (which may be greater than 2). integrity checking can
                 be performed in polynomial time provided we restrict
                 ourselves to {\em n-cyclic\/} jds. Finally, we
                 generalize a well-known characterization for acyclic
                 jds by proving that n-cyclicity is equivalent to
                 ``n-wise consistency implies global consistency.'' As a
                 consequence, consistency checking can be performed in
                 polynomial time if we restrict ourselves to n-cyclic
                 jds, for a tired value of n, not necessarily equal to
                 2.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Antwerp, Antwerp, Belg",
  affiliationaddress = "Univ of Antwerp, Antwerp, Belg",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CYCLICITY; database systems; DECOMPOSITION; JOIN
                 DEPENDENCIES; MATHEMATICAL TECHNIQUES --- Graph Theory;
                 Relational",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms. {\bf G.2.2}: Mathematics
                 of Computing, DISCRETE MATHEMATICS, Graph Theory, Graph
                 algorithms. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Schema and subschema. {\bf
                 G.2.2}: Mathematics of Computing, DISCRETE MATHEMATICS,
                 Graph Theory, Trees.",
}

@Article{Sacco:1986:FTE,
  author =       "Giovanni Maria Sacco",
  title =        "Fragmentation: a technique for Efficient Query
                 Processing",
  journal =      j-TODS,
  volume =       "11",
  number =       "2",
  pages =        "113--133",
  month =        jun,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: University of Torino, TR., Aug.
                 1983.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-2/p113-sacco/p113-sacco.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-2/p113-sacco/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5638.html",
  abstract =     "A `divide and conquer' strategy to compute natural
                 joins by sequential scans on unordered relations is
                 described. This strategy is shown to always be better
                 than merging scans when both relations must be sorted
                 before joining, and generally better in practical cases
                 when only the largest relation must be sorted.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ di Torino",
  affiliationaddress = "Turin, Italy",
  annote =       "Join by hashing: Create fragments by hashing, as many
                 fragments as buffers can be allocated in memory. Then
                 repeat that for the other relation. Then do a nested
                 unsorted join, as Kim, W. 1980, on the fragment
                 pairs.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Economics; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- Algorithms; database systems;
                 divide-and-conquer algorithms; economics;
                 fragmentation; natural joins, join hash partitioning
                 overflow avoidance recursion parallelism TODS,
                 algorithms; performance; query processing",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing.",
}

@Article{Beeri:1986:IAL,
  author =       "Catriel Beeri and Michael Kifer",
  title =        "An Integrated Approach to Logical Design of Relational
                 Database Schemes",
  journal =      j-TODS,
  volume =       "11",
  number =       "2",
  pages =        "134--158",
  month =        jun,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "848 633",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-2/p134-beeri/p134-beeri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-2/p134-beeri/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214291.html",
  abstract =     "We propose a new approach to the design of relational
                 database schemes. The main features of the approach are
                 the following:\par

                 A combination of the traditional decomposition and
                 synthesis approaches, thus allowing the use of both
                 functional and multivalued dependencies.
                 \par

                 Separation of structural dependencies relevant for the
                 design process from integrity constraints, that is,
                 constraints that do not bear any structural information
                 about the data and which should therefore be discarded
                 at the design stage. This separation is supported by a
                 simple syntactic test filtering out nonstructural
                 dependencies.\par

                 Automatic correction of schemes which lack certain
                 desirable properties.",
  acknowledgement = ack-nhfb,
  affiliation =  "Hebrew Univ of Jerusalem, Jerusalem, Isr",
  affiliationaddress = "Hebrew Univ of Jerusalem, Jerusalem, Isr",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "acyclic schemes, design; database systems;
                 decomposition; functional dependencies; multivalued
                 dependencies; relational; synthesis; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Normal forms.",
}

@Article{Mendelson:1986:IIC,
  author =       "Haim Mendelson and Aditya N. Saharia",
  title =        "Incomplete Information Costs and Database Design",
  journal =      j-TODS,
  volume =       "11",
  number =       "2",
  pages =        "159--185",
  month =        jun,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-2/p159-mendelson/p159-mendelson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-2/p159-mendelson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5678.html",
  abstract =     "This paper presents a methodology for trading-off the
                 cost of incomplete information against the data-related
                 costs in the design of database systems. It
                 investigates how the usage patterns of the database,
                 defined by the characteristics of information requests
                 presented to it, affect its conceptual design. The
                 construction of minimum-cost answers to information
                 requests for a variety of query types and cost
                 structures is also studied. The resulting costs of
                 incomplete database information are balanced against
                 the data-related costs in the derivation of the optimal
                 design.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Rochester, Rochester, NY, USA",
  affiliationaddress = "Univ of Rochester, Rochester, NY, USA",
  annote =       "information value, missing data, decision theory
                 framework, applied to ships in the Mediterranean.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Economics; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data-related costs, design; database systems; design;
                 economics; incomplete information costs; optimization;
                 theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.1.1}: Information Systems,
                 MODELS AND PRINCIPLES, Systems and Information Theory.
                 {\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval.",
}

@Article{Ginsburg:1986:CTS,
  author =       "Seymour Ginsburg and Katsumi Tanaka",
  title =        "Computation-Tuple Sequences and Object Histories",
  journal =      j-TODS,
  volume =       "11",
  number =       "2",
  pages =        "186--212",
  month =        jun,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "848 634",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-2/p186-ginsburg/p186-ginsburg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-2/p186-ginsburg/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5924.html",
  abstract =     "A record-based, algebraically-oriented model is
                 introduced for describing data for ``object histories''
                 (with computation), such as checking accounts, credit
                 card accounts, taxes, schedules, and so on. The model
                 consists of sequences of computation tuples defined by
                 a computation-tuple sequence scheme (CSS). The CSS has
                 three major features (in addition to input data):
                 computation (involving previous computation tuples),
                 ``uniform'' constraints (whose satisfaction by a
                 computation-tuple sequence $u$ implies satisfaction by
                 every interval of $u$ ), and specific sequences with
                 which to start the valid computation-tuple sequences. A
                 special type of CSS, called ``local,'' is singled out
                 for its relative simplicity in maintaining the validity
                 of a computation-tuple sequence. A necessary and
                 sufficient condition for a CSS to be equivalent to at
                 least one local CSS is given. Finally, the notion of
                 ``local bisimulatability'' is introduced for regarding
                 two CSS as conveying the same information, and two
                 results on local bisimulatability in connection with
                 local CSS are established.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Southern California, Los Angeles, CA, USA",
  affiliationaddress = "Univ of Southern California, Los Angeles, CA,
                 USA",
  annote =       "Sequential history tuples and objects with input,
                 computation, and result. Some constraints applied per
                 sequential entry cause satisfaction of global
                 constraints. Temporal issues are very specific.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computation history; data description; database state
                 transitions; database systems; theory; transaction
                 processing, algorithms; verification",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.1.0}: Information
                 Systems, MODELS AND PRINCIPLES, General. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing.",
}

@Article{Garg:1986:OPK,
  author =       "Anil K. Garg and C. C. Gotlieb",
  title =        "Order-Preserving Key Transformations",
  journal =      j-TODS,
  volume =       "11",
  number =       "2",
  pages =        "213--234",
  month =        jun,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-2/p213-garg/p213-garg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-2/p213-garg/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/5923.html",
  abstract =     "File organizations based on conventional hash
                 functions provide faster access to the stored records
                 in comparison with tree-like file structures. Tree
                 structures such as B** plus -trees and ISAM do provide
                 for sequential processing, but require considerable
                 storage for the indices. When sequential processing is
                 needed a table that performs an order-preserving
                 transformation on keys can be used. H is an
                 order-preserving key transform if H(K//1) greater than
                 equivalent to H(K//2), for all keys K//1 greater than
                 K//2. We present methodologies for constructing such
                 key transforms, and illustrate them for some real-life
                 key sets. Storage requirements for the table needed to
                 carry out the transformation are less than those needed
                 for the indices.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Toronto, Toronto, Ont, Can",
  affiliationaddress = "Univ of Toronto, Toronto, Ont, Can",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Management; Measurement;
                 Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access methods, algorithms; data processing; database
                 systems; design; dynamic files; file organization; key
                 transformations; management; measurement;
                 order-preserving hashing; performance; theory",
  subject =      "{\bf E.5}: Data, FILES, Organization/structure. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods.",
}

@Article{Shapiro:1986:JPD,
  author =       "Leonard D. Shapiro",
  title =        "Join Processing in Database Systems with Large Main
                 Memories",
  journal =      j-TODS,
  volume =       "11",
  number =       "3",
  pages =        "239--264",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/database.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-3/p239-shapiro/p239-shapiro.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-3/p239-shapiro/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/6315.html",
  abstract =     "We study algorithms for computing the equijoin of two
                 relations in a system with a standard architecture hut
                 with large amounts of main memory. Our algorithms are
                 especially efficient when the main memory available is
                 a significant fraction of the size of one of the
                 relations to he joined; but they can be applied
                 whenever there is memory equal to approximately the
                 square root of the size of one relation. We present a
                 new algorithm which is a hybrid of two hash-based
                 algorithms and which dominates the other algorithms we
                 present, including sort-merge. Even in a virtual memory
                 environment, the hybrid algorithm dominates all the
                 others we study.\par

                 Finally, we describe how three popular tools to
                 increase the efficiency of joins, namely filters, Babb
                 arrays, and semijoins, can he grafted onto any of our
                 algorithms.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; memory query evaluation classical simple
                 hybrid hash joins TODS, algorithms; performance",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.0}: Information
                 Systems, DATABASE MANAGEMENT, General. {\bf H.2.6}:
                 Information Systems, DATABASE MANAGEMENT, Database
                 Machines. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design.",
}

@Article{Gavish:1986:SQO,
  author =       "Bezalel Gavish and Arie Segev",
  title =        "Set Query Optimization in Distributed Database
                 Systems",
  journal =      j-TODS,
  volume =       "11",
  number =       "3",
  pages =        "265--293",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-3/p265-gavish/p265-gavish.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-3/p265-gavish/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/6488.html",
  abstract =     "This paper addresses the problem of optimizing queries
                 that involve set operations (set queries) in a
                 distributed relational database system. A particular
                 emphasis is put on the optimization of such queries in
                 horizontally partitioned database systems. A
                 mathematical programming model of the set query problem
                 is developed and its NP-completeness is proved.
                 Solution procedures are proposed and computational
                 results presented. One of the main results of the
                 computational experiments is that, for many queries,
                 the solution procedures are not sensitive to errors in
                 estimating the size of results of set operations.",
  acknowledgement = ack-nhfb,
  annote =       "The time complexity is NP-complete. Three
                 approximations.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; languages; theory",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf G.2.2}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Trees.
                 {\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing.",
}

@Article{Lafortune:1986:STM,
  author =       "St{\'e}phane Lafortune and Eugene Wong",
  title =        "A State Transition Model for Distributed Query
                 Processing",
  journal =      j-TODS,
  volume =       "11",
  number =       "3",
  pages =        "294--322",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/database.bib; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/des.bib; Misc/Discrete.event.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-3/p294-lafortune/p294-lafortune.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-3/p294-lafortune/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/6460.html",
  abstract =     "A state transition model for the optimization of query
                 processing in a distributed database system is
                 presented. The problem is parameterized by means of a
                 state describing the amount of processing that has been
                 performed at each site where the database is located. A
                 state transition occurs each time a new join or
                 semijoin is executed. Dynamic programming is used to
                 compute recursively the costs of the states and the
                 globally optimal solution, taking into account
                 communication and local processing costs. The state
                 transition model is general enough to account for the
                 possibility of parallel processing among the various
                 sites, as well as for redundancy in the database. The
                 model also permits significant reductions of the
                 necessary computations by taking advantage of simple
                 additivity and site-uniformity properties of a cost
                 model, and of clever strategies that improve on the
                 basic dynamic programming algorithm.",
  acknowledgement = ack-nhfb,
  bib =          "koz",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory, Optimization
                 TODS",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Distributed
                 systems. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing.",
}

@Article{Lozinskii:1986:POI,
  author =       "Eliezer L. Lozinskii",
  title =        "A Problem-Oriented Inferential Database System",
  journal =      j-TODS,
  volume =       "11",
  number =       "3",
  pages =        "323--356",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68T20)",
  MRnumber =     "87k:68025",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-3/p323-lozinskii/p323-lozinskii.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-3/p323-lozinskii/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/6419.html",
  abstract =     "Recently developed inferential database systems face
                 some common problems: a very fast growth of search
                 space and difficulties in recognizing inference
                 termination (especially for recursive axioms). These
                 shortcomings stem mainly from the fact that the
                 inference process is usually separated from database
                 operations. A problem-oriented inferential system i8
                 described which refers to the database prior to query
                 (or subquery) processing, so that the inference from
                 the very beginning is directed by data relevant to the
                 query. A multiprocessor implementation of the system is
                 presented based on a computer network conforming to
                 database relations and axioms. The system provides an
                 efficient indication of query termination, and is
                 complete in the sense that it produces all correct
                 answers to a query in a finite time.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory;
                 Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory;
                 verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving, Deduction. {\bf C.1.3}: Computer
                 Systems Organization, PROCESSOR ARCHITECTURES, Other
                 Architecture Styles, Data-flow architectures.",
}

@Article{Osborn:1986:DRD,
  author =       "Sylvia L. Osborn and T. E. Heaven",
  title =        "The Design of a Relational Database System with
                 Abstract Data Types for Domains",
  journal =      j-TODS,
  volume =       "11",
  number =       "3",
  pages =        "357--373",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-3/p357-osborn/p357-osborn.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-3/p357-osborn/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/6461.html",
  abstract =     "An extension to the relational model is described in
                 which domains can he arbitrarily defined as abstract
                 data types. Operations on these data types include
                 primitive operations, aggregates, and transformations.
                 It is shown that these operations make the query
                 language complete in the sense of Chandra and Harel.
                 The system has been designed in such a way that new
                 data types and their operations can be defined with a
                 minimal amount of interaction with the database
                 management system.",
  acknowledgement = ack-nhfb,
  annote =       "Operations on simple objects, operations on aggregates
                 and `transformations' can be defined on relations. It
                 is possible to implement a transitive closure RAD uses
                 the data dictionary. ---Ong, Fogg and Stonebraker.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf D.3.3}: Software,
                 PROGRAMMING LANGUAGES, Language Constructs and
                 Features, Abstract data types. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages.",
}

@Article{Gawlick:1986:RIW,
  author =       "Dieter Gawlick",
  title =        "Report on the International Workshop on
                 High-Performance Transaction Systems",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "375--377",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p375-gawlick/p375-gawlick.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p375-gawlick/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/17346.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; performance",
  subject =      "{\bf A.0}: General Literature, GENERAL. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing. {\bf C.4}: Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS, Reliability,
                 availability, and serviceability.",
}

@Article{Mohan:1986:TMR,
  author =       "C. Mohan and B. Lindsay and R. Obermarck",
  title =        "Transaction Management in the {R*} Distributed
                 Database Management System",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "378--396",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p378-mohan/p378-mohan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p378-mohan/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/7266.html",
  abstract =     "This paper deals with the transaction management
                 aspects of the R* distributed database system. It
                 concentrates primarily on the description of the R*
                 commit protocols, Presumed Abort (PA) and Presumed
                 Commit (PC). PA and PC are extensions of the
                 well-known, two-phase (2P) commit protocol. PA is
                 optimized for read-only transactions and a class of
                 multisite update transactions, and PC is optimized for
                 other classes of multisite update transactions. The
                 optimizations result in reduced intersite message
                 traffic and log writes, and, consequently, a better
                 response time. The paper also discusses R*'s approach
                 toward distributed deadlock detection and resolution.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM, San Jose, CA, USA",
  affiliationaddress = "IBM, San Jose, CA, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "commit protocols; concurrency control, RSTAR TODS,
                 algorithms; database systems; deadlock victim
                 selection; design; distributed; optimization;
                 reliability; transaction management",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Concurrency. {\bf D.4.1}: Software,
                 OPERATING SYSTEMS, Process Management, Deadlocks. {\bf
                 D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Synchronization. {\bf D.4.5}: Software,
                 OPERATING SYSTEMS, Reliability, Fault-tolerance. {\bf
                 D.4.7}: Software, OPERATING SYSTEMS, Organization and
                 Design, Distributed systems. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Recovery
                 and restart. {\bf H.2.7}: Information Systems, DATABASE
                 MANAGEMENT, Database Administration, Logging and
                 recovery.",
}

@Article{Bayer:1986:CTR,
  author =       "Rudolf Bayer",
  title =        "Consistency of Transactions and Random Batch",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "397--404",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p397-bayer/p397-bayer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p397-bayer/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214287.html",
  abstract =     "A synchronization technique and scheduling strategy is
                 described, which allows us to run a batch process
                 simultaneously with on-line transactions. The batch
                 process and the transactions are serialized in such a
                 way that consistency level 3 is achieved.",
  acknowledgement = ack-nhfb,
  affiliation =  "Technische Univ Muenchen, West Ger",
  affiliationaddress = "Technische Univ Muenchen, West Ger",
  classification = "723; 913",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control, algorithms; consistency of
                 transactions; database systems; design; performance;
                 random batch; scheduling; synchronization",
  subject =      "{\bf D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management. {\bf D.4.7}: Software, OPERATING SYSTEMS,
                 Organization and Design. {\bf E.5}: Data, FILES. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems.",
}

@Article{ONeil:1986:ETM,
  author =       "Patrick E. O'Neil",
  title =        "The {Escrow} Transactional Method",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "405--430",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p405-o_neil/p405-o_neil.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p405-o_neil/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/7265.html",
  abstract =     "A method is presented for permitting record updates by
                 long-lived transactions without forbidding simultaneous
                 access by other users to records modified. Earlier
                 methods presented separately by Gawlick and Reuter are
                 comparable but concentrate on ``hot-spot'' situations,
                 where even short transactions cannot lock frequently
                 accessed fields without causing bottlenecks. The Escrow
                 Method offered here is designed to support nonblocking
                 record updates by transactions that are ``long lived''
                 and thus require long periods to complete.
                 Recoverability of intermediate results prior to commit
                 thus becomes a design goal, so that updates as of a
                 given time can be guaranteed against memory or media
                 failure while still retaining the prerogative to abort.
                 This guarantee basically completes phase one of a
                 two-phase commit, and several advantages result: (1) As
                 with Gawlick's and Reuter's methods, high-concurrency
                 items in the database will not act as a bottleneck; (2)
                 transaction commit of different updates can be
                 performed asynchronously, allowing natural distributed
                 transactions; indeed, distributed transactions in the
                 presence of delayed messages or occasional line
                 disconnection become feasible in a way that we argue
                 will tie up minimal resources for the purpose intended;
                 and (3) it becomes natural to allow for human
                 interaction in the middle of a transaction without loss
                 of concurrent access or any special difficulty for the
                 application programmer. The Escrow Method, like
                 Gawlick's Fast Path and Reuter's Method, requires the
                 database system to be an ``expert'' about the type of
                 transactional updates performed, most commonly updates
                 involving incremental changes to aggregate quantities.
                 However, the Escrow Method is extendable to other types
                 of updates.",
  acknowledgement = ack-nhfb,
  affiliation =  "Computer Corp of America, Cambridge, MA, USA",
  affiliationaddress = "Computer Corp of America, Cambridge, MA, USA",
  annote =       "For aggregate values (counts, sum) concurrency control
                 can use soft tolerances and keep them in escrow",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; design; escrow transactional method;
                 long-lived transactions; multiuser environment, locking
                 quantities, not variables TODS, algorithms; nested
                 transactions; performance; theory; two-phase commit",
  subject =      "{\bf D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Concurrency. {\bf D.4.1}: Software,
                 OPERATING SYSTEMS, Process Management, Deadlocks. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Recovery and restart. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Escrow. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing.",
}

@Article{Hsu:1986:PTP,
  author =       "Meichun Hsu and Arvola Chan",
  title =        "Partitioned Two-Phase Locking",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "431--446",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p431-hsu/p431-hsu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p431-hsu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/7477.html",
  abstract =     "In a large integrated database, there often exists an
                 ``information hierarchy,'' where both raw data and
                 derived data are stored and used together. Therefore,
                 among update transactions, there will often be some
                 that perform only read accesses from a certain (i.e.,
                 the ``raw'' data) portion of the database and write
                 into another (i.e., the ``derived'' data) portion. A
                 conventional concurrency control algorithm would have
                 treated such transactions as regular update
                 transactions and subjected them to the usual protocols
                 for synchronizing update transactions. In this paper
                 such transactions are examined more closely. The
                 purpose is to devise concurrency control methods that
                 allow the computation of derived information to proceed
                 without interfering with the updating of raw data.
                 \par

                 The first part of the paper presents a proof method for
                 correctness of concurrency control algorithms in a
                 hierarchically decomposed database. The proof method
                 provides a framework for understanding the intricacies
                 in dealing with hierarchically decomposed databases.
                 The second part of the paper is an application of the
                 proof method to show the correctness of a
                 two-phase-locking- based algorithm, called partitioned
                 two-phase locking, for hierarchically decomposed
                 databases. This algorithm is a natural extension to the
                 Version Pool method proposed previously in the
                 literature.",
  acknowledgement = ack-nhfb,
  affiliation =  "Harvard Univ, Cambridge, MA, USA",
  affiliationaddress = "Harvard Univ, Cambridge, MA, USA",
  annote =       "revisions also for update",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; concurrency
                 control; database systems; theory; transaction
                 processing, algorithms; two-phase locking",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design.",
}

@Article{Luk:1986:EEL,
  author =       "W. S. Luk and Steve Kloster",
  title =        "{ELFS}: {English} Language from {SQL}",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "447--472",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Dec 8 08:54:10 MST 1996",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this paper we describe a system which, given a
                 query in SQL-like relational database language, will
                 display its meaning in clear, unambiguous natural
                 language. The syntax-driven translation mechanism is
                 independent of the application domain. It has direct
                 applications in designing computer-based SQL tutorial
                 systems and program debugging systems. The research
                 results obtained in the paper will also be useful in
                 query optimization and design of a more user-friendly
                 language front-end for casual users.",
  acknowledgement = ack-nhfb,
  affiliation =  "Simon Fraser Univ, Burnaby, BC, Can",
  affiliationaddress = "Simon Fraser Univ, Burnaby, BC, Can",
  annote =       "display meaning in natural language is independent of
                 the application domain.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- Program Debugging; database
                 systems; ELFS; Query Languages; relational database
                 language; SQL",
}

@Article{Sacco:1986:BMR,
  author =       "Giovanni Maria Sacco and Mario Schkolnick",
  title =        "Buffer Management in Relational Database Systems",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "473--498",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p473-sacco/p473-sacco.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p473-sacco/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/7336.html",
  abstract =     "The hot-set model, characterizing the buffer
                 requirements of relational queries, is presented. This
                 model allows the system to determine the optimal buffer
                 space to be allocated to a query; it can also be used
                 by the query optimizer to derive efficient execution
                 plans accounting for the available buffer space, and by
                 a query scheduler to prevent thrashing. The hot-set
                 model is compared with the working-set model. A
                 simulation study is presented.",
  acknowledgement = ack-nhfb,
  acmcr =        "8708-0695",
  affiliation =  "Univ di Torino",
  affiliationaddress = "Turin, Italy",
  annote =       "The hot-set model provides a more meaningful measure
                 of cost than simple I/O counts.",
  classification = "723; 913",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "buffer management; database systems; performance;
                 query optimizer, algorithms; query processing;
                 relational; scheduling; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design. {\bf
                 D.4.2}: Software, OPERATING SYSTEMS, Storage
                 Management.",
}

@Article{Ariav:1986:TOD,
  author =       "Gad Ariav",
  title =        "A Temporally Oriented Data Model",
  journal =      j-TODS,
  volume =       "11",
  number =       "4",
  pages =        "499--527",
  month =        dec,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1986-11-4/p499-ariav/p499-ariav.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1986-11-4/p499-ariav/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/7350.html",
  abstract =     "The research into time and data models has so far
                 focused on the identification of extensions to the
                 classical relational model that would provide it with
                 ``adequate'' semantic capacity to deal with time. The
                 temporally oriented data model (TODM) presented in this
                 paper is a result of a different approach, namely, it
                 directly operationalizes the pervasive
                 three-dimensional metaphor for time. One of the main
                 results is thus the development of the notion of the
                 data cube: a three-dimensional and inherently temporal
                 data construct where time, objects, and attributes are
                 the primary dimensions of stored data. TODM's cube adds
                 historical depth to the tabular notions of data and
                 provides a framework for storing and retrieving data
                 within their temporal context. The basic operations in
                 the model allow the formation of new cubic views from
                 existing ones, or viewing data as one moves up and down
                 in time within cubes.\par

                 This paper introduces TODM, a consistent set of
                 temporally oriented data constructs, operations, and
                 constraints, and then presents TOSQL, a corresponding
                 end-user's SQL-like query syntax. The model is a
                 restricted but consistent superset of the relational
                 model, and the query syntax incorporates temporal
                 notions in a manner that likewise avoids penalizing
                 users who are interested solely in the current view of
                 data (rather than in a temporal perspective). The
                 naturalness of the spatial reference to time and the
                 added semantic capacity of TODM come with a price--the
                 definitions of the cubic constructs and basic
                 operations are relatively cumbersome. As rudimentary as
                 it is, TODM nonetheless provides a comprehensive basis
                 for formulating an external data model for a temporally
                 oriented database.",
  acknowledgement = ack-nhfb,
  affiliation =  "New York Univ, New York, NY, USA",
  affiliationaddress = "New York Univ, New York, NY, USA",
  annote =       "at least one timestamp, time of record, plus other
                 temporal --- event stamps.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data cube; database systems; information modeling;
                 languages; relational; temporally oriented data model;
                 theory; TODM, design; TOSQL",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML). {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages.",
}

@Article{Albano:1986:OSG,
  author =       "Antonio Albano and S. Alfo and Luca Cardelli and Renzo
                 Orsini",
  title =        "An Overview of {SIDEREUS}: a Graphical Database Schema
                 Editor for {Galileo}",
  journal =      j-TODS,
  volume =       "11",
  number =       "??",
  pages =        "568--571",
  month =        "????",
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:51:20 1996",
  bibsource =    "Distributed/gesturing.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Durand:1986:FMS,
  author =       "Charles Durand",
  title =        "Forward Multidimensional Search with Applications to
                 Information Retrieval",
  journal =      j-TODS,
  volume =       "??",
  number =       "??",
  pages =        "??--??",
  month =        sep,
  year =         "1986",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:51:25 1996",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Submitted.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Spyratos:1987:PMD,
  author =       "Nicolas Spyratos",
  title =        "The partition model: a deductive database model",
  journal =      j-TODS,
  volume =       "12",
  number =       "1",
  pages =        "1--37",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Institut National de la
                 Recherche en Informatique et Automatique, TR-286, Apr.
                 1983.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-1/p1-spyratos/p1-spyratos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-1/p1-spyratos/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22718.html",
  abstract =     "We present a new database model in which each
                 attribute is modeled by a family of disjoint subsets of
                 an underlying population of objects. Such a family is
                 called a partitioning, and the set of all partitionings
                 is turned into a lattice by appropriately defining
                 product and sum. A database is seen as a function from
                 a sublattice into the lattice of partitionings. The
                 model combines the following features:\par

                 (1) syntactic simplicity (essentially that of the
                 relational model),\par

                 (2) powerful means for the specification of semantic
                 information (in the form of lattice equations), and
                 \par

                 (3) deductive capability (essentially that of set
                 theory).\par

                 The relational model of data and the basic constructs
                 of semantic modeling can be embedded into our model in
                 a simple and straightforward manner.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ de Paris-Sud, Orsay, Fr",
  affiliationaddress = "Univ de Paris-Sud, Orsay, Fr",
  annote =       "Type hierarchies and lattices.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database semantics; database systems; deductive
                 database model; partition model; theory",
  subject =      "{\bf F.3.2}: Theory of Computation, LOGICS AND
                 MEANINGS OF PROGRAMS, Semantics of Programming
                 Languages, Algebraic approaches to semantics. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models.",
}

@Article{Wu:1987:ASM,
  author =       "C. T. Wu and Walter A. Burkhard",
  title =        "Associative Searching in Multiple Storage Units",
  journal =      j-TODS,
  volume =       "12",
  number =       "1",
  pages =        "38--64",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Discusses interpolation hashing, a multidimensional
                 variant of linear hashing.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-1/p38-wu/p38-wu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-1/p38-wu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/12048.html",
  abstract =     "A file maintenance model, called the multiple random
                 access storage units model, is introduced. Storage
                 units can be accessed simultaneously, and the parallel
                 processing of an associative query is achieved by
                 distributing data evenly among the storage units.
                 Maximum parallelism is obtained when data satisfying an
                 associative query are evenly distributed for every
                 possible query. An allocation scheme called $M$-cycle
                 allocation is proposed to maintain large files of data
                 on multiple random access storage units. The allocation
                 scheme provides an efficient and straightforward
                 indexing over multidimensional key spaces and supports
                 the parallel processing of orthogonal range queries.
                 Our analysis shows that $M$-cycle allocation achieves
                 the near-optimum parallelism for processing the
                 orthogonal range queries. Moreover, there is no
                 duplication of records and no increase in
                 insertion\slash deletion cost.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, San Diego, CA, USA",
  affiliationaddress = "Univ of California, San Diego, CA, USA",
  classification = "723; 903",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "associative searching; data processing --- file
                 organization; database systems; design; file
                 maintenance model; information science --- information
                 retrieval; multiple storage units; performance; random
                 access, algorithms; theory",
  subject =      "{\bf E.5}: Data, FILES. {\bf E.1}: Data, DATA
                 STRUCTURES. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design.",
}

@Article{Lomet:1987:PEF,
  author =       "David B. Lomet",
  title =        "Partial Expansions for File Organizations with an
                 Index",
  journal =      j-TODS,
  volume =       "12",
  number =       "1",
  pages =        "65--84",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-1/p65-lomet/p65-lomet.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-1/p65-lomet/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/12049.html",
  abstract =     "A new way to increase file space in dynamically
                 growing files is introduced in which substantial
                 improvement in file utilization can be achieved. It
                 makes use of partial expansions in which, instead of
                 doubling the space associated with some part of the
                 file, the space grows at a slower rate. Unlike previous
                 versions of partial expansion in which the number of
                 buckets involved in file growth is increased by less
                 than a factor of two, the new method expands file space
                 by increasing bucket size via `elastic buckets'. This
                 permits partial expansions to be used with a wide range
                 of indexed files, including B-trees. The results of
                 using partial expansions are analyzed, and the analysis
                 confirmed by a simulation study. The analysis and
                 simulation demonstrate that the file utilization gains
                 are substantial and that fears of excessive insertion
                 cost resulting from more frequent file growth are
                 unfounded.",
  acknowledgement = ack-nhfb,
  affiliation =  "Wang Inst of Graduate Studies, Tyngboro, MA, USA",
  affiliationaddress = "Wang Inst of Graduate Studies, Tyngboro, MA,
                 USA",
  annote =       "a way to increase file space with substantial
                 improvement in file utilization elastic buckets come in
                 a number of sizes.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; database systems; File Organization;
                 FILE UTILIZATION; INSERTION COST; PARTIAL EXPANSIONS",
  subject =      "{\bf D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, Access methods. {\bf D.4.3}: Software,
                 OPERATING SYSTEMS, File Systems Management, File
                 organization. {\bf H.2.2}: Information Systems,
                 DATABASE MANAGEMENT, Physical Design, Access methods.
                 {\bf H.3.2}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Storage, File
                 organization.",
}

@Article{Fedorowicz:1987:DPE,
  author =       "Jane Fedorowicz",
  title =        "Database Performance Evaluation in an Indexed File
                 Environment",
  journal =      j-TODS,
  volume =       "12",
  number =       "1",
  pages =        "85--110",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-1/p85-fedorowicz/p85-fedorowicz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-1/p85-fedorowicz/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/13675.html",
  abstract =     "The use of database systems for managerial decision
                 making often incorporates information-retrieval
                 capabilities with numeric report generation. Of great
                 concern to the user of such a system is the response
                 time associated with issuing a query to the database.
                 This study presents a procedure for estimating response
                 time for one of the most frequently encountered
                 physical storage mechanisms, the indexed file. The
                 model provides a fairly high degree of accuracy, but is
                 simple enough so that the cost of applying the model is
                 not exorbitant. The model incorporates the knowledge
                 that the distribution of access key occurrences is
                 known to follow Zipf's law. It first estimates the
                 access time required to complete the query, which
                 includes the time needed for all input and output
                 transactions, and CPU time used in performing the
                 search. The effects of multiple users on an
                 individual's response time are then assessed using a
                 simple regression estimation technique. The two-step
                 procedure allows for the separation of access time from
                 multiuser influences.",
  acknowledgement = ack-nhfb,
  affiliation =  "Boston Univ, Boston, MA, USA",
  affiliationaddress = "Boston Univ, Boston, MA, USA",
  annote =       "a procedure for estimating response time; distribution
                 of access key occurrences follow Zipf's law. Early
                 version with Kellogg, J. L. Model provides a fairly
                 high degree of accuracy but is simple. The effects of
                 multiple users are assessed using simple regression
                 estimation.",
  classification = "723; 912; 922",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing --- File organization; database
                 performance; database systems; indexed file
                 environment; MANAGEMENT --- Information Systems;
                 multiple users, design; Performance; performance;
                 response time; statistical methods --- regression
                 analysis; Zipf's law",
  subject =      "{\bf D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, File organization. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.3.2}: Information Systems, INFORMATION
                 STORAGE AND RETRIEVAL, Information Storage, File
                 organization. {\bf H.3.3}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Search
                 and Retrieval, Retrieval models.",
}

@Article{Ozsoyoglu:1987:NNF,
  author =       "Z. Meral {\"O}zsoyo{\u{g}}lu and Li-Yan Yuan",
  title =        "A New Normal Form for Nested Relations",
  journal =      j-TODS,
  volume =       "12",
  number =       "1",
  pages =        "111--136",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "886 100",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-1/p111-ozsoyoglu/p111-ozsoyoglu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-1/p111-ozsoyoglu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/13676.html",
  abstract =     "We consider nested relations whose schemes are
                 structured as trees, called scheme trees, and introduce
                 a normal form for such relations, called the nested
                 normal form. Given a set of attributes $U$, and a set
                 of multivalued dependencies (MVDs) $M$ over these
                 attributes, we present an algorithm to obtain a nested
                 normal form decomposition of $U$ with respect to $M$.
                 Such a decomposition has several desirable properties,
                 such as explicitly representing a set of full and
                 embedded MVDs implied by $M$, and being a faithful and
                 nonredundant representation of $U$. Moreover, if the
                 given set of MVDs is conflict-free, then the nested
                 normal form decomposition is also
                 dependency-preserving. Finally, we show that if $M$ is
                 conflict-free, then the set of root-to-leaf paths of
                 scheme trees in nested normal form decomposition is
                 precisely the unique 4NF decomposition $ [9, 16] $ of
                 $U$ with respect to $M$.",
  acknowledgement = ack-nhfb,
  affiliation =  "Case Western Reserve Univ, Cleveland, OH, USA",
  affiliationaddress = "Case Western Reserve Univ, Cleveland, OH, USA",
  annote =       "non-first normal form.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; data structures; database systems ---
                 design; decomposition, algorithms; design; multivalued
                 dependency; nested relations; normal form; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Normal
                 forms.",
}

@Article{Christodoulakis:1987:ARP,
  author =       "Stavros Christodoulakis",
  title =        "Analysis of Retrieval Performance for Records and
                 Objects using Optical Disk Technology",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "137--169",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p137-christodoulakis/p137-christodoulakis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p137-christodoulakis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/23015.html",
  abstract =     "In this paper we examine the problem of object and
                 record retrieval from optical disks. General objects
                 (such as images, documents, etc.) may be long and their
                 length may have high variance. We assume that all the
                 components of an object are stored consecutively in
                 storage to speed-up retrieval performance. We first
                 present an optical disk model and an optimal schedule
                 for retrieval of records and objects which qualify in a
                 single query on a file stored on an optical disk
                 device. We then provide {\em exact\/} and {\em
                 approximate\/} analytic results for evaluating the
                 retrieval performance for objects from an optical disk.
                 The analysis provides some basic analytic tools for
                 studying the performance of various file and database
                 organizations for optical disks. The results involve
                 probability distribution of block accesses, probability
                 distributions of span accesses, and probability
                 distribution of seek times. Record retrieval is an
                 important special case. This analysis differs from
                 similar ones in database environments in the following
                 respects: (1) the large size and large variance of the
                 size of objects; (2) crossing of track boundaries by
                 objects; (3) the capability for span access that
                 optical disks provide (e.g., when the optical assembly
                 is located in a given position, information can be read
                 from a number of consecutive tracks (span) with a small
                 additional cost).",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Waterloo, Waterloo, Ont, Can",
  affiliationaddress = "Univ of Waterloo, Waterloo, Ont, Can",
  classification = "723; 741; 903",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Measurement; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data storage, optical --- storage devices; information
                 retrieval; information science; optical disk
                 technology, measurement; performance; retrieval
                 performance; theory",
  subject =      "{\bf C.4}: Computer Systems Organization, PERFORMANCE
                 OF SYSTEMS, Modeling techniques. {\bf D.4.3}: Software,
                 OPERATING SYSTEMS, File Systems Management, File
                 organization.",
}

@Article{Herlihy:1987:DQA,
  author =       "Maurice Herlihy",
  title =        "Dynamic Quorum Adjustment for Partitioned Data",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "170--194",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p170-herlihy/p170-herlihy.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p170-herlihy/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22953.html",
  abstract =     "A partition occurs when functioning sites in a
                 distributed system are unable to communicate. This
                 paper introduces a new method for managing replicated
                 data objects in the presence of partitions. Each
                 operation provided by a replicated object has a set of
                 quorums, which are sets of sites whose cooperation
                 suffices to execute the operation. The method permits
                 an object's quorums to be adjusted dynamically in
                 response to failures and recoveries. A transaction that
                 is unable to progress using one set of quorums may
                 switch to another, more favorable set, and transactions
                 in different partitions may progress using different
                 sets. This method has three novel aspects: (1) it
                 supports a wider range of quorums than earlier
                 proposals, (2) it scales up effectively to large
                 systems because quorum adjustments do not require
                 global reconfiguration, and (3) it systematically
                 exploits the semantics of typed objects to support more
                 flexible quorum adjustment.",
  acknowledgement = ack-nhfb,
  affiliation =  "Carnegie-Mellon Univ, Pittsburgh, PA, USA",
  affiliationaddress = "Carnegie-Mellon Univ, Pittsburgh, PA, USA",
  annote =       "Each operation provided by a replicated object has a
                 set of quorums, sites whose cooperation suffices to
                 execute the operation.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer systems, digital --- distributed; database
                 systems; distributed; dynamic quorum adjustment;
                 languages; partitioned data, algorithms; reliability",
  subject =      "{\bf D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Abstract data types. {\bf
                 D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, Distributed file systems. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf D.3.3}: Software, PROGRAMMING
                 LANGUAGES, Language Constructs and Features, Data types
                 and structures. {\bf D.4.5}: Software, OPERATING
                 SYSTEMS, Reliability, Fault-tolerance. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing.",
}

@Article{Ellis:1987:CLH,
  author =       "Carla Schlatter Ellis",
  title =        "Concurrency in Linear Hashing",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "195--217",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite{ACM:1985:PFA}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p195-ellis/p195-ellis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p195-ellis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22954.html",
  abstract =     "Concurrent access to complex shared data structures,
                 particularly structures useful as database indices, has
                 long been of interest in the database community. In
                 dynamic databases, tree structures such as B-trees have
                 been used as indices because of their ability to handle
                 growth; whereas hashing has been used for fast access
                 in relatively static databases. Recently, a number of
                 techniques for dynamic hashing have appeared. They
                 address the major deficiency of traditional hashing
                 when applied to databases that experience significant
                 change in the amount of data being stored. This paper
                 presents a solution that allows concurrency in one of
                 these dynamic hashing data structures, namely linear
                 hash files. The solution is based on locking protocols
                 and minor modifications in the data structures.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Rochester, USA",
  affiliationaddress = "Univ of Rochester, USA",
  annote =       "Searching can proceed in parallel with splits. Also
                 discusses distributed access.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "complex shared data structures, algorithms; concurrent
                 access; data processing; Data Structures; database
                 systems; linear hashing",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf D.4.1}: Software, OPERATING
                 SYSTEMS, Process Management,
                 Multiprocessing/multiprogramming. {\bf E.1}: Data, DATA
                 STRUCTURES. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design. {\bf H.3.2}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Storage, File organization.",
}

@Article{Valduriez:1987:JI,
  author =       "Patrick Valduriez",
  title =        "Join Indices",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "218--246",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Compares join indices with inverted indices, clustered
                 indices, B+ trees, linked lists, and hybrid hash
                 techniques.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p218-valduriez/p218-valduriez.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p218-valduriez/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22955.html",
  abstract =     "In new application areas of relational database
                 systems, such as artificial intelligence, the join
                 operator is used more extensively than in conventional
                 applications. In this paper, we propose a simple data
                 structure, called a join index, for improving the
                 performance of joins in the context of complex queries.
                 For most of the joins, updates to join indices incur
                 very little overhead. Some properties of a join index
                 are (i) its efficient use of memory and adaptiveness to
                 parallel execution, (ii) its compatibility with other
                 operations (including select and union), (iii) its
                 support for abstract data type join predicates, (iv)
                 its support for multirelation clustering, and (v) its
                 use in representing directed graphs and in evaluating
                 recursive queries. Finally, the analysis of the join
                 algorithm using join indices shows its excellent
                 performance.",
  acknowledgement = ack-nhfb,
  affiliation =  "Microelectronics \& Computer Technology Corp, Austin,
                 TX, USA",
  affiliationaddress = "Microelectronics \& Computer Technology Corp,
                 Austin, TX, USA",
  annote =       "arrays of combined indices are maintained to
                 precompute joins among tuples. The technique is very
                 similar to that implemented as ADABAS correlators.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data types; computer programming ---
                 Algorithms; data processing --- Data Structures;
                 database systems; design; join algorithm, including
                 semi-join join index with rid list from selection index
                 TODS, algorithms; JOIN index; multirelation clustering;
                 performance; Relational",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf E.1}: Data, DATA
                 STRUCTURES, Trees. {\bf E.5}: Data, FILES,
                 Organization/structure. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf H.3.1}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Content Analysis and Indexing, Indexing methods.",
}

@Article{Snodgrass:1987:TQL,
  author =       "Richard Snodgrass",
  title =        "The {Temporal Query Language TQUEL}",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "247--298",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p247-snodgrass/p247-snodgrass.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p247-snodgrass/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22956.html",
  abstract =     "Recently, attention has been focused on {\em temporal
                 databases}, representing an enterprise over time. We
                 have developed a new language, {\em Tquel}, to query a
                 temporal database. TQuel was designed to be a minimal
                 extension, both syntactically and semantically, of
                 Quel, the query language in the Ingres relational
                 database management system. This paper discusses the
                 language informally, then provides a tuple relational
                 calculus semantics for the TQuel statements that differ
                 from their Quel counterparts, including the
                 modification statements. The three additional temporal
                 constructs defined in Tquel are shown to be direct
                 semantic analogues of Quel's where clause and target
                 list. We also discuss reducibility of the semantics to
                 Quel's semantics when applied to a static database.
                 TQuel is compared with ten other query languages
                 supporting time.",
  acmcrnumber =  "8712-1006",
  affiliation =  "Univ of North Carolina, Chapel Hill, NC, USA",
  affiliationaddress = "Univ of North Carolina, Chapel Hill, NC, USA",
  annote =       "Describes extensions to Quel to handle temporal
                 queries. Three kinds of temporal information are
                 handled: `Transaction time', when information was
                 stored in the database, `valid time' when the stored
                 info models reality, and `user-defined time' explicitly
                 stored by user in the database.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; database systems;
                 relational calculus; temporal databases; temporal query
                 language; theory; TQUEL; tuple calculus, languages",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.7}: Information
                 Systems, DATABASE MANAGEMENT, Database Administration,
                 Logging and recovery. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Query languages. {\bf
                 H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, TQUEL.",
}

@Article{Wong:1987:MIR,
  author =       "S. K. M. Wong and W. Ziarko and V. V. Raghavan and P.
                 C. N. Wong",
  title =        "On Modeling of Information Retrieval Concepts in
                 Vector Spaces",
  journal =      j-TODS,
  volume =       "12",
  number =       "2",
  pages =        "299--321",
  month =        jun,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-2/p299-wong/p299-wong.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-2/p299-wong/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/22957.html",
  abstract =     "The Vector Space Model (VSM) has been adopted in
                 information retrieval as a means of coping with inexact
                 representation of documents and queries, and the
                 resulting difficulties in determining the relevance of
                 a document relative to a given query. The major problem
                 in employing this approach is that the explicit
                 representation of term vectors is not known a priori.
                 Consequently, earlier researchers made the assumption
                 that the vectors corresponding to terms are pairwise
                 orthogonal. Such an assumption is clearly unrealistic.
                 Although attempts have been made to compensate for this
                 assumption by some separate, corrective steps, such
                 methods are ad hoc and, in most cases, formally
                 inconsistent.\par

                 In this paper, a generalization of the VSM, called the
                 GVSM, is advanced. The developments provide a solution
                 not only for the computation of a measure of similarity
                 (correlation) between terms, but also for the
                 incorporation of these similarities into the retrieval
                 process.\par

                 The major strength of the GVSM derives from the fact
                 that it is theoretically sound and elegant.
                 Furthermore, experimental evaluation of the model on
                 several test collections indicates that the performance
                 is better than that of the VSM. Experiments have been
                 performed on some variations of the GVSM, and all these
                 results have also been compared to those of the VSM,
                 based on inverse document frequency weighting. These
                 results and some ideas for the efficient implementation
                 of the GVSM are discussed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Regina, Regina, Sask, Can",
  affiliationaddress = "Univ of Regina, Regina, Sask, Can",
  annote =       "The space for both documents and queries is an
                 n-dimensional vector space. In GVSM, terms are not
                 assumed to be linearly independent; measure of
                 independence is based on the number of common
                 documents.",
  classification = "723; 903",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Experimentation; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "document representation; experimentation; generalized
                 vector space; Information Retrieval; information
                 retrieval systems --- Mathematical Models; information
                 science; languages; query representation, design;
                 theory",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Retrieval models. {\bf H.3.1}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Content Analysis and
                 Indexing, Thesauruses. {\bf H.3.1}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Content
                 Analysis and Indexing, Indexing methods. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval.",
}

@Article{Rybinski:1987:FOL,
  author =       "Henryk Rybi{\'n}ski",
  title =        "On First-Order-Logic Databases",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "325--349",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (03B70)",
  MRnumber =     "88j:68033",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p325-rybinski/p325-rybinski.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p325-rybinski/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/27630.html",
  abstract =     "The use of first-order logic as database logic is
                 shown to be powerful enough for formalizing and
                 implementing not only relational but also hierarchical
                 and network-type databases. It enables one to treat all
                 the types of databases in a uniform manner. This paper
                 focuses on the database language for heterogeneous
                 databases. The language is shown to be general enough
                 to specify constraints for a particular type of
                 database, so that a specification of database type can
                 be ``translated'' to the specification given in the
                 database language, creating a ``logical environment''
                 for different views that can be defined by users. Owing
                 to the fact that any database schema is seen as a
                 first-order theory expressed by a finite set of
                 sentences, the problems concerned with completeness and
                 compactness of the database logic discussed by Jacobs
                 (``On Database Logic,'' {\em J. ACM 29\/}, 2 (Apr.
                 1982), 310-332) are avoided.",
  acknowledgement = ack-nhfb,
  annote =       "Successor of Jacobs's work. Language Ld can specify
                 constraints for any database type, which can then be
                 `translated' to a particular database domain.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer metatheory --- formal logic; database logic;
                 design; first-order logic; hierarchical databases;
                 languages; network databases; relational databases;
                 theory, database systems",
  review =       "ACM Computing Reviews, Jan 1989",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML). {\bf F.4.1}: Theory of
                 Computation, MATHEMATICAL LOGIC AND FORMAL LANGUAGES,
                 Mathematical Logic. {\bf I.2.4}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Knowledge
                 Representation Formalisms and Methods, Predicate
                 logic.",
}

@Article{Stonebraker:1987:EDS,
  author =       "Michael Stonebraker and Jeff Anton and Eric Hanson",
  title =        "Extending a Database System with Procedures",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "350--376",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: UCB/ERL memo M85/59, 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p350-stonebraker/p350-stonebraker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p350-stonebraker/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/27631.html",
  abstract =     "This paper suggests that more powerful database
                 systems (DBMS) can be built by supporting database
                 procedures as full-fledged database objects. In
                 particular, allowing fields of a database to be a
                 collection of queries in the query language of the
                 system is shown to allow the natural expression of
                 complex data relationships. Moreover, many of the
                 features present in object-oriented systems and
                 semantic data models can be supported by this facility.
                 \par

                 In order to implement this construct, extensions to a
                 typical relational query language must be made, and
                 considerable work on the execution engine of the
                 underlying DBMS must be accomplished. This paper
                 reports on the extensions for one particular query
                 language and data manager and then gives performance
                 figures for a prototype implementation. Even though the
                 performance of the prototype is competitive with that
                 of a conventional system, suggestions for improvement
                 are presented.",
  acknowledgement = ack-nhfb,
  annote =       "Stored procedures follow DBTG suggestions from 1971.
                 The INGRES+ results were `competitive'.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database procedures, design; database systems;
                 object-oriented systems; relational query language;
                 semantic data models",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing.",
}

@Article{Ozsoyoglu:1987:RMM,
  author =       "Z. Meral {\"O}zsoyo{\u{g}}lu and Li-Yan Yuan",
  title =        "Reduced {MVDs} and Minimal Covers",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "377--394",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "88h:68017",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p377-ozsoyoolu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214286.html",
  abstract =     "Multivalued dependencies (MVDs) are data dependencies
                 that appear frequently in the ``real world'' and play
                 an important role in designing relational database
                 schemes. Given a set of MVDs to constrain a database
                 scheme, it is desirable to obtain an equivalent set of
                 MVDs that do not have any redundancies. In this paper
                 we define such a set of MVDs, called reduced MVDs, and
                 present an algorithm to obtain reduced MVDs. We also
                 define a minimal cover of a set of MVDs, which is a set
                 of reduced MVDs, and give an efficient method to find
                 such a minimal cover. The significance and properties
                 of reduced MVDs are also discussed in the context of
                 database design (e.g., 4NF decomposition) and
                 conflict-free MVDs.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; computer programming --- algorithms;
                 database schema design; database systems; design;
                 minimal covers; multivalued dependencies; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Normal
                 forms.",
}

@Article{Faloutsos:1987:OSE,
  author =       "Christos Faloutsos and Stavros Christodoulakis",
  title =        "Optimal Signature Extraction and Information Loss",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "395--428",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p395-faloutsos/p395-faloutsos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p395-faloutsos/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214285.html",
  abstract =     "Signature files seem to be a promising access method
                 for text and attributes. According to this method, the
                 documents (or records) are stored sequentially in one
                 file (``text file''), while abstractions of the
                 documents (``signatures'') are stored sequentially in
                 another file (``signature file''). In order to resolve
                 a query, the signature file is scanned first, and many
                 nonqualifying documents are immediately rejected. We
                 develop a framework that includes primary key hashing,
                 multiattribute hashing, and signature files. Our effort
                 is to find the optimal signature extraction method.
                 \par

                 The main contribution of this paper is that we present
                 optimal and efficient suboptimal algorithms for
                 assigning words to signatures in several environments.
                 Another contribution is that we use information theory,
                 and study the relationship of the false drop
                 probability $ F_d $ and the information that is lost
                 during signature extraction. We give tight lower bounds
                 on the achievable $ F_d $ and show that a simple
                 relationship holds between the two quantities in the
                 case of optimal signature extraction with uniform
                 occurrence and query frequencies. We examine hashing as
                 a method to map words to signatures (instead of the
                 optimal way), and show that the same relationship holds
                 between $ F_d $ and {\em loss}, indicating that an
                 invariant may exist between these two quantities for
                 every signature extraction method.",
  acknowledgement = ack-nhfb,
  annote =       "superimposed coding",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing; database systems; information loss;
                 optimal signature extraction; performance; signature
                 files, design",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf E.4}: Data,
                 CODING AND INFORMATION THEORY, Data compaction and
                 compression. {\bf E.5}: Data, FILES. {\bf H.3.2}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Storage, File organization.",
}

@Article{Ibaraki:1987:SC,
  author =       "Toshihide Ibaraki and Tiko Kameda and Toshimi
                 Minoura",
  title =        "Serializability with Constraints",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "429--452",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "909 139",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p429-ibaraki/p429-ibaraki.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p429-ibaraki/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214284.html",
  abstract =     "This paper deals with the serializability theory for
                 single-version and multiversion database systems. We
                 first introduce the concept of {\em disjoint-interval
                 topological sort\/} ({\em DITS}, for short) of an
                 arc-labeled directed acyclic graph. It is shown that a
                 history is serializable if and only if its {\em
                 transaction IO graph\/} has a DITS. We then define
                 several subclasses of serializable histories, based on
                 the constraints imposed by write-write, write-read,
                 read-write, or read-read conflicts, and investigate
                 inclusion relationships among them. In terms of DITS,
                 we give a sufficient condition for a class of
                 serializable histories to be polynomially recognizable,
                 which is then used to show that a new class of
                 histories, named WRW, can be recognized in polynomial
                 time. We also present NP-completeness results for the
                 problem of testing membership in some other classes.
                 \par

                 In the second half of this paper, we extend these
                 results to multiversion database systems. The inclusion
                 relationships among multiversion classes defined by
                 constraints, such as write-write and write-read, are
                 investigated. One such class coincides with class
                 DMVSR, introduced by Papadimitriou and Kanellakis, and
                 gives a simple characterization of this class. It is
                 shown that for most constraints, multiversion classes
                 properly contain the corresponding single-version
                 classes. Complexity results for the membership testing
                 are also discussed.",
  acknowledgement = ack-nhfb,
  annote =       "classification and properties of conflict graphs, with
                 and without versions.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer systems programming --- sorting; concurrency
                 control; database systems; disjoint-interval
                 topological sort, algorithms; serializability; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing.",
}

@Article{Wolfson:1987:OLC,
  author =       "Ouri Wolfson",
  title =        "The Overhead of Locking (and Commit) Protocols in
                 Distributed Databases",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "453--471",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p453-wolfson/p453-wolfson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p453-wolfson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/28053.html",
  abstract =     "The main purpose of a locking protocol is to ensure
                 correct interleaving of actions executed by concurrent
                 transactions. The locking protocol consists of a set of
                 rules dictating how accessed entities should be locked
                 and unlocked. As a result of obeying the rules,
                 transactions in a distributed database incur an
                 overhead. We propose three measures of evaluating this
                 overhead, each most suitable to a different type of
                 underlying communication network. Then, using a graph
                 theoretic model, we analyze and compare three protocols
                 according to each measure: two-phase locking, two-phase
                 locking with a fixed order imposed on the database
                 entities (ensuring deadlock freedom), and the tree
                 protocol. In practice, a transaction also executes the
                 two-phase commit protocol in order to guarantee
                 atomicity. Therefore, the combined overhead of each
                 locking protocol and the two-phase commit protocol is
                 also determined.",
  acknowledgement = ack-nhfb,
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Measurement; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "commit protocols; computer networks --- protocols;
                 concurrency control, algorithms; database systems;
                 locking protocols; measurement; message passing;
                 performance; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf C.4}: Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS. {\bf C.2.4}:
                 Computer Systems Organization, COMPUTER-COMMUNICATION
                 NETWORKS, Distributed Systems, Distributed databases.
                 {\bf D.2.8}: Software, SOFTWARE ENGINEERING, Metrics,
                 Performance measures. {\bf H.2.2}: Information Systems,
                 DATABASE MANAGEMENT, Physical Design. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing.",
}

@Article{Su:1987:CBD,
  author =       "Stanley Y. W. Su and Jozo Dujmovic and D. S. Batory
                 and S. B. Navathe and Richard Elnicki",
  title =        "A Cost-Benefit Decision Model: Analysis, Comparison,
                 and Selection of Data Management Systems",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "472--520",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p472-su/p472-su.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p472-su/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/33403.html",
  abstract =     "This paper describes a general cost-benefit decision
                 model that is applicable to the evaluation, comparison,
                 and selection of alternative products with a
                 multiplicity of features, such as complex computer
                 systems. The application of this model is explained and
                 illustrated using the selection of data management
                 systems as an example.\par

                 The model has the following features: (1) it is
                 mathematically based on an extended continuous logic
                 and a theory of complex criteria; (2) the
                 decision-making procedure is very general yet
                 systematic, well-structured, and quantitative; (3) the
                 technique is based on a comprehensive cost analysis and
                 an elaborate analysis of benefits expressed in terms of
                 the decision maker's preferences. The decision
                 methodology, when applied to the problem of selecting a
                 data management system, takes into consideration the
                 life cycle of a DMS and the objectives and goals for
                 the new systems under evaluation. It allows the cost
                 and preference analyses to be carried out separately
                 using two different models. The model for preference
                 analysis makes use of comprehensive performance (or
                 preference) parameters and allows what we call a
                 ``logic scoring of preferences'' using continuous
                 values between zero and one, to express the degree with
                 which candidate systems satisfy stated requirements. It
                 aggregates preference parameters based on their
                 relative weights and logical relationships to compute a
                 global performance (preference) score for each system.
                 The cost model incorporates an aggregation of costs
                 which may be estimated over different time horizons and
                 discounted at appropriate discount rates. A procedure
                 to establish an overall ranking of alternative systems
                 based on their global preference scores and global
                 costs is also discussed.",
  acknowledgement = ack-nhfb,
  annote =       "The LPS model results from a NBS study. Preference
                 scoring, facility costs.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cost-benefit decision model; data management systems,
                 SYWSU TODS; database systems; mathematical models",
  subject =      "{\bf H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration. {\bf K.6.3}: Computing
                 Milieux, MANAGEMENT OF COMPUTING AND INFORMATION
                 SYSTEMS, Software Management. {\bf D.2.9}: Software,
                 SOFTWARE ENGINEERING, Management, Cost estimation.",
}

@Article{Keller:1987:CBS,
  author =       "Arthur M. Keller",
  title =        "Comment on {Bancilhon} and {Spyratos}' {``Update
                 semantics and relational views''}",
  journal =      j-TODS,
  volume =       "12",
  number =       "3",
  pages =        "521--523",
  month =        sep,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Bancilhon:1981:USR}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-3/p521-keller/p521-keller.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-3/p521-keller/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/214296.html",
  acknowledgement = ack-nhfb,
  annote =       "a small relaxation of constraints increases the
                 feasibility of view update",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; theory",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Retrieval models. {\bf H.2.1}: Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Data models.",
}

@Article{Abiteboul:1987:IFS,
  author =       "Serge Abiteboul and Richard Hull",
  title =        "{IFO}: a Formal Semantic Database Model",
  journal =      j-TODS,
  volume =       "12",
  number =       "4",
  pages =        "525--565",
  month =        dec,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "88i:68017",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Also published in/as: ACM SIGACT-SIGMOD Symp. on
                 Principles of Database Systems, 1984 (short version).
                 Also published in/as: USC CSD, TR-84-304, Apr. 1984.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-4/p525-abiteboul/p525-abiteboul.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-4/p525-abiteboul/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/32205.html",
  abstract =     "A new, formally defined database model is introduced
                 that combines fundamental principles of ``semantic''
                 database modeling in a coherent fashion. Using a
                 graph-based formalism, the IFO model provides
                 mechanisms for representing structured objects, and
                 functional and ISA relationships between them. A number
                 of fundamental results concerning semantic data
                 modeling are obtained in the context of the IFO model.
                 Notably, the types of object structure that can arise
                 as a result of multiple uses of ISA relationships and
                 object construction are described. Also, a natural,
                 formal definition of update propagation is given, and
                 it is shown that (under certain conditions) a correct
                 update always exists.",
  acknowledgement = ack-nhfb,
  annote =       "A graph-based formal semantic database model. Well
                 written. The model appears to be well founded.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; theory; verification",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf I.2.4}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Knowledge
                 Representation Formalisms and Methods, Semantic
                 networks.",
}

@Article{Ozsoyoglu:1987:ERA,
  author =       "G. {\"O}zsoyo{\u{g}}lu and Z. M. {\"O}zsoyo{\u{g}}lu
                 and V. Matos",
  title =        "Extending Relational Algebra and Relational Calculus
                 with Set-Valued Attributes and Aggregate Functions",
  journal =      j-TODS,
  volume =       "12",
  number =       "4",
  pages =        "566--592",
  month =        dec,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "920 253",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-4/p566-ozsoyoglu/p566-ozsoyoglu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-4/p566-ozsoyoglu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/32219.html",
  abstract =     "In commercial network database management systems,
                 set-valued fields and aggregate functions are commonly
                 supported. However, the relational database model, as
                 defined by Codd, does not include set-valued attributes
                 or aggregate functions. Recently, Klug extended the
                 relational model by incorporating aggregate functions
                 and by defining relational algebra and calculus
                 languages.\par

                 In this paper, relational algebra and relational
                 calculus database query languages (as defined by Klug)
                 are extended to manipulate set-valued attributes and to
                 utilize aggregate functions. The expressive power of
                 the extended languages is shown to be equivalent. We
                 extend the relational algebra with three new operators,
                 namely, pack, unpack, and aggregation-by-template. The
                 extended languages form a theoretical framework for
                 statistical database query languages.",
  acknowledgement = ack-nhfb,
  annote =       "Nested relations.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; languages; theory; verification;
                 {\"O}zsoyo{\u{g}}lu Nested NF2 TODS",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages.",
}

@Article{Palley:1987:URM,
  author =       "Michael A. Palley and Jeffrey S. Simonoff",
  title =        "The Use of Regression Methodology for the Compromise
                 of Confidential Information in Statistical Databases",
  journal =      j-TODS,
  volume =       "12",
  number =       "4",
  pages =        "593--608",
  month =        dec,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-4/p593-palley/p593-palley.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-4/p593-palley/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42174.html",
  abstract =     "A regression methodology based technique can be used
                 to compromise confidentiality in a statistical
                 database. This holds true even when the DBMS prevents
                 application of regression methodology to the database.
                 Existing inference controls, including cell
                 restriction, perturbation, and table restriction
                 approaches, are shown to be generally ineffective
                 against this compromise technique. The effect of
                 incomplete supplemental knowledge on the regression
                 methodology based compromise technique is examined.
                 Finally, some potential complicators of this disclosure
                 scheme are introduced.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Legal Aspects; Management; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "legalaspects; management; security, security TODS",
  subject =      "{\bf H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration. {\bf H.2.8}: Information
                 Systems, DATABASE MANAGEMENT, Database applications.
                 {\bf K.4.1}: Computing Milieux, COMPUTERS AND SOCIETY,
                 Public Policy Issues, Privacy. {\bf K.6.m}: Computing
                 Milieux, MANAGEMENT OF COMPUTING AND INFORMATION
                 SYSTEMS, Miscellaneous, Security*. {\bf H.2.0}:
                 Information Systems, DATABASE MANAGEMENT, General,
                 Security, integrity, and protection. {\bf G.3}:
                 Mathematics of Computing, PROBABILITY AND STATISTICS,
                 Statistical computing.",
}

@Article{Agrawal:1987:CCP,
  author =       "Rakesh Agrawal and Michael J. Carey and Miron Livny",
  title =        "Concurrency Control Performance Modeling: Alternatives
                 and Implications",
  journal =      j-TODS,
  volume =       "12",
  number =       "4",
  pages =        "609--654",
  month =        dec,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/real.time.bib",
  note =         "Also published in/as: ACM SIGMOD Conf. on the
                 Management of Data, 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-4/p609-agrawal/p609-agrawal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-4/p609-agrawal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/32220.html",
  abstract =     "A number of recent studies have examined the
                 performance of concurrency control algorithms for
                 database management systems. The results reported to
                 date, rather than being definitive, have tended to be
                 contradictory. In this paper, rather than presenting
                 ``yet another algorithm performance study,'' we
                 critically investigate the assumptions made in the
                 models used in past studies and their implications. We
                 employ a fairly complete model of a database
                 environment for studying the relative performance of
                 three different approaches to the concurrency control
                 problem under a variety of modeling assumptions. The
                 three approaches studied represent different extremes
                 in how transaction conflicts are dealt with, and the
                 assumptions addressed pertain to the nature of the
                 database system's resources, how transaction restarts
                 are modeled, and the amount of information available to
                 the concurrency control algorithm about transactions'
                 reference strings. We show that differences in the
                 underlying assumptions explain the seemingly
                 contradictory performance results. We also address the
                 question of how realistic the various assumptions are
                 for actual database systems.",
  acknowledgement = ack-nhfb,
  annote =       "This paper is unique in that it studies the
                 implications of fundamental assumptions regarding
                 system resources, transactions restarts, and
                 predeclarations of writes. Essential reading.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "performance",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf D.4.8}: Software, OPERATING
                 SYSTEMS, Performance, Simulation. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing. {\bf D.4.2}: Software,
                 OPERATING SYSTEMS, Storage Management.",
}

@Article{Sacks-Davis:1987:MAM,
  author =       "R. Sacks-Davis and A. Kent and K. Ramamohanarao",
  title =        "Multikey Access Methods Based on Superimposed Coding
                 Techniques",
  journal =      j-TODS,
  volume =       "12",
  number =       "4",
  pages =        "655--696",
  month =        dec,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1987-12-4/p655-sacks-davis/p655-sacks-davis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1987-12-4/p655-sacks-davis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/32222.html",
  abstract =     "Both single-level and two-level indexed descriptor
                 schemes for multikey retrieval are presented and
                 compared. The descriptors are formed using superimposed
                 coding techniques and stored using a bit-inversion
                 technique. A fast-batch insertion algorithm for which
                 the cost of forming the bit-inverted file is less than
                 one disk access per record is presented. For large data
                 files, it is shown that the two-level implementation is
                 generally more efficient for queries with a small
                 number of matching records. For queries that specify
                 two or more values, there is a potential problem with
                 the two-level implementation in that costs may accrue
                 when blocks of records match the query but individual
                 records within these blocks do not. One approach to
                 overcoming this problem is to set bits in the
                 descriptors based on pairs of indexed terms. This
                 approach is presented and analyzed.",
  acknowledgement = ack-nhfb,
  annote =       "Expressions for the cost of a two-level and one-level
                 scheme.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "descriptors; Design; hashing; partial match retrieval;
                 performance; record signatures; superimposed coding,
                 design",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.3.2}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Storage, File organization. {\bf H.4.1}:
                 Information Systems, INFORMATION SYSTEMS APPLICATIONS,
                 Office Automation. {\bf I.7.m}: Computing
                 Methodologies, TEXT PROCESSING, Miscellaneous.",
}

@Article{Elhardt:1987:SQO,
  author =       "K. Elhardt",
  title =        "Support for Query Optimization by Optimal Predicate
                 Splitting",
  journal =      j-TODS,
  volume =       "??",
  number =       "??",
  pages =        "??--??",
  month =        mar,
  year =         "1987",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Accepted.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Vianu:1988:DFO,
  author =       "Victor Vianu",
  title =        "A Dynamic Framework for Object Projection Views",
  journal =      j-TODS,
  volume =       "13",
  number =       "1",
  pages =        "1--22",
  month =        mar,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "933 215",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-1/p1-vianu/p1-vianu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-1/p1-vianu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42202.html",
  abstract =     "User views in a relational database obtained through a
                 single projection ('projection views') are considered
                 in a new framework. Specifically, such views, where
                 each tuple in the view represents an object
                 ('object-projection views'), are studied using the
                 dynamic relational model, which captures the evolution
                 of the database through consecutive updates. Attribute
                 sets that yield object-projection views are
                 characterized using the static and dynamic functional
                 dependencies satisfied by the database.
                 Object-projection views are then described using the
                 static and dynamic functional dependencies `inherited'
                 from the original database. Finally, the impact of
                 dynamic constraints on the view update problem is
                 studied in a limited context. This paper demonstrates
                 that new, useful information about views can be
                 obtained by looking at the evolution of the database as
                 captured by the dynamic relational model.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, San Diego, CA, USA",
  affiliationaddress = "Univ of California, San Diego, CA, USA",
  annote =       "Assumes simplest translation to the underlying
                 database. Very restrictive unirelational database FDs
                 and DFDs, no exclusions or deletions. Technically very
                 good. Rejecting update object projections.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Management; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database schema, design; database systems; dynamic
                 constraints; functional dependency; languages;
                 management; object projection views; relational;
                 theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Data description
                 languages (DDL).",
}

@Article{Sellis:1988:MQO,
  author =       "Timos K. Sellis",
  title =        "Multiple-Query Optimization",
  journal =      j-TODS,
  volume =       "13",
  number =       "1",
  pages =        "23--52",
  month =        mar,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-1/p23-sellis/p23-sellis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-1/p23-sellis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42203.html",
  abstract =     "Some recently proposed extensions to relational
                 database systems, as well as to deductive database
                 systems, require support for multiple-query processing.
                 For example, in a database system enhanced with
                 inference capabilities, a simple query involving a rule
                 with multiple definitions may expand to more than one
                 actual query that has to be run over the database. It
                 is an interesting problem then to come up with
                 algorithms that process these queries together instead
                 of one query at a time. The main motivation for
                 performing such an interquery optimization lies in the
                 fact that queries may share common data. We examine the
                 problem of multiple-query optimization in this paper.
                 The first major contribution of the paper is a
                 systematic look at the problem, along with the
                 presentation and analysis of algorithms that can be
                 used for multiple-query optimization. The second
                 contribution lies in the presentation of experimental
                 results. Our results show that using multiple-query
                 processing algorithms may reduce execution cost
                 considerably.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, Berkeley, CA, USA",
  affiliationaddress = "Univ of California, Berkeley, CA, USA",
  annote =       "Two architectures: Interleaving the results of locally
                 optimal access plans or a global optimizer.
                 Experimental results, a decrease of 20-50 percent show
                 in I/O and CPU time. ---Sava-Segal.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "common access paths; computer programming ---
                 algorithms; database systems; deductive databases;
                 heuristic methods, global query TODS, algorithms;
                 multiple-query optimization; optimization; performance;
                 Relational",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models. {\bf I.2.8}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Problem Solving, Control
                 Methods, and Search, Heuristic methods. {\bf H.3.4}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Systems and Software.",
}

@Article{Shasha:1988:CSS,
  author =       "Dennis Shasha and Nathan Goodman",
  title =        "Concurrent Search Structure Algorithms",
  journal =      j-TODS,
  volume =       "13",
  number =       "1",
  pages =        "53--90",
  month =        mar,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-1/p53-shasha/p53-shasha.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-1/p53-shasha/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42204.html",
  abstract =     "A dictionary is an abstract data type supporting the
                 actions member, insert, and delete. A search structure
                 is a data structure used to implement a dictionary.
                 Examples include B trees, hash structures, and
                 unordered lists. Concurrent algorithms on search
                 structures can achieve more parallelism than standard
                 concurrency control methods would suggest, by
                 exploiting the fact that many different search
                 structure states represent one dictionary state. We
                 present a framework for verifying such algorithms and
                 for inventing new ones. We give several examples, one
                 of which exploits the structure of Banyan family
                 interconnection networks. We also discuss the
                 interaction between concurrency control and recovery as
                 applied to search structures.",
  acknowledgement = ack-nhfb,
  affiliation =  "New York Univ, New York, NY, USA",
  affiliationaddress = "New York Univ, New York, NY, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data type; Banyan interconnection networks;
                 computer programming --- algorithms; concurrent
                 algorithms, algorithms; data processing; Data
                 Structures; design; dictionary; performance; search
                 structure",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf D.4.1}: Software, OPERATING
                 SYSTEMS, Process Management, Concurrency. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.",
}

@Article{Finkelstein:1988:PDD,
  author =       "S. J. Finkelstein and M. Schkolnick and P. Tiberio",
  title =        "Physical Database Design for Relational Databases",
  journal =      j-TODS,
  volume =       "13",
  number =       "1",
  pages =        "91--128",
  month =        mar,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Also published in/as: IBM Research Report No. RJ5034,
                 Nov. 1986, preprint IBM Aug. 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-1/p91-finkelstein/p91-finkelstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-1/p91-finkelstein/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42205.html",
  abstract =     "This paper describes the concepts used in the
                 implementation of DBDSGN, an experimental physical
                 design tool for relational databases developed at the
                 IBM San Jose Research Laboratory. Given a workload for
                 System R (consisting of a set of SQL statements and
                 their execution frequencies), DBDSGN suggests physical
                 configurations for efficient performance. Each
                 configuration consists of a set of indices and an
                 ordering for each table. Workload statements are
                 evaluated only for atomic configurations of indices,
                 which have only one index per table. Costs for any
                 configuration can be obtained from those of the atomic
                 configurations. DBDSGN uses information supplied by the
                 System R optimizer both to determine which columns
                 might be worth indexing and to obtain estimates of the
                 cost of executing statements in different
                 configurations. The tool finds efficient solutions to
                 the index-selection problem; if we assume the cost
                 estimates supplied by the optimizer are the actual
                 execution costs, it finds the optimal solution.
                 Optionally, heuristics can be used to reduce execution
                 time. The approach taken by DBDSGN in solving the
                 index-selection problem for multiple-table statements
                 significantly reduces the complexity of the problem.
                 DBDSGN's principles were used in the Relational Design
                 Tool (RDT), an IBM product based on DBDSGN, which
                 performs design for SQL/DS, a relational system based
                 on System R. System R actually uses DBDSGN's suggested
                 solutions as the tool expects because cost estimates
                 and other necessary information can be obtained from
                 System R using a new SQL statement, the EXPLAIN
                 statement. This illustrates how a system can export a
                 model of its internal assumptions and behavior so that
                 other systems (such as tools) can share this model.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM",
  affiliationaddress = "IBM",
  annote =       "DBDSGN led to IBM RDT; input are relational tables and
                 a set of queries expected to be run; produces
                 specifications of indexes, clustered; it uses the
                 actual DBMS (SQL/DS) optimizer.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance;
                 Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "DBDSGN; design; experimentation; IBM DBDSGN tool TODS,
                 database systems; optimization; performance; physical
                 database design; relational; System R, algorithms;
                 theory",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf D.4.8}: Software,
                 OPERATING SYSTEMS, Performance, Modeling and
                 prediction. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf E.1}: Data,
                 DATA STRUCTURES, Tables.",
}

@Article{Raju:1988:FFD,
  author =       "K. V. S. V. N. Raju and Arun K. Majumdar",
  title =        "Fuzzy Functional Dependencies and Lossless Join
                 Decomposition of Fuzzy Relational Database Systems",
  journal =      j-TODS,
  volume =       "13",
  number =       "2",
  pages =        "129--166",
  month =        jun,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-2/p129-raju/p129-raju.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-2/p129-raju/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42344.html",
  abstract =     "This paper deals with the application of fuzzy logic
                 in a relational database environment with the objective
                 of capturing more meaning of the data. It is shown that
                 with suitable interpretations for the fuzzy membership
                 functions, a fuzzy relational data model can be used to
                 represent ambiguities in data values as well as
                 impreciseness in the association among them. Relational
                 operators for fuzzy relations have been studied, and
                 applicability of fuzzy logic in capturing integrity
                 constraints has been investigated. By introducing a
                 fuzzy resemblance measure EQUAL for comparing domain
                 values, the definition of classical functional
                 dependency has been generalized to fuzzy functional
                 dependency (ffd). The implication problem of ffds has
                 been examined and a set of sound and complete inference
                 axioms has been proposed. Next, the problem of lossless
                 join decomposition of fuzzy relations for a given set
                 of fuzzy functional dependencies is investigated. It is
                 proved that with a suitable restriction on EQUAL, the
                 design theory of a classical relational database with
                 functional dependencies can be extended to fuzzy
                 relations satisfying fuzzy functional dependencies.",
  acknowledgement = ack-nhfb,
  affiliation =  "Ahdhra Univ, Visakhapatnam, India",
  affiliationaddress = "Ahdhra Univ, Visakhapatnam, India",
  annote =       "Extend Armstrong's axioms to fuzzy domains. The
                 concept works out theoretically, but may not
                 necessarily agree with the intuition of the database
                 user. This paper is highly theoretical and notationally
                 intimidating. Yet worth to peruse.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; fuzzy functional dependencies; fuzzy
                 relational database systems, design; lossless join
                 decomposition; mathematical techniques --- fuzzy sets;
                 relational; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf E.5}: Data, FILES. {\bf I.2.3}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Deduction and Theorem Proving, Uncertainty, ``fuzzy,''
                 and probabilistic reasoning.",
}

@Article{Winslett:1988:MBA,
  author =       "Marianne Winslett",
  title =        "A Model-Based Approach to Updating Databases with
                 Incomplete Information",
  journal =      j-TODS,
  volume =       "13",
  number =       "2",
  pages =        "167--196",
  month =        jun,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/bibdb.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-2/p167-winslett/p167-winslett.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-2/p167-winslett/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42386.html",
  abstract =     "Suppose one wishes to construct, use, and maintain a
                 database of facts about the real world, even though the
                 state of that world is only partially known. In the
                 artificial intelligence domain, this problem arises
                 when an agent has a base set of beliefs that reflect
                 partial knowledge about the world, and then tries to
                 incorporate new, possibly contradictory knowledge into
                 this set of beliefs. In the database domain, one facet
                 of this situation is the well-known null values
                 problem. We choose to represent such a database as a
                 logical theory, and view the models of the theory as
                 representing possible states of the world that are
                 consistent with all known information.\par

                 How can new information be incorporated into the
                 database? For example, given the new information that
                 ``$b$ or $c$ is true,'' how can one get rid of all
                 outdated information about $b$ and $c$, add the new
                 information, and yet in the process not disturb any
                 other information in the database? In current-day
                 database management systems, the difficult and tedious
                 burden of determining exactly what to add and remove
                 from the database is placed on the user. The goal of
                 our research was to relieve users of that burden, by
                 equipping the database management system with update
                 algorithms that can automatically determine what to add
                 and remove from the database. \par

                 Under our approach, new information about the state of
                 the world is input to the database management system as
                 a well-formed formula that the state of the world is
                 now known to satisfy. We have constructed database
                 update algorithms to interpret this update formula and
                 incorporate the new information represented by the
                 formula into the database without further assistance
                 from the user. In this paper we show how to embed the
                 incomplete database and the incoming information in the
                 language of mathematical logic, explain the semantics
                 of our update operators, and discuss the algorithms
                 that implement these operators.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Illinois, Urbana, IL, USA",
  affiliationaddress = "Univ of Illinois, Urbana, IL, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Human Factors; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; database systems;
                 database updates; humanfactors; incomplete information;
                 languages; theory; uncertainty, algorithms; update
                 algorithms",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems. {\bf H.3.0}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, General. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Retrieval
                 models. {\bf I.2.3}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem
                 Proving.",
}

@Article{Eich:1988:DCC,
  author =       "Margaret H. Eich and David L. Wells",
  title =        "Database Concurrency Control using Data Flow Graphs",
  journal =      j-TODS,
  volume =       "13",
  number =       "2",
  pages =        "197--227",
  month =        jun,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "943 408",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-2/p197-eich/p197-eich.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-2/p197-eich/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/42345.html",
  abstract =     "A specialized data flow graph, {\em Database Flow
                 Graph\/} ({\em DBFG\/}) is introduced. DBFGs may be
                 used for scheduling database operations, particularly
                 in an MIMD database machine environment. A DBFG
                 explicitly maintains intertransaction and
                 intratransaction dependencies, and is constructed from
                 the Transaction Flow Graphs (TFG) of active
                 transactions. A TFG, in turn, is the generalization of
                 a query tree used, for example, in DIRECT [15].
                 \par

                 All DBFG schedules are serializable and deadlock free.
                 Operations needed to create and maintain the DBFG
                 structure as transactions are added or removed from the
                 system are discussed. Simulation results show that DBFG
                 scheduling performs as well as two-phase locking.",
  acknowledgement = ack-nhfb,
  affiliation =  "Southern Methodist Univ, Dallas, TX, USA",
  affiliationaddress = "Southern Methodist Univ, Dallas, TX, USA",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data flow graphs; database concurrency control;
                 database systems; deadlock, algorithms; distributed;
                 mathematical techniques --- graph theory; performance;
                 serializability",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Deadlock avoidance. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Transaction processing.",
}

@Article{Batory:1988:ICE,
  author =       "D. S. Batory and T. Y. Leung and T. E. Wise",
  title =        "Implementation Concepts for an Extensible Data Model
                 and Data Language",
  journal =      j-TODS,
  volume =       "13",
  number =       "3",
  pages =        "231--262",
  month =        sep,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Un. Texas, TR-86-24, Oct.
                 1986.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-3/p231-batory/p231-batory.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-3/p231-batory/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/45062.html",
  abstract =     "Future database systems must feature extensible data
                 models and data languages in order to accommodate the
                 novel data types and special-purpose operations that
                 are required by nontraditional database applications.
                 In this paper, we outline a functional data model and
                 data language that are targeted for the semantic
                 interface of GENESIS, an extensible DBMS. The model and
                 language are generalizations of FQL [11] and DAPLEX
                 [40], and have an implementation that fits ideally with
                 the modularity required by extensible database
                 technologies. We explore different implementations of
                 functional operators and present experimental evidence
                 that they have efficient implementations. We also
                 explain the advantages of a functional front-end to 1NF
                 databases, and show how our language and implementation
                 are being used to process queries on both 1NF and 1NF
                 relations.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Texas at Austin, Austin, TX, USA",
  affiliationaddress = "Univ of Texas at Austin, Austin, TX, USA",
  annote =       "Illustrates a functional data model of GENESIS.
                 Defines GDM, the data model, and GDL, the data
                 language. GDM productions are stream rewrite rules and
                 GDM computations are expressed as streams of tokens.
                 Promotes extensibility. To build a non-1NF query
                 processor is an example.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming languages; data language;
                 database systems; functional data model; languages;
                 lazy evaluation, TODS genesis functional language join
                 joins lazy eager nested relations, design; Management;
                 performance; stream translators",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Data manipulation languages (DML). {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing. {\bf I.1.3}: Computing Methodologies,
                 ALGEBRAIC MANIPULATION, Languages and Systems,
                 Evaluation strategies. {\bf I.1.3}: Computing
                 Methodologies, ALGEBRAIC MANIPULATION, Languages and
                 Systems, Nonprocedural languages.",
}

@Article{Apers:1988:DAD,
  author =       "Peter M. G. Apers",
  title =        "Data Allocation in Distributed Database Systems",
  journal =      j-TODS,
  volume =       "13",
  number =       "3",
  pages =        "263--304",
  month =        sep,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-3/p263-apers/p263-apers.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-3/p263-apers/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/45063.html",
  abstract =     "The problem of allocating the data of a database to
                 the sites of a communication network is investigated.
                 This problem deviates from the well-known file
                 allocation problem in several aspects. First, the
                 objects to be allocated are not known a priori; second,
                 these objects are accessed by schedules that contain
                 transmissions between objects to produce the result. A
                 model that makes it possible to compare the cost of
                 allocations is presented; the cost can be computed for
                 different cost functions and for processing schedules
                 produced by arbitrary query processing algorithms.
                 \par

                 For minimizing the total transmission cost, a method is
                 proposed to determine the fragments to be allocated
                 from the relations in the conceptual schema and the
                 queries and updates executed by the users.\par

                 For the same cost function, the complexity of the data
                 allocation problem is investigated. Methods for
                 obtaining optimal and heuristic solutions under various
                 ways of computing the cost of an allocation are
                 presented and compared.\par

                 Two different approaches to the allocation management
                 problem are presented and their merits are discussed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Vrije Univ, Enschede, Neth",
  affiliationaddress = "Vrije Univ, Enschede, Neth",
  annote =       "Fragments are allocated. The strength of the paper is
                 its rigor, the weakness is in the applicability of the
                 model.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Measurement; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer networks; computer software --- software
                 engineering; data allocation; database systems; design;
                 distributed; dynamic schedules; greedy method;
                 heuristic allocations, algorithms; measurement; optimal
                 allocations; static schedules; theory",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 D.2.8}: Software, SOFTWARE ENGINEERING, Metrics,
                 Performance measures.",
}

@Article{Storey:1988:MCU,
  author =       "Veda C. Storey and Robert C. Goldstein",
  title =        "A Methodology for Creating User Views in Database
                 Design",
  journal =      j-TODS,
  volume =       "13",
  number =       "3",
  pages =        "305--338",
  month =        sep,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-3/p305-storey/p305-storey.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-3/p305-storey/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/45064.html",
  abstract =     "The View Creation System (VCS) is an expert system
                 that engages a user in a dialogue about the information
                 requirements for some application, develops an
                 Entity-Relationship model for the user's database view,
                 and then converts the E-R model to a set of Fourth
                 Normal Form relations. This paper describes the
                 knowledge base of VCS. That is, it presents a formal
                 methodology, capable of mechanization as a computer
                 program, for accepting requirements from a user,
                 identifying and resolving inconsistencies,
                 redundancies, and ambiguities, and ultimately producing
                 a normalized relational representation. Key aspects of
                 the methodology are illustrated by applying VCS's
                 knowledge base to an actual database design task.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Rochester, Rochester, NY, USA",
  affiliationaddress = "Univ of Rochester, Rochester, NY, USA",
  annote =       "Engages the user in a dialogue about the information
                 requirements for some application, develops an E-R
                 model and Fourth Normal Form relations.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming; database systems; design; expert
                 system; knowledge base; user views, design; view
                 creation system (VCS)",
  subject =      "{\bf I.2.1}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Applications and Expert Systems. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.2.7}: Information Systems,
                 DATABASE MANAGEMENT, Database Administration.",
}

@Article{Diederich:1988:NMF,
  author =       "Jim Diederich and Jack Milton",
  title =        "New Methods and Fast Algorithms for Database
                 Normalization",
  journal =      j-TODS,
  volume =       "13",
  number =       "3",
  pages =        "339--365",
  month =        sep,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 196",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "Also published in/as: UCD, Math, revised Jun. 1987.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-3/p339-diederich/p339-diederich.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-3/p339-diederich/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/44499.html",
  abstract =     "A new method for computing minimal covers is presented
                 using a new type of closure that allows significant
                 reductions in the number of closures computed for
                 normalizing relations. Benchmarks are reported
                 comparing the new and the standard techniques.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of California, Davis, CA, USA",
  affiliationaddress = "Univ of California, Davis, CA, USA",
  annote =       "Modify existing algorithms to make the process
                 faster.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; database
                 normalization; database systems; design; fast
                 algorithms; functional dependency; management;
                 redundant dependencies; relational database,
                 algorithms",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms.",
}

@Article{Larson:1988:LHS,
  author =       "Per-{\AA}ke Larson",
  title =        "Linear Hashing with Separators --- {A} Dynamic Hashing
                 Scheme Achieving One-Access Retrieval",
  journal =      j-TODS,
  volume =       "13",
  number =       "3",
  pages =        "366--388",
  month =        sep,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-3/p366-larson/p366-larson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-3/p366-larson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/44500.html",
  abstract =     "A new dynamic hashing scheme is presented. Its most
                 outstanding feature is that any record can be retrieved
                 in exactly one disk access. This is achieved by using a
                 small amount of supplemental internal storage that
                 stores enough information to uniquely determine the
                 current location of any record. The amount of internal
                 storage required is small: typically one byte for each
                 page of the file. The necessary address computation,
                 insertion, and expansion algorithms are presented and
                 the performance is studied by means of simulation. The
                 new method is the first practical method offering
                 one-access retrieval for large dynamic files.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ of Waterloo, Waterloo, Ont, Can",
  affiliationaddress = "Univ of Waterloo, Waterloo, Ont, Can",
  annote =       "A small amount of supplemental internal storage stores
                 enough information to uniquely determine the current
                 location of any record.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "address computation; data processing --- File
                 organization; design; dynamic hashing schemes;
                 extendible hashing; linear hashing; linear probing;
                 Management; one-access retrieval; open addressing,
                 algorithms; open addressing, database systems;
                 performance",
  review =       "ACM CR 8811-0850",
  subject =      "{\bf D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, Access methods. {\bf D.4.3}: Software,
                 OPERATING SYSTEMS, File Systems Management, File
                 organization. {\bf E.5}: Data, FILES,
                 Organization/structure. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods.",
}

@Article{Roth:1988:EAC,
  author =       "Mark A. Roth and Henry F. Korth and Abraham
                 Silberschatz",
  title =        "Extended Algebra and Calculus for Nested Relational
                 Databases",
  journal =      j-TODS,
  volume =       "13",
  number =       "4",
  pages =        "389--417",
  month =        dec,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 197",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/database.bib;
                 Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  note =         "See comment \cite{Tansel:1992:MRH}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-4/p389-roth/p389-roth.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-4/p389-roth/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/49347.html",
  abstract =     "Relaxing the assumption that relations are always in
                 First-Normal-Form (1NF) necessitates a reexamination of
                 the fundamentals of relational database theory. In this
                 paper we take a first step towards unifying the various
                 theories of $ \not $1NF databases. We start by
                 determining an appropriate model to couch our
                 formalisms in. We then define an extended relational
                 calculus as the theoretical basis for our $ \not $1NF
                 relational calculus. We define a class of $ \not $1NF
                 relations with certain ``good'' properties and extend
                 our algebra operators to work within this domain. We
                 prove certain desirable equivalences that hold only if
                 we restrict our language to this domain.",
  acknowledgement = ack-nhfb,
  affiliation =  "Texas Univ., Austin, TX, USA",
  affiliationaddress = "Austin, TX, USA",
  annote =       "Only two new operators, nest and unnest.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "languages; mathematical techniques --- algebra; nested
                 relational databases; non-first normal form database;
                 partitioned normal forms; query languages; relational;
                 relational calculus; theory, extended algebra and
                 calculus, nested relations, non-first normal form,
                 partitioned normal form, database systems",
  owner =        "curtis",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Normal
                 forms. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Query languages.",
}

@Article{Gadia:1988:HRM,
  author =       "Shashi K. Gadia",
  title =        "A Homogeneous Relational Model and Query Languages for
                 Temporal Databases",
  journal =      j-TODS,
  volume =       "13",
  number =       "4",
  pages =        "418--448",
  month =        dec,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 198",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-4/p418-gadia/p418-gadia.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-4/p418-gadia/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/50065.html",
  abstract =     "In a temporal database, time values are associated
                 with data item to indicate their periods of validity.
                 We propose a model for temporal databases within the
                 framework of the classical database theory. Our model
                 is realized as a temporal parameterization of static
                 relations. We do not impose any restrictions upon the
                 schemes of temporal relations. The classical concepts
                 of normal forms and dependencies are easily extended to
                 our model, allowing a suitable design for a database
                 scheme. We present a relational algebra and a tuple
                 calculus for our model and prove their equivalence. Our
                 data model is homogeneous in the sense that the periods
                 of validity of all the attributes in a given tuple of a
                 temporal relation are identical. We discuss how to
                 relax the homogeneity requirement to extend the
                 application domain of our approach.",
  acknowledgement = ack-nhfb,
  affiliation =  "Iowa State Univ., Ames, IA, USA",
  affiliationaddress = "Ames, IA, USA",
  annote =       "temporal tuples and temporal relations; the temporal
                 domain varies from tuple to tuple but it is constant
                 with respect to attributes.",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; mathematical techniques ---
                 algebra; query languages; relational; relational
                 algebra; relational calculus; temporal databases;
                 theory, database systems; tuple calculus",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages.",
}

@Article{Brosda:1988:URR,
  author =       "Volkert Brosda and Gottfried Vossen",
  title =        "Update and Retrieval in a Relational Database Through
                 a Universal Schema Interface",
  journal =      j-TODS,
  volume =       "13",
  number =       "4",
  pages =        "449--485",
  month =        dec,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 199",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-4/p449-brosda/p449-brosda.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-4/p449-brosda/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/49884.html",
  abstract =     "A database system that is based on the universal
                 relation (UR) model aims at freeing its users from
                 specifying access paths on both the physical and on the
                 logical levels. All information about the logical
                 structure of the database (i.e., its conceptual scheme)
                 is hidden from users; they need only to know the
                 attribute names, which now carry all the semantics of
                 the database.\par

                 Previous work on UR interfaces has concentrated on the
                 design and implementation of query languages that serve
                 to facilitate retrieval of data from a relational
                 database. On the other hand, updates are always handled
                 as before, which means that users still have to know
                 the logical structure of the database in case they want
                 to insert, delete, or modify tuples.\par

                 In this paper the concepts underlying a UR interface,
                 which is really ``universal,'' are presented; it is
                 based on the UR model, and it permits not only queries
                 but also updates: Combinations of attributes that may
                 participate in an update-operation (``objects'') have
                 to be specified during the design phase of the
                 database, and are then embodied into the database
                 scheme by an extended synthesis algorithm. They form
                 the basis for any insertion or deletion operation. A
                 precise definition of ``insertable'' tuples, and of the
                 insert- and delete-operation in this new context, is
                 given. It is then shown that these operations modify a
                 database state in such a way that a representative
                 instance always exists. This is accomplished by
                 providing a more detailed version of Sagiv's uniqueness
                 condition and by exploring the structure of nonunique
                 objects. Since the underlying database always has a
                 representative instance, this instance can be used to
                 define the window function for retrieval. It is shown
                 that it is still possible to compute windows by a union
                 of minimal extension joins.",
  acknowledgement = ack-nhfb,
  affiliation =  "Rheinisch-Westfalische Tech. Hochschule Aachen, West
                 Germany",
  affiliationaddress = "Aachen, West Ger",
  classification = "723; 903",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer interfaces; database systems; database
                 updates; information science --- information retrieval;
                 languages; minimal extension joins; relational; theory;
                 universal relation interface; universal scheme
                 interfaces, design",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf D.2.2}: Software,
                 SOFTWARE ENGINEERING, Tools and Techniques, User
                 interfaces. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data manipulation languages
                 (DML). {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Transaction processing. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing.",
}

@Article{Gottlob:1988:PUS,
  author =       "Georg Gottlob and Paolo Paolini and Roberto Zicari",
  title =        "Properties and Update Semantics of Consistent Views",
  journal =      j-TODS,
  volume =       "13",
  number =       "4",
  pages =        "486--524",
  month =        dec,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 200",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-4/p486-gottlob/p486-gottlob.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-4/p486-gottlob/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/50068.html",
  abstract =     "The problem of translating view updates to database
                 updates is considered. Both databases and views are
                 modeled as data abstractions. A data abstraction
                 consists of a set of states and of a set of primitive
                 update operators representing state transition
                 functions. It is shown how complex update programs can
                 be built from primitive update operators and how view
                 update programs are translated into database update
                 programs. Special attention is paid to a class of views
                 that we call ``consistent.'' Loosely speaking, a
                 consistent view is a view with the following property:
                 If the effect of a view update program on a view state
                 is determined, then the effect of the corresponding
                 database update is unambiguously determined. Thus, in
                 order to know how to translate a given view update into
                 a database update, it is sufficient to be aware of a
                 functional specification of such a program. We show
                 that consistent views have a number of interesting
                 properties with respect to the concurrency of
                 (high-level) update transactions. Moreover we show that
                 the class of consistent views includes as a subset the
                 class of views that translate updates under maintenance
                 of a constant complement. However, we show that there
                 exist consistent views that do not translate under
                 constant complement. The results of Bancilhon and
                 Spyratos [6] are generalized in order to capture the
                 update semantics of the entire class of consistent
                 views. In particular we show that the class of
                 consistent views is obtained if we relax the
                 requirement of a constant complement by allowing the
                 complement to decrease according to a suitable partial
                 order.",
  acknowledgement = ack-nhfb,
  affiliation =  "Inst. for Appl. Math., CNR, Genoa, Italy",
  affiliationaddress = "Genoa, Italy",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency, design; consistent views; data
                 abstractions; database systems; theory; update
                 semantics; view updates",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems.",
}

@Article{Lomet:1988:SBD,
  author =       "David B. Lomet",
  title =        "A Simple Bounded Disorder File Organization with Good
                 Performance",
  journal =      j-TODS,
  volume =       "13",
  number =       "4",
  pages =        "525--551",
  month =        dec,
  year =         "1988",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1988-13-4/p525-lomet/p525-lomet.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1988-13-4/p525-lomet/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/50067.html",
  abstract =     "A bounded-disorder (BD) file is one in which data are
                 organized into nodes that are indexed, e.g., by means
                 of a B-tree. The data nodes are multibucket nodes that
                 are accessed by hashing. In this paper we present two
                 important improvements to the BD organization as
                 originally described. First, records in a data node
                 that overflow their designated primary bucket are
                 stored in a single overflow bucket which is itself a
                 bucket of the data node. Second, when file space needs
                 to be increased, partial expansions are used that
                 employ elastic buckets. Analysis and simulation results
                 demonstrate that this variant of the BD organization
                 has utilization, random access performance, and file
                 growth performance that can be competitive with good
                 extendible hashing methods, while supporting
                 high-performance sequential access. The simplicity of
                 the organization results in simple algorithms for
                 realizing the organization.",
  acknowledgement = ack-nhfb,
  affiliation =  "Digital Equip. Corp., Nashua, NH, USA",
  annote =       "Index entries refer to large data nodes which are
                 treated as separate hashed files.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bounded disorder file organization; computer
                 programming --- algorithms; computer simulation; data
                 processing; design; dynamic files; file organization;
                 index sequential access; indexed files; performance;
                 storage management, algorithms",
  review =       "ACM CR 8904-0253",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.3.2}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Storage, File organization. {\bf D.4.3}:
                 Software, OPERATING SYSTEMS, File Systems Management,
                 Access methods. {\bf D.4.3}: Software, OPERATING
                 SYSTEMS, File Systems Management, File organization.",
}

@Article{Freytag:1989:TRQ,
  author =       "Johann Christoph Freytag and Nathan Goodman",
  title =        "On the Translation of Relational Queries into
                 Iterative Programs",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "1--27",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p1-freytag/p1-freytag.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p1-freytag/",
  abstract =     "This paper investigates the problem of translating
                 set-oriented query specifications into iterative
                 programs. The translation uses techniques of functional
                 programming and program transformation.\par

                 We present two algorithms that generate iterative
                 programs from algebra-based query specifications. The
                 first algorithm translates query specifications into
                 recursive programs. Those are simplified by sets of
                 transformation rules before the algorithm generates the
                 final iterative form. The second algorithm uses a
                 two-level translation that generates iterative programs
                 faster than the first algorithm. On the first level a
                 small set of transformation rules performs structural
                 simplification before the functional combination on the
                 second level yields the final iterative form.",
  acknowledgement = ack-nhfb,
  affiliation =  "Eur. Comput. Ind. Res. Centre, Munich, West Germany",
  affiliationaddress = "Cambridge, MA, USA",
  annote =       "Functional programming is used for the formulation of
                 simple algebraic rules",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algebraic specification; algorithms; computer
                 programming --- algorithms; functional programming;
                 iterative programs; languages; mathematical techniques
                 --- algebra; performance; program transformation;
                 relational; relational queries; theory, database
                 systems",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf D.1.1}: Software,
                 PROGRAMMING TECHNIQUES, Applicative (Functional)
                 Programming. {\bf I.2.2}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Automatic Programming, Program
                 transformation. {\bf H.2.5}: Information Systems,
                 DATABASE MANAGEMENT, Heterogeneous Databases, Program
                 translation.",
}

@Article{Ahad:1989:ECP,
  author =       "Rafiul Ahad and K. V. {Bapa Rao} and Dennis McLeod",
  title =        "On Estimating the Cardinality of the Projection of a
                 Database Relation",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "28--40",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p28-ahad/p28-ahad.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p28-ahad/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/62034.html",
  abstract =     "We present an analytical formula for estimating the
                 cardinality of the projection on certain attributes of
                 a subset of a relation in a relational database. This
                 formula takes into account a priori knowledge of the
                 semantics of the real-world objects and relationships
                 that the database is intended to represent.
                 Experimental testing of the formula shows that it has
                 an acceptably low percentage error, and that its
                 worst-case error is smaller than the best-known
                 formula. Furthermore, the formula presented here has
                 the advantage that it does not require a scan of the
                 relation.",
  acknowledgement = ack-nhfb,
  affiliation =  "Maryland Univ., College Park, MD, USA",
  affiliationaddress = "College Park, MD, USA",
  annote =       "Uses normal distribution estimates.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Experimentation; Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "application semantics; cardinality estimation, data
                 semantics selectivity estimation TODS, design; computer
                 simulation; data semantics; database systems;
                 experimentation; measurement; performance; relational",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing.",
}

@Article{Ramesh:1989:VDT,
  author =       "R. Ramesh and A. J. G. Babu and J. Peter Kincaid",
  title =        "Variable-Depth Trie Index Optimization: Theory and
                 Experimental Results",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "41--74",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p41-ramesh/p41-ramesh.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p41-ramesh/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/77249.html",
  abstract =     "We develop an efficient approach to Trie index
                 optimization. A {\em Trie\/} is a data structure used
                 to index a file having a set of attributes as record
                 identifiers. In the proposed methodology, a file is
                 horizontally partitioned into subsets of records using
                 a Trie index whose depth of indexing is allowed to
                 vary. The retrieval of a record from the file proceeds
                 by ``stepping through'' the index to identify a subset
                 of records in the file in which a binary search is
                 performed. This paper develops a taxonomy of
                 optimization problems underlying variable-depth Trie
                 index construction. All these problems are solvable in
                 polynomial time, and their characteristics are studied.
                 Exact algorithms and heuristics for their solution are
                 presented. The algorithms are employed in CRES-an
                 expert system for editing written narrative material,
                 developed for the Department of the Navy. CRES uses
                 several large-to-very-large dictionary files for which
                 Trie indexes are constructed using these algorithms.
                 Computational experience with CRES shows that search
                 and retrieval using variable-depth Trie indexes can be
                 as much as six times faster than pure binary search.
                 The space requirements of the Tries are reasonable. The
                 results show that the variable-depth Tries constructed
                 according to the proposed algorithms are viable and
                 efficient for indexing large-to-very-large files by
                 attributes in practical applications.",
  acknowledgement = ack-nhfb,
  affiliation =  "State Univ. of New York, Buffalo, NY, USA",
  affiliationaddress = "Buffalo, NY, USA",
  classification = "723; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance;
                 Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "artificial intelligence --- expert systems; computer
                 editing; computer programming --- algorithms; CRES
                 expert system; data processing; data structures;
                 design; experimentation; information science ---
                 information retrieval; optimization; performance;
                 theory; trie index optimization, algorithms",
  subject =      "{\bf E.1}: Data, DATA STRUCTURES, Trees. {\bf E.5}:
                 Data, FILES. {\bf E.2}: Data, DATA STORAGE
                 REPRESENTATIONS, Contiguous representations. {\bf
                 H.3.1}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Content Analysis and Indexing. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval, Search process. {\bf
                 E.2}: Data, DATA STORAGE REPRESENTATIONS, Linked
                 representations.",
}

@Article{Gladney:1989:DRD,
  author =       "H. M. Gladney",
  title =        "Data Replicas in Distributed Information Services",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "75--97",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; Distributed/CCR.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "{\it Reviews}: Computing Reviews, Vol. 30, No. 11,
                 November 1989",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p75-gladney/p75-gladney.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p75-gladney/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/62035.html",
  abstract =     "In an information distribution network in which
                 records are repeatedly read, it is cost-effective to
                 keep read-only copies in work locations. This paper
                 presents a method of updating replicas that need not be
                 immediately synchronized with the source data or with
                 each other. The method allows an arbitrary mapping from
                 source records to replica records. It is fail-safe,
                 maximizes workstation autonomy, and is well suited to a
                 network with slow, unreliable, and/or expensive
                 communications links.\par

                 The algorithm is a manipulation of queries, which are
                 represented as short encodings. When a response is
                 generated, we record which portion of the source
                 database was used. Later, when the source data are
                 updated, this information is used to identify obsolete
                 replicas. For each workstation, the identity of
                 obsolete replicas is saved until a workstation process
                 asks for this information. This workstation process
                 deletes each obsolete replica, and replaces it by an
                 up-to-date version either promptly or the next time the
                 application asks for this particular item. Throughout,
                 queries are grouped so that the impact of each source
                 update transaction takes effect atomically at each
                 workstation.\par

                 Optimizations of the basic algorithm are outlined.
                 These overlap change dissemination with user service,
                 allow the mechanism to be hidden within the data
                 delivery subsystem, and permit very large networks.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Almaden Res. Center, San Jose, CA, USA",
  affiliationaddress = "San Jose, CA, USA",
  annote =       "Server and workstations.",
  classification = "723; 903; 921",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Languages; Performance;
                 Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; design;
                 distributed; information services; languages;
                 optimization, algorithms; performance; reliability;
                 reviews, database systems",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed applications. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems.",
}

@Article{McLeish:1989:FRS,
  author =       "Mary McLeish",
  title =        "Further Results on the Security of Partitioned Dynamic
                 Statistical Databases",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "98--113",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p98-mcleish/p98-mcleish.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p98-mcleish/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/62036.html",
  abstract =     "Partitioning is a highly secure approach to protecting
                 statistical databases. When updates are introduced,
                 security depends on putting restrictions on the sizes
                 of partition sets which may be queried. To overcome
                 this problem, attempts have been made to add ``dummy''
                 records. Recent work has shown that this leads to high
                 information loss.\par

                 This paper reconsiders the restrictions on the size of
                 partitioning sets required to achieve a high level of
                 security. Updates of two records at a time were studied
                 earlier, and security was found to hold if the sizes of
                 the partition sets were kept even. In this paper an
                 extended model is presented, allowing very general
                 updates to be performed. The security problem is
                 thoroughly studied, giving if and only if conditions.
                 The earlier result is shown to be part of a corollary
                 to the main theorem of this paper. Alternatives to
                 adding dummy records are presented and the practical
                 implications of the theory for the database manager are
                 discussed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Guelph Univ., Ont., Canada",
  affiliationaddress = "Guelph, Ont, Can",
  classification = "723; 922",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Security; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data processing --- security of data; database
                 security, algorithms; database systems; partitioned
                 databases; performance; security; statistical
                 databases; statistical methods; theory",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 G.3}: Mathematics of Computing, PROBABILITY AND
                 STATISTICS. {\bf K.4.1}: Computing Milieux, COMPUTERS
                 AND SOCIETY, Public Policy Issues, Privacy.",
}

@Article{Bic:1989:ADD,
  author =       "Lubomir Bic and Robert L. Hartmann",
  title =        "{AGM}: a Dataflow Database Machine",
  journal =      j-TODS,
  volume =       "14",
  number =       "1",
  pages =        "114--146",
  month =        mar,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-1/p114-bic/p114-bic.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-1/p114-bic/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/62037.html",
  abstract =     "In recent years, a number of database machines
                 consisting of large numbers of parallel processing
                 elements have been proposed. Unfortunately, there are
                 two main limitations in database processing that
                 prevent a high degree of parallelism; these are the
                 available I/O bandwidth of the underlying storage
                 devices and the concurrency control mechanisms
                 necessary to guarantee data integrity. The main problem
                 with conventional approaches is the lack of a
                 computational model capable of utilizing the potential
                 of any significant number of processing elements and
                 storage devices and, at the same time, preserving the
                 integrity of the database.\par

                 This paper presents a database model and its associated
                 architecture, which is based on the principles of
                 data-driven computation. According to this model, the
                 database is represented as a network in which each node
                 is conceptually an independent, asynchronous processing
                 element, capable of communicating with other nodes by
                 exchanging messages along the network arcs. To answer a
                 query, one or more such messages, called tokens, are
                 created and injected into the network. These then
                 propagate asynchronously through the network in search
                 of results satisfying the given query.\par

                 The asynchronous nature of processing permits the model
                 to be mapped onto a computer architecture consisting of
                 large numbers of independent disk units and processing
                 elements. This increases both the available I/O
                 bandwidth as well as the processing potential of the
                 machine. At the same time, new concurrency control and
                 error recovery mechanisms are necessary to cope with
                 the resulting parallelism.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Irvine, CA, USA",
  affiliationaddress = "Irvine, CA, USA",
  classification = "722; 723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Active Graph Machine, design; AGM Model; Computer
                 Architecture; Database Machines; Database Systems;
                 Dataflow Computing; hardware support token flow on
                 multi-processor data connection graph TODS; languages;
                 Models; performance",
  subject =      "{\bf H.2.6}: Information Systems, DATABASE MANAGEMENT,
                 Database Machines. {\bf C.1.2}: Computer Systems
                 Organization, PROCESSOR ARCHITECTURES, Multiple Data
                 Stream Architectures (Multiprocessors), Parallel
                 processors. {\bf C.1.3}: Computer Systems Organization,
                 PROCESSOR ARCHITECTURES, Other Architecture Styles,
                 Data-flow architectures. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf C.4}: Computer Systems Organization,
                 PERFORMANCE OF SYSTEMS, Modeling techniques.",
}

@Article{Yu:1989:FER,
  author =       "C. T. Yu and W. Meng and S. Park",
  title =        "A Framework for Effective Retrieval",
  journal =      j-TODS,
  volume =       "14",
  number =       "2",
  pages =        "147--167",
  month =        jun,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "1 072 201",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-2/p147-yu/p147-yu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-2/p147-yu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/63519.html",
  abstract =     "The aim of an effective retrieval system is to yield
                 high recall and precision (retrieval effectiveness).
                 The nonbinary independence model, which takes into
                 consideration the number of occurrences of terms in
                 documents, is introduced. It is shown to be optimal
                 under the assumption that terms are independent. It is
                 verified by experiments to yield significant
                 improvement over the binary independence model. The
                 nonbinary model is extended to normalized vectors and
                 is applicable to more general queries.\par

                 Various ways to alleviate the consequences of the term
                 independence assumption are discussed. Estimation of
                 parameters required for the nonbinary independence
                 model is provided, taking into consideration that a
                 term may have different meanings.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "8912-0901",
  affiliation =  "Dept. of Electr. Eng. and Comput. Sci., Illinois
                 Univ., Chicago, IL, USA",
  affiliationaddress = "Chicago, IL, USA",
  annote =       "probabilistic models with parameters estimated from
                 previously retrieved relevant and irrelevant
                 documents.",
  classification = "723; 903; 921; 922",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Database Systems; Information Retrieval; Information
                 Science; Mathematical Techniques --- Estimation;
                 Nonbinary Independence Model; Parameter Estimation;
                 Retrieval Effectiveness",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Retrieval models. {\bf H.3.3}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Search
                 and Retrieval, Query formulation. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval, Clustering.",
}

@Article{Embley:1989:NNF,
  author =       "David W. Embley",
  title =        "{NFQL}: The {Natural Forms Query Language}",
  journal =      j-TODS,
  volume =       "14",
  number =       "2",
  pages =        "168--211",
  month =        jun,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Brigham Young Un., TR-CS-87-6,
                 Mar. 1987.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-2/p168-embley/p168-embley.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-2/p168-embley/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/64125.html",
  abstract =     "A means by which ordinary forms can be exploited to
                 provide a basis for nonprocedural specification of
                 information processing is discussed. The Natural Forms
                 Query Language (NFQL) is defined. In NFQL data
                 retrieval requests and computation specifications are
                 formulated by sketching ordinary forms to show what
                 data are desired and update operations are specified by
                 altering data on filled-in forms. The meaning of a form
                 depends on a store of knowledge that includes extended
                 abstract data types for defining elementary data items,
                 a database scheme defined by an entity-relationship
                 model, and a conceptual model of an ordinary form.
                 Based on this store of knowledge, several issues are
                 addressed and resolved in the context of NFQL. These
                 issues include automatic generation of query
                 expressions from weak specifications, the view update
                 problem, power and completeness, and a heuristic
                 approach to resolving computational relationships. A
                 brief status report of an implementation of NFQL is
                 also given.",
  acknowledgement = ack-nhfb,
  affiliation =  "Brigham Young Univ., Prov., UT, USA",
  affiliationaddress = "Provo, UT, USA",
  annote =       "Uses surrogate keys in the ER model.",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database systems; entity-relationship model; forms
                 oriented interfaces; informal software specification;
                 information science --- information retrieval;
                 languages; natural forms query language; nonprocedural
                 specification; query languages; relational database,
                 algorithms; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, NFQL. {\bf H.2.1}: Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Data models. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.4.1}:
                 Information Systems, INFORMATION SYSTEMS APPLICATIONS,
                 Office Automation. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Query
                 formulation. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data description languages
                 (DDL). {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data manipulation languages
                 (DML). {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Query languages.",
}

@Article{Atzeni:1989:EOS,
  author =       "Paolo Atzeni and Edward P. F. Chan",
  title =        "Efficient Optimization of Simple Chase Join
                 Expressions",
  journal =      j-TODS,
  volume =       "14",
  number =       "2",
  pages =        "212--230",
  month =        jun,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 202",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-2/p212-atzeni/p212-atzeni.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-2/p212-atzeni/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/63520.html",
  abstract =     "Simple chase join expressions are relational algebra
                 expressions, involving only projection and join
                 operators, defined on the basis of the functional
                 dependencies associated with the database scheme. They
                 are meaningful in the weak instance model, because for
                 certain classes of schemes, including independent
                 schemes, the total projections of the representative
                 instance can be computed by means of unions of simple
                 chase join expressions. We show how unions of simple
                 chase join expressions can be optimized efficiently,
                 without constructing and chasing the corresponding
                 tableaux. We also present efficient algorithms for
                 testing containment and equivalence, and for optimizing
                 individual simple chase join expressions.",
  acknowledgement = ack-nhfb,
  affiliation =  "IASI-CNR, Rome, Italy",
  affiliationaddress = "Rome, Italy",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer programming --- algorithms; containment
                 testing; database systems; equivalence testing,
                 algorithms; functional dependencies; languages;
                 optimization; performance; relational; relational
                 algebra expressions; simple chase join expressions;
                 theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Normal forms.",
}

@Article{Ramakrishna:1989:FOU,
  author =       "M. V. Ramakrishna and Per-{\AA}ke Larson",
  title =        "File Organization Using Composite Perfect Hashing",
  journal =      j-TODS,
  volume =       "14",
  number =       "2",
  pages =        "231--263",
  day =          "1",
  month =        jun,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib; UnCover
                 library database",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-2/p231-ramakrishna/p231-ramakrishna.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-2/p231-ramakrishna/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/63521.html",
  abstract =     "Perfect hashing refers to hashing with no overflows.
                 We propose and analyze a composite perfect hashing
                 scheme for large external files. The scheme guarantees
                 retrieval of any record in a single disk access.
                 Insertions and deletions are simple, and the file size
                 may vary considerably without adversely affecting the
                 performance. A simple variant of the scheme supports
                 efficient range searches in addition to being a
                 completely dynamic file organization scheme. These
                 advantages are achieved at the cost of a small amount
                 of additional internal storage and increased cost of
                 insertions.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Michigan State Univ., East
                 Lansing, MI, USA",
  affiliationaddress = "East Lansing, MI, USA",
  classification = "723; 903",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "composite perfect hashing; data processing; database
                 systems; design; dynamic file organization, algorithms;
                 experimentation; file organization; information science
                 --- information retrieval; large external files;
                 performance; range searches",
  subject =      "{\bf E.2}: Data, DATA STORAGE REPRESENTATIONS,
                 Hash-table representations. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Sorting and searching. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods.",
}

@Article{ElAbaddi:1989:MAP,
  author =       "Amr {El Abaddi} and Sam Toueg",
  title =        "Maintaining Availability in Partitioned Replicated
                 Databases",
  journal =      j-TODS,
  volume =       "14",
  number =       "2",
  pages =        "264--290",
  month =        jun,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 072 203",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Compendex database; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: ACM SIGACT-SIGMOD Symp. on
                 Principles of Database Systems, Cambridge MA, Mar.
                 1986.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-2/p264-abbadi/p264-abbadi.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-2/p264-abbadi/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/63501.html",
  abstract =     "In a replicated database, a data item may have copies
                 residing on several sites. A replica control protocol
                 is necessary to ensure that data items with several
                 copies behave as if they consist of a single copy, as
                 far as users can tell. We describe a new replica
                 control protocol that allows the accessing of data in
                 spite of site failures and network partitioning. This
                 protocol provides the database designer with a large
                 degree of flexibility in deciding the degree of data
                 availability, as well as the cost of accessing data.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Cornell Univ., Ithaca, NY,
                 USA",
  affiliationaddress = "Ithaca, NY, USA",
  classification = "723",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "computer operating systems; concurrency control;
                 database availability, algorithms; database systems;
                 design; distributed; partitioned replicated databases;
                 partitioning failures; performance; reliability;
                 replica control; serializability",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, Distributed file systems. {\bf D.4.5}:
                 Software, OPERATING SYSTEMS, Reliability,
                 Fault-tolerance. {\bf C.2.2}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS, Network
                 Protocols. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design, Access methods.",
}

@Article{Hudson:1989:CSA,
  author =       "Scott E. Hudson and Roger King",
  title =        "{Cactis}: a Self-Adaptive, Concurrent Implementation
                 of an Object-Oriented Database Management System",
  journal =      j-TODS,
  volume =       "14",
  number =       "3",
  pages =        "291--321",
  month =        sep,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-3/p291-hudson/p291-hudson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-3/p291-hudson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/68013.html",
  abstract =     "Cactis is an object-oriented, multiuser DBMS developed
                 at the University of Colorado. The system supports
                 functionally-defined data and uses techniques based on
                 attributed graphs to optimize the maintenance of
                 functionally-defined data. The implementation is
                 self-adaptive in that the physical organization and the
                 update algorithms dynamically change in order to reduce
                 disk access. The system is also concurrent. At any
                 given time there are some number of computations that
                 must be performed to bring the database up to date;
                 these computations are scheduled independently and are
                 performed when the expected cost to do so is minimal.
                 The DBMS runs in the Unix/C Sun workstation
                 environment. Cactis is designed to support applications
                 that require rich data modeling capabilities and the
                 ability to specify functionally-defined data, but that
                 also demand good performance. Specifically, Cactis is
                 intended for use in the support of such applications as
                 VLSI and PCB design, and software environments.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Arizona Univ., Tucson, AZ,
                 USA",
  annote =       "attributes of objects can be derived; dynamic
                 definition of a scheme, depending on the user's
                 requirements functionally defined data, intelligent
                 update algorithms",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; clustering derived update propagation
                 TODS; design; performance",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Cactis. {\bf H.2.1}: Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Data models. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design. {\bf J.6}: Computer Applications,
                 COMPUTER-AIDED ENGINEERING.",
}

@Article{Sheard:1989:AVD,
  author =       "Tim Sheard and David Stemple",
  title =        "Automatic Verification of Database Transaction
                 Safety",
  journal =      j-TODS,
  volume =       "14",
  number =       "3",
  pages =        "322--368",
  month =        sep,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: U. Mass, COINS TR-88-29, Apr.
                 1988, also TR-86-30, 1986.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-3/p322-sheard/p322-sheard.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-3/p322-sheard/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/68014.html",
  abstract =     "Maintaining the integrity of databases is one of the
                 promises of database management systems. This includes
                 assuring that integrity constraints are invariants of
                 database transactions. This is very difficult to
                 accomplish efficiently in the presence of complex
                 constraints and large amounts of data. One way to
                 minimize the amount of processing required to maintain
                 database integrity over transaction processing is to
                 prove at compile-time that transactions cannot, if run
                 atomically, disobey integrity constraints. We report on
                 a system that performs such verification for a robust
                 set of constraint and transaction classes. The system
                 accepts database schemas written in a more or less
                 traditional style and accepts programs in a high-level
                 programming language. Automatic verification fast
                 enough to be effective on current workstation hardware
                 is performed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. and Inf. Sci., Massachusetts Univ.,
                 Amherst, MA, USA",
  annote =       "LISP; inference techniques based on Boyer--Moore.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Reliability; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; reliability; verification",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Schema and subschema. {\bf D.2.4}: Software,
                 SOFTWARE ENGINEERING, Program Verification.",
}

@Article{Blakeley:1989:UDR,
  author =       "Jos{\'e} A. Blakeley and Neil Coburn and Per-{\AA}ke
                 Larson",
  title =        "Updating Derived Relations: Detecting Irrelevant and
                 Autonomously Computable Updates",
  journal =      j-TODS,
  volume =       "14",
  number =       "3",
  pages =        "369--400",
  month =        sep,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 073 202",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Indiana Un., TR-235, Nov. 1987.
                 Also published in \cite{Kambayashi:1986:TIC}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-3/p369-blakeley/p369-blakeley.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-3/p369-blakeley/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/68015.html",
  abstract =     "Consider a database containing not only base relations
                 but also stored derived relations (also called
                 materialized or concrete views). When a base relation
                 is updated, it may also be necessary to update some of
                 the derived relations. This paper gives sufficient and
                 necessary conditions for detecting when an update of a
                 base relation cannot affect a derived relation (an
                 irrelevant update), and for detecting when a derived
                 relation can be correctly updated using no data other
                 than the derived relation itself and the given update
                 operation (an autonomously computable update). The
                 class of derived relations considered is restricted to
                 those defined by {\em PSJ\/}-expressions, that is, any
                 relational algebra expressions constructed from an
                 arbitrary number of project, select and join operations
                 (but containing no self-joins). The class of update
                 operations consists of insertions, deletions, and
                 modifications, where the set of tuples to be deleted or
                 modified is specified by a selection condition on
                 attributes of the relation being updated.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Waterloo Univ., Ont., Canada",
  annote =       "Identity implementation; rigorous mathematical proofs;
                 expressions that are capable of being tested constitute
                 a large and commonly occurring class.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Schema and subschema. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods.",
}

@Article{Mackert:1989:ISU,
  author =       "Lothar F. Mackert and Guy M. Lohman",
  title =        "Index Scans Using a Finite {LRU} Buffer: a Validated
                 {I/O} Model",
  journal =      j-TODS,
  volume =       "14",
  number =       "3",
  pages =        "401--424",
  month =        sep,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: IBM, TR-RC-4836, Sep. 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-3/p401-mackert/p401-mackert.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-3/p401-mackert/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/68016.html",
  abstract =     "Indexes are commonly employed to retrieve a portion of
                 a file or to retrieve its records in a particular
                 order. An accurate performance model of indexes is
                 essential to the design, analysis, and tuning of file
                 management and database systems, and particularly to
                 database query optimization. Many previous studies have
                 addressed the problem of estimating the number of disk
                 page fetches when randomly accessing $k$ records out of
                 $N$ given records stored on $T$ disk pages. This paper
                 generalizes these results, relaxing two assumptions
                 that usually do not hold in practice: unlimited buffer
                 and unique records for each key value. Experiments show
                 that the performance of an index scan is very sensitive
                 to buffer size limitations and multiple records per key
                 value. A model for these more practical situations is
                 presented and a formula derived for estimating the
                 performance of an index scan. We also give a
                 closed-form approximation that is easy to compute. The
                 theoretical results are validated using the $R$ *
                 distributed relational database system. Although we use
                 database terminology throughout the paper, the model is
                 more generally applicable whenever random accesses are
                 made using keys.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Almaden Res. Center, San Jose, CA, USA",
  annote =       "consider limited buffer and now unique records for
                 each key value.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Measurement; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; measurement; performance; theory, clustering
                 non-clustering B-tree lookup cost Starburst System R
                 IBM Almaden TODS",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.3.2}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Storage, File organization. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.3.3}: Information Systems, INFORMATION
                 STORAGE AND RETRIEVAL, Information Search and
                 Retrieval, Retrieval models. {\bf H.3.1}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Content
                 Analysis and Indexing, Indexing methods. {\bf E.5}:
                 Data, FILES, Organization/structure. {\bf C.4}:
                 Computer Systems Organization, PERFORMANCE OF SYSTEMS,
                 Performance attributes.",
}

@Article{Manolopoulos:1989:PTH,
  author =       "Y. Manolopoulos and J. G. Kollias",
  title =        "Performance of a Two-Headed Disk System when Serving
                 Database Queries Under the Scan Policy",
  journal =      j-TODS,
  volume =       "14",
  number =       "3",
  pages =        "425--442",
  month =        sep,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-3/p425-manolopoulos/p425-manolopoulos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-3/p425-manolopoulos/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/68017.html",
  abstract =     "Disk drives with movable two-headed arms are now
                 commercially available. The two heads are separated by
                 a fixed number of cylinders. A major problem for
                 optimizing disk head movement, when answering database
                 requests, is the specification of the optimum number of
                 cylinders separating the two heads. An earlier
                 analytical study assumed a FCFS model and concluded
                 that the optimum separation distance should be equal to
                 0.44657 of the number of cylinders $N$ of the disk.
                 This paper considers that the SCAN scheduling policy is
                 used in file access, and it applies combinatorial
                 analysis to derive exact formulas for the expected head
                 movement. Furthermore, it is proven that the optimum
                 separation distance is $ N / 2 - 1 $ ($ \lceil N / 2 -
                 1 \rceil $ and $ \lfloor N / 2 - 1 \rfloor $ ) if $N$
                 is even (odd). In addition, a comparison with a
                 single-headed disk system operating under the same
                 scheduling policy shows that if the two heads are
                 optimally spaced, then the mean seek distance is less
                 than one-half of the value obtained with one head. In
                 fact that the SCAN policy is used for many database
                 applications (for example,batching and secondary key
                 retrieval) demonstrates the potential of two-headed
                 disk systems for improving the performance of database
                 systems.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Electr. Eng., Aristotelian Univ. of
                 Thessaloniki, Greece",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; hardware support I/O seek machine TODS,
                 algorithms; performance",
  subject =      "{\bf D.4.2}: Software, OPERATING SYSTEMS, Storage
                 Management, Secondary storage. {\bf D.4.1}: Software,
                 OPERATING SYSTEMS, Process Management, Scheduling. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Sequencing and scheduling.",
}

@Article{Dreizen:1989:ISR,
  author =       "Howard M. Dreizen and Shi-Kuo Chang",
  title =        "Imprecise Schema: a Rationale for Relations with
                 Embedded Subrelations",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "447--479",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Misc/is.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p447-dreizen/p447-dreizen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p447-dreizen/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76903.html",
  abstract =     "{\em Exceptional conditions\/} are anomalous data
                 which meet the intent of a schema but not the schema
                 definition, represent a small proportion of the
                 database extension, and may become known only after the
                 schema is in use. Admission of exceptional conditions
                 is argued to suggest a representation that locally
                 stretches the schema definition by use of relations
                 with embedded subrelations. Attempted normalization of
                 these relations to 1NF does not yield the static schema
                 typically associated with such transformations. A class
                 of relations, termed Exceptional Condition Nested Form
                 (ECNF), is defined which allows the necessary
                 representation of exceptional conditions while
                 containing sufficient restrictions to prevent arbitrary
                 and chaotic inclusion of embedded subrelations. Queries
                 on a subset of exceptional conditions, the {\em
                 exceptional constraints}, are provided an
                 interpretation via an algorithm that transforms ECNF
                 relations into 1NF relations containing two types of
                 null values. Extensions of relational algebraic
                 operators, suitable for interactive query navigation,
                 are defined for use with ECNF relations containing all
                 forms of exceptional conditions.",
  acknowledgement = ack-nhfb,
  affiliation =  "Illinois Inst. of Technol., Chicago, IL, USA",
  annote =       "If the data contain only a few `exceptional' tuples,
                 their effect might better be confined to a local schema
                 change rather than changing the global schema; the
                 effect of processing is only partly answered.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing.",
}

@Article{Motro:1989:IVC,
  author =       "Amihai Motro",
  title =        "Integrity = Validity + Completeness",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "480--502",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p480-motro/p480-motro.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p480-motro/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76904.html",
  abstract =     "Database integrity has two complementary components:
                 {\em validity}, which guarantees that all false
                 information is excluded from the database, and {\em
                 completeness}, which guarantees that all true
                 information is included in the database. This article
                 describes a uniform model of integrity for relational
                 databases, that considers both validity and
                 completeness. To a large degree, this model subsumes
                 the prevailing model of integrity (i.e., integrity
                 constraints). One of the features of the new model is
                 the determination of the integrity of answers issued by
                 the database system in response to user queries. To
                 users, answers that are accompanied with such detailed
                 certifications of their integrity are more meaningful.
                 First, the model is defined and discussed. Then, a
                 specific mechanism is described that implements this
                 model. With this mechanism, the determination of the
                 integrity of an answer is a process analogous to the
                 determination of the answer itself.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ. of Southern California, Los Angeles, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models.",
}

@Article{Farrag:1989:USK,
  author =       "Abdel Aziz Farrag and M. Tamer {\"O}zsu",
  title =        "Using Semantic Knowledge of Transactions to Increase
                 Concurrency",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "503--525",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Accepted. Also published in/as: Un. Alberta, DCS,
                 TR-85-11, Jul. 1985.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p503-farrag/p503-farrag.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p503-farrag/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76905.html",
  abstract =     "When the only information available about transactions
                 is syntactic information, serializability is the main
                 correctness criterion for concurrency control.
                 Serializability requires that the execution of each
                 transaction must appear to every other transaction as a
                 single atomic step (i.e., the execution of the
                 transaction cannot be interrupted by other
                 transactions). Many researchers, however, have realized
                 that this requirement is unnecessarily strong for many
                 applications and can significantly increase transaction
                 response time. To overcome this problem, a new approach
                 for controlling concurrency that exploits the semantic
                 information available about transactions to allow
                 controlled nonserializable interleavings has recently
                 been proposed. This approach is useful when the cost of
                 producing only serializable interleavings is
                 unacceptably high. The main drawback of the approach is
                 the extra overhead incurred by utilizing the semantic
                 information. We examine this new approach in this paper
                 and discuss its strengths and weaknesses. We introduce
                 a new formalization for the concurrency control problem
                 when semantic information is available about the
                 transactions. This semantic information takes the form
                 of transaction types, transaction steps, and
                 transaction break-points. We define a new class of
                 ``safe'' schedules called relatively consistent (RC)
                 schedules. This class contains serializable as well as
                 nonserializable schedules. We prove that the execution
                 of an RC schedule cannot violate consistency and
                 propose a new concurrency control mechanism that
                 produces only RC schedules. Our mechanism assumes fewer
                 restrictions on the interleavings among transactions
                 than previously introduced semantic-based mechanisms.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dalhousie Univ., Halifax, NS, Canada",
  annote =       "By setting breakpoints to interleave compatible
                 transactions. But aborts can require rollbacks or
                 offsetting transactions. Compatible transactions can
                 interleave. Nested compatibility in interleaving
                 transactions.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.",
}

@Article{Ozsoyoglu:1989:QPT,
  author =       "Gultekin {\"O}zsoyo{\u{g}}lu and Victor Matos and Z.
                 Meral {\"O}zsoyo{\u{g}}lu",
  title =        "Query Processing Techniques in the
                 Summary-Table-by-Example Database Query Language",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "526--573",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p526-ozsoyoglu/p526-ozsoyoglu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p526-ozsoyoglu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76906.html",
  abstract =     "Summary-Table-by-Example (STBE) is a graphical
                 language suitable for statistical database
                 applications. STBE queries have a hierarchical subquery
                 structure and manipulate summary tables and relations
                 with set-valued attributes.\par

                 The hierarchical arrangement of STBE queries naturally
                 implies a tuple-by-tuple subquery evaluation strategy
                 (similar to the nested loops join implementation
                 technique) which may not be the best query processing
                 strategy. In this paper we discuss the query processing
                 techniques used in STBE. We first convert an STBE query
                 into an ``extended'' relational algebra (ERA)
                 expression. Two transformations are introduced to
                 remove the hierarchical arrangement of subqueries so
                 that query optimization is possible. To solve the
                 ``empty partition'' problem of aggregate function
                 evaluation, directional join (one-sided outer-join) is
                 utilized. We give the algebraic properties of the ERA
                 operators to obtain an ``improved'' ERA expression.
                 Finally we briefly discuss the generation of
                 alternative implementations of a given ERA expression.
                 \par

                 STBE is implemented in a prototype statistical database
                 management system. We discuss the STBE-related features
                 of the implemented system.",
  acknowledgement = ack-nhfb,
  affiliation =  "Case Western Reserve Univ., Cleveland, OH, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; languages; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Data models. {\bf G.3}:
                 Mathematics of Computing, PROBABILITY AND STATISTICS,
                 Statistical software.",
}

@Article{Grady:1989:EJO,
  author =       "Dani{\`e}le Grady and Claude Puech",
  title =        "On the Effect of Join Operations on Relation Sizes",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "574--603",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 073 203",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p574-grady/p574-grady.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p574-grady/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76907.html",
  abstract =     "We propose a generating function approach to the
                 problem of evaluating the sizes of derived relations in
                 a relational database framework. We present a model of
                 relations and show how to use it to deduce
                 probabilistic estimations of derived relation sizes.
                 These are found to asymptotically follow normal
                 distributions under a variety of assumptions.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ. Paris-Sud, Orsay, France",
  annote =       "Polynomial generating function is given; derived
                 relation sizes asymptotically follow normal
                 distributions.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; generating functions; performance;
                 selectivity estimation; theory; TODS",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.2.1}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Combinatorics,
                 Generating functions. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Computations on discrete structures. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Schema and subschema. {\bf H.3.3}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Search and Retrieval.",
}

@Article{Lang:1989:UAB,
  author =       "Sheau-Dong Lang and James R. Driscoll and Jiann H.
                 Jou",
  title =        "A Unified Analysis of Batched Searching of Sequential
                 and Tree-Structured Files",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "604--618",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10 (68P20)",
  MRnumber =     "1 073 204",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1989-14-4/p604-lang/p604-lang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1989-14-4/p604-lang/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/76908.html",
  abstract =     "A direct and unified approach is used to analyze the
                 efficiency of batched searching of sequential and
                 tree-structured files. The analysis is applicable to
                 arbitrary search distributions, and closed-form
                 expressions are obtained for the expected batched
                 searching cost and savings. In particular, we consider
                 a search distribution satisfying Zipf's law for
                 sequential files and four types of uniform (random)
                 search distribution for sequential and tree-structured
                 files. These results unify and extend earlier research
                 on batched searching and estimating block accesses for
                 database systems.",
  acknowledgement = ack-nhfb,
  affiliation =  "Central Florida Univ., Orlando, FL, USA",
  annote =       "closed-form expressions for the number of accesses
                 needed given arbitrary search distributions.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval, Search
                 process. {\bf H.3.2}: Information Systems, INFORMATION
                 STORAGE AND RETRIEVAL, Information Storage, File
                 organization. {\bf H.2.2}: Information Systems,
                 DATABASE MANAGEMENT, Physical Design, Access methods.",
}

@Article{Motro:1989:QDK,
  author =       "A. Motro and Q. Yuan",
  title =        "Querying Database Knowledge",
  journal =      j-TODS,
  volume =       "14",
  number =       "4",
  pages =        "??--??",
  month =        dec,
  year =         "1989",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in \cite{Garcia-Molina:1990:PAS}.",
  annote =       "The describe statement inquires about the meaning of a
                 concept under specified circumstances",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  xxnote =       "This paper does not seem to be published in TODS.",
}

@Article{Liu:1990:IMI,
  author =       "Ken-Chih C. Liu and Rajshekhar Sunderraman",
  title =        "Indefinite and Maybe Information in Relational
                 Databases",
  journal =      j-TODS,
  volume =       "15",
  number =       "1",
  pages =        "1--39",
  month =        mar,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 073 205",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-1/p1-liu/p1-liu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-1/p1-liu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/77644.html",
  abstract =     "This paper extends the relational model to represent
                 indefinite and maybe kinds of incomplete information. A
                 data structure, called an I-table, which is capable of
                 representing indefinite and maybe facts, is introduced.
                 The information content of I-tables is precisely
                 defined, and an operator to remove redundant facts is
                 presented. The relational algebra is then extended in a
                 semantically correct way to operate on I-tables.
                 Queries are posed in the same way as in conventional
                 relational algebra; however, the user may now expect
                 indefinite as well as maybe answers.",
  acknowledgement = ack-nhfb,
  affiliation =  "AT\&T Bell Labs., Naperville, IL, USA",
  annote =       "an I-table is capable of representing indefinite and
                 maybe facts, is introduced; an operator to remove
                 redundant facts is presented. The relational algebra is
                 then extended; user may now expect indefinite as well
                 as maybe answers",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf E.1}: Data, DATA STRUCTURES, Tables.
                 {\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval.",
}

@Article{Langerak:1990:VUR,
  author =       "Rom Langerak",
  title =        "View Updates in Relational Databases with an
                 Independent Scheme",
  journal =      j-TODS,
  volume =       "15",
  number =       "1",
  pages =        "40--66",
  month =        mar,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 073 206",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-1/p40-langerak/p40-langerak.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-1/p40-langerak/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/77645.html",
  abstract =     "A view on a database is a mapping that provides a user
                 or application with a suitable way of looking at the
                 data. Updates specified on a view have to be translated
                 into updates on the underlying database. We study the
                 view update translation problem for a relational data
                 model in which the base relations may contain (indexed)
                 nulls.\par

                 The representative instance is considered to be the
                 correct representation of all data in the database; the
                 class of views that is studied consists of total
                 projections of the representative instance. Only
                 independent database schemes are considered, that is,
                 schemes for which global consistency is implied by
                 local consistency. A view update can be an insertion, a
                 deletion, or a modification of a single view tuple.
                 \par

                 It is proven that the constant complement method of
                 Bancilhon and Spyratos is too restrictive to be useful
                 in this context. Structural properties of extension
                 joins are derived that are important for understanding
                 views. On the basis of these properties, minimal
                 algorithms for translating a single view-tuple update
                 are given.",
  acknowledgement = ack-nhfb,
  affiliation =  "Fac. of Inf., Twente Univ., Enschede, Netherlands",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; theory; views TODS, algorithms",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models. {\bf H.3.3}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Search
                 and Retrieval.",
}

@Article{Whang:1990:QOM,
  author =       "Kyu-Young Y. Whang and Ravi Krishnamurthy",
  title =        "Query Optimization in a Memory-Resident Domain
                 Relational Calculus Database System",
  journal =      j-TODS,
  volume =       "15",
  number =       "1",
  pages =        "67--95",
  month =        mar,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-1/p67-whang/p67-whang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-1/p67-whang/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/77646.html",
  abstract =     "We present techniques for optimizing queries in
                 memory-resident database systems. Optimization
                 techniques in memory-resident database systems differ
                 significantly from those in conventional disk-resident
                 database systems. In this paper we address the
                 following aspects of query optimization in such systems
                 and present specific solutions for them: (1) a new
                 approach to developing a CPU-intensive cost model; (2)
                 new optimization strategies for main-memory query
                 processing; (3) new insight into join algorithms and
                 access structures that take advantage of memory
                 residency of data; and (4) the effect of the operating
                 system's scheduling algorithm on the memory-residency
                 assumption. We present an interesting result that a
                 major cost of processing queries in memory-resident
                 database systems is incurred by evaluation of
                 predicates. We discuss optimization techniques using
                 the Office-by-Example (OBE) that has been under
                 development at IBM Research. We also present the
                 results of performance measurements, which prove to be
                 excellent in the current state of the art. Despite
                 recent work on memory-resident database systems, query
                 optimization aspects in these systems have not been
                 well studied. We believe this paper opens the issues of
                 query optimization in memory-resident database systems
                 and presents practical solutions to them.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Thomas J. Watson Res. Center, Yorktown Heights,
                 NY, USA",
  annote =       "Office-by-example extends the concept of
                 query-by-example (QBE); disks are used only for
                 permanent storage of data and backup; The technique is
                 not a heuristic since it employs a systematic search,
                 but uses the branch-and-bound algorithm. Uses the
                 nested-loop join with use of indexes. An index is an
                 array of tuple identifiers. Assess uses binary search.
                 When an index entry is inserted, the upper half of the
                 index is block-copied. In a 3081 processor copying 1 MB
                 of memory takes less than 0.1 second. Queries in OBE
                 are in the canonical form, have no substructures. Pure
                 demand paging is not suitable, the system has a global
                 goal for paging activities. The set of virtual machines
                 on the dispatch list is determined.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Languages; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "experimentation; languages; Large Main Memory TODS,
                 algorithms; performance",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design, Access methods.",
}

@Article{Herlihy:1990:AVA,
  author =       "Maurice Herlihy",
  title =        "Apologizing Versus Asking Permission: Optimistic
                 Concurrency Control for Abstract Data Types",
  journal =      j-TODS,
  volume =       "15",
  number =       "1",
  pages =        "96--124",
  month =        mar,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68N25 (68P15 68Q65)",
  MRnumber =     "1 073 207",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-1/p96-herlihy/p96-herlihy.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-1/p96-herlihy/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/77647.html",
  abstract =     "An optimistic concurrency control technique is one
                 that allows transactions to execute without
                 synchronization, relying on commit-time validation to
                 ensure serializability. Several new optimistic
                 concurrency control techniques for objects in
                 decentralized distributed systems are described here,
                 their correctness and optimality properties are proved,
                 and the circumstances under which each is likely to be
                 useful are characterized.\par

                 Unlike many methods that classify operations only as
                 Reads or Writes, these techniques systematically
                 exploit type-specific properties of objects to validate
                 more interleavings. Necessary and sufficient validation
                 conditions can be derived directly from an object's
                 data type specification. These techniques are also
                 modular: they can be applied selectively on a
                 per-object (or even per-operation) basis in conjunction
                 with standard pessimistic techniques such as two-phase
                 locking, permitting optimistic methods to be introduced
                 exactly where they will be most effective.\par

                 These techniques can be used to reduce the algorithmic
                 complexity of achieving high levels of concurrency,
                 since certain scheduling decisions that are NP-complete
                 for pessimistic schedulers can be validated after the
                 fact in time, independent of the level of concurrency.
                 These techniques can also enhance the availability of
                 replicated data, circumventing certain tradeoffs
                 between concurrency and availability imposed by
                 comparable pessimistic techniques.",
  acknowledgement = ack-nhfb,
  affiliation =  "Carnegie Mellon Univ., Pittsburgh, PA, USA",
  annote =       "new optimistic techniques for objects; exploit
                 type-specific properties of objects to validate
                 interleavings. These techniques reduce the complexity
                 of achieving high levels of concurrency and enhance the
                 availability of replicated data. Deals with hot spots
                 such as counters, account balances, or queues.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Reliability; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "ADT TODS, algorithms; design; reliability;
                 verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency. {\bf D.3.3}: Software, PROGRAMMING
                 LANGUAGES, Language Constructs and Features, Abstract
                 data types. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Distributed systems. {\bf D.4.5}:
                 Software, OPERATING SYSTEMS, Reliability, Verification.
                 {\bf D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, Distributed file systems.",
}

@Article{Wald:1990:EAF,
  author =       "Joseph A. Wald and Paul G. Sorenson",
  title =        "Explaining Ambiguity in a Formal Query Language",
  journal =      j-TODS,
  volume =       "15",
  number =       "2",
  pages =        "125--161",
  month =        jun,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-2/p125-wald/p125-wald.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-2/p125-wald/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/78923.html",
  abstract =     "The problem of generating reasonable natural
                 language-like responses to queries formulated in
                 nonnavigational query languages with logical data
                 independence is addressed. An extended ER model, the
                 Entity-Relationship-Involvement model, is defined which
                 assists in providing a greater degree of logical data
                 independence and the generation of natural language
                 explanations of a query processor's interpretation of a
                 query. These are accomplished with the addition of the
                 concept of an involvement to the model. Based on
                 involvement definitions in a formally defined data
                 definition language, DDL, an innovative strategy for
                 generating explanations is outlined and exemplified. In
                 the conclusion, possible extensions to the approach are
                 given.",
  acknowledgement = ack-nhfb,
  affiliation =  "Schlumberger Lab. for Comput. Sci., Austin, TX, USA",
  annote =       "sorting out paths in the ER model.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models.",
}

@Article{Chakravarthy:1990:LBA,
  author =       "Upen S. Chakravarthy and John Grant and Jack Minker",
  title =        "Logic-Based Approach to Semantic Query Optimization",
  journal =      j-TODS,
  volume =       "15",
  number =       "2",
  pages =        "162--207",
  month =        jun,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-2/p162-chakravarthy/p162-chakravarthy.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-2/p162-chakravarthy/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/78924.html",
  abstract =     "The purpose of semantic query optimization is to use
                 semantic knowledge (e.g., integrity constraints) for
                 transforming a query into a form that may be answered
                 more efficiently than the original version. In several
                 previous papers we described and proved the correctness
                 of a method for semantic query optimization in
                 deductive databases couched in first-order logic. This
                 paper consolidates the major results of these papers
                 emphasizing the techniques and their applicability for
                 optimizing relational queries. Additionally, we show
                 how this method subsumes and generalizes earlier work
                 on semantic query optimization. We also indicate how
                 semantic query optimization techniques can be extended
                 to databases that support recursion and integrity
                 constraints that contain disjunction, negation, and
                 recursion.",
  acknowledgement = ack-nhfb,
  annote =       "consolidate results emphasizing the techniques and
                 their applicability for optimizing relational queries;
                 recursion and integrity constraints that contain
                 disjunction, negation, and recursion.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf I.2.3}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem Proving,
                 Logic programming. {\bf H.2.0}: Information Systems,
                 DATABASE MANAGEMENT, General, Security, integrity, and
                 protection. {\bf I.2.8}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Problem Solving, Control
                 Methods, and Search, Heuristic methods. {\bf I.2.8}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Problem Solving, Control Methods, and Search, Plan
                 execution, formation, generation.",
}

@Article{Whang:1990:LTP,
  author =       "Kyu-Young Whang and Brad T. {Vander-Zanden} and Howard
                 M. Taylor",
  title =        "A Linear-Time Probabilistic Counting Algorithm for
                 Database Applications",
  journal =      j-TODS,
  volume =       "15",
  number =       "2",
  pages =        "208--229",
  month =        jun,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib;
                 Theory/ProbAlgs.bib",
  note =         "A probabilistic technique called linear counting,
                 based on hashing, for counting the number of unique
                 values in the presence of duplicates is presented in
                 this paper.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-2/p208-whang/p208-whang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-2/p208-whang/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/78925.html",
  abstract =     "We present a probabilistic algorithm for counting the
                 number of unique values in the presence of duplicates.
                 This algorithm has $O$ ($q$ ) time complexity, where
                 $q$ is the number of values including duplicates, and
                 produces an estimation with an arbitrary accuracy
                 prespecified by the user using only a small amount of
                 space. Traditionally, accurate counts of unique values
                 were obtained by sorting, which has $O$ ($q$ log $q$ )
                 time complexity. Our technique, called {\em linear
                 counting}, is based on hashing. We present a
                 comprehensive theoretical and experimental analysis of
                 linear counting. The analysis reveals an interesting
                 result: A load factor (number of unique values/hash
                 table size) much larger than 1.0 (e.g., 12) can be used
                 for accurate estimation (e.g., 1\% of error). We present
                 this technique with two important applications to
                 database problems: namely, (1) obtaining the column
                 cardinality (the number of unique values in a column of
                 a relation) and (2) obtaining the join selectivity (the
                 number of unique values in the join column resulting
                 from an unconditional join divided by the number of
                 unique join column values in the relation to he
                 joined). These two parameters are important statistics
                 that are used in relational query optimization and
                 physical database design.",
  acknowledgement = ack-nhfb,
  affiliation =  "Korea Adv. Inst. of Sci. and Technol., Seoul, South
                 Korea",
  annote =       "Counting the number of unique values in the presence
                 of duplicates; $ O(n) $ time complexity based on
                 hashing.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance;
                 Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; experimentation; hashing sampling TODS,
                 algorithms; performance; theory",
  subject =      "{\bf G.3}: Mathematics of Computing, PROBABILITY AND
                 STATISTICS, Probabilistic algorithms (including Monte
                 Carlo). {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Access methods.",
}

@Article{Jajodia:1990:DVA,
  author =       "Sushil Jajodia and David Mutchler",
  title =        "Dynamic Voting Algorithms for Maintaining the
                 Consistency of a Replicated Database",
  journal =      j-TODS,
  volume =       "15",
  number =       "2",
  pages =        "230--280",
  month =        jun,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-2/p230-jajodia/p230-jajodia.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-2/p230-jajodia/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/78926.html",
  abstract =     "There are several replica control algorithms for
                 managing replicated files in the face of network
                 partitioning due to site or communication link
                 failures. Pessimistic algorithms ensure consistency at
                 the price of reduced availability; they permit at most
                 one (distinguished) partition to process updates at any
                 given time. The best known pessimistic algorithm, {\em
                 voting}, is a ``static'' algorithm, meaning that all
                 potential distinguished partitions can be listed in
                 advance. We present a dynamic extension of voting
                 called {\em dynamic voting}. This algorithm permits
                 updates in a partition provided it contains more than
                 half of the {\em up-to-date\/} copies of the replicated
                 file. We also present an extension of dynamic voting
                 called {\em dynamic voting with linearly ordered
                 copies\/} (abbreviated as {\em dynamic-linear\/}).
                 These algorithms are dynamic because the order in which
                 past distinguished partitions were created plays a role
                 in the selection of the next distinguished partition.
                 Our algorithms have all the virtues of ordinary voting,
                 including its simplicity, and provide improved
                 availability as well. We provide two stochastic models
                 to support the latter claim. In the first (site) model,
                 sites may fail but communication links are infallible;
                 in the second (link) model the reverse is true. We
                 prove that under the site model, dynamic-linear has
                 greater availability than any static algorithm,
                 including weighted voting, if there are four or more
                 sites in the network. In the link model, we consider
                 all biconnected five-site networks and a wide variety
                 of failure and repair rates. In all cases considered,
                 dynamic-linear had greater availability than any static
                 algorithm.",
  acknowledgement = ack-nhfb,
  affiliation =  "George Mason Univ., Fairfax, VA, USA",
  annote =       "mathematical analysis shows that dynamic-linear is
                 better than static voting algorithms",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; reliability",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf D.4.3}: Software, OPERATING
                 SYSTEMS, File Systems Management, Distributed file
                 systems. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Distributed systems. {\bf C.4}:
                 Computer Systems Organization, PERFORMANCE OF SYSTEMS,
                 Reliability, availability, and serviceability.",
}

@Article{Dasgupta:1990:FCC,
  author =       "Partha Dasgupta and Zvi M. Kedem",
  title =        "The Five-Color Concurrency Control Protocol:
                 Non-Two-Phase Locking in General Databases",
  journal =      j-TODS,
  volume =       "15",
  number =       "2",
  pages =        "281--307",
  month =        jun,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 156 124",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-2/p281-dasgupta/p281-dasgupta.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-2/p281-dasgupta/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/78927.html",
  abstract =     "Concurrency control protocols based on two-phase
                 locking are a popular family of locking protocols that
                 preserve serializability in general (unstructured)
                 database systems. A concurrency control algorithm (for
                 databases with no inherent structure) is presented that
                 is practical, non two-phase, and allows varieties of
                 serializable logs not possible with any commonly known
                 locking schemes. All transactions are required to
                 predeclare the data they intend to read or write. Using
                 this information, the protocol anticipates the
                 existence (or absence) of possible conflicts and hence
                 can allow non-two-phase locking.\par

                 It is well known that serializability is characterized
                 by acyclicity of the conflict graph representation of
                 interleaved executions. The two-phase locking protocols
                 allow only {\em forward\/} growth of the paths in the
                 graph. The {\em Five Color\/} protocol allows the
                 conflict graph to grow in any direction (avoiding
                 two-phase constraints) and prevents cycles in the graph
                 by maintaining transaction access information in the
                 form of data-item markers. The read and write set
                 information can also be used to provide relative
                 immunity from deadlocks.",
  acknowledgement = ack-nhfb,
  affiliation =  "Georgia Inst. of Technol., Atlanta, GA, USA",
  annote =       "allow varieties of serializable logs not possible with
                 known locking schemes; the protocol anticipates the
                 existence of possible conflicts and hence can allow
                 non-two-phase locking.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; predeclared lock sets
                 TODS",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency.",
}

@Article{Moore:1990:DTA,
  author =       "James C. Moore and William B. Richmond and Andrew B.
                 Whinston",
  title =        "A Decision-Theoretic Approach to Information
                 Retrieval",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "311--340",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "91h:68037",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p311-moore/p311-moore.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p311-moore/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/88597.html",
  abstract =     "We present the file search problem in a
                 decision-theoretic framework, and discuss a variation
                 of it that we call the common index problem. The goal
                 of the common index problem is to return the best
                 available record in the file, where {\em best\/} is in
                 terms of a class of user preferences. We use dynamic
                 programming to construct an optimal algorithm using two
                 different optimality criteria, and we develop
                 sufficient conditions for obtaining complete
                 information.",
  acknowledgement = ack-nhfb,
  affiliation =  "Purdue Univ., West Lafayette, IN, USA",
  annote =       "searching a file for a best record rather than a
                 specific one; the assumption is that the preferences of
                 the ith individual can be represented as a composite
                 where preferences are based on the same index for all
                 users",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Economics; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; approximate algorithms; design; economics;
                 economics of information; information retrieval;
                 theory",
  subject =      "{\bf E.5}: Data, FILES, Sorting/searching. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Search
                 process.",
}

@Article{Westland:1990:SOC,
  author =       "J. Christopher Westland",
  title =        "Scaling Up Output Capacity and Performance Results
                 from Information Systems Prototypes",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "341--358",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p341-westland/p341-westland.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p341-westland/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/87943.html",
  abstract =     "The advantage of information system prototyping arises
                 from its predict problems and end-user satisfaction
                 with a system early in the development process, before
                 significant commitments of time and effort have been
                 made. Predictions of problems and end-user satisfaction
                 have risen in importance with the increasing complexity
                 of business information systems and the exponential
                 growth of database size. This research investigates the
                 reporting of information to an end user, and the
                 process of inferring from a prototype to a full-scale
                 information system. This inference is called {\em
                 scaling up}, and is an important part of the systems
                 development planning process. The research investigates
                 information systems reporting from a linguistic
                 perspective, where a database is used as a central
                 receptacle for information storage. It then
                 investigates the manner in which reporting statistics
                 from the prototype information system may be used to
                 infer the behavior and performance of the full-scale
                 system. An example is presented for the application of
                 the algorithm, and the final section discusses the
                 usefulness, application, and implications of the
                 algorithm developed in this research.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ. of Southern California, Los Angeles, CA, USA",
  annote =       "a coarse introduction, mainly the recall of
                 information retrieval systems; the mathematics is wrong
                 and too simple.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; inclusion-exclusion principle;
                 performance",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval. {\bf
                 C.4}: Computer Systems Organization, PERFORMANCE OF
                 SYSTEMS, Measurement techniques.",
}

@Article{Alonso:1990:DCI,
  author =       "Rafael Alonso and Daniel Barbara and H{\'e}ctor
                 Garc{\'\i}a-Molina",
  title =        "Data Caching Issues in an Information Retrieval
                 System",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "359--384",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p359-alonso/p359-alonso.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p359-alonso/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/87848.html",
  abstract =     "Currently, a variety of information retrieval systems
                 are available to potential users\ldots{}. While in many
                 cases these systems are accessed from personal
                 computers, typically no advantage is taken of the
                 computing resources of those machines (such as local
                 processing and storage). In this paper we explore the
                 possibility of using the user's local storage
                 capabilities to cache data at the user's site. This
                 would improve the response time of user queries albeit
                 at the cost of incurring the overhead required in
                 maintaining multiple copies. In order to reduce this
                 overhead it may be appropriate to allow copies to
                 diverge in a controlled fashion\ldots{}. Thus, we
                 introduce the notion of quasi-copies, which embodies
                 the ideas sketched above. We also define the types of
                 deviations that seem useful, and discuss the available
                 implementation strategies.",
  acknowledgement = ack-nhfb,
  affiliation =  "Princeton Univ., NJ, USA",
  annote =       "focus: cache coherency in distributed information
                 retrieval systems one central server site, data cached
                 at client site. Less response time, but more overhead
                 in maintaining multiple copies. To reduce overhead,
                 allow copies to diverge in a controlled fashion -
                 notion of ``quasi copies'' and quasi-caching. - users
                 can precisely define limits for divergence of
                 quasi-copies. - reduces update propagation overhead -
                 main difference with materialized views is that here
                 user can establish degree of coherency - implementation
                 possibilities: invalidate/refresh out-of-date data,
                 include automatic expiration date etc. Paper surveys
                 various implementation strategies and their tradeoffs.
                 users give two types of conditions on quasi-caches:
                 selection and coherency. - selection conditions specify
                 which object images will be cached at the user site. -
                 modifiers: - Add/drop (add to cache or remove) -
                 compulsory or advisory (whether caching is to be
                 enforced or to be taken as a hint) - query optimizer
                 can take advantage if caching is compulsory - advisory
                 selection gives greater system flexibility - static /
                 dynamic (static => objects selected once when the
                 condition is issued by a user, dynamic => changes in
                 data cause objects to be added/dropped dynamically). -
                 triggering delay: specifies acceptable delay for
                 dynamic selections - coherency conditions define the
                 allowable deviations between an object and its images.
                 - default: image must have a valid value (though
                 out-of-date) - delay: how much time an image may lag
                 behind an object - version: acceptable lag of how many
                 versions - periodic: image to be refreshed periodically
                 - arithmetic: deviations limited by the difference
                 between the values of the object and its image. - can
                 also have inter-object consistency constraints.
                 implementation issues - transmission delays and
                 failures: ``null'' messages sent out by central site to
                 check if client is alive etc. - what to propagate: -
                 data message: contains new values to overwrite old ones
                 in cache - invalidation message: only identifies
                 invalid object to be purged from the cache, but does
                 not contain new values. - version number message:
                 provides new version numbers only - no new data -
                 implicit invalidation: no message from central site,
                 cache images automatically invalidated after a certain
                 time. - when to propagate: - last minute: delayed until
                 a selection/coherency condition is about to be
                 violated. - immediately: as soon as updates occur -
                 delayed update at central site, so no cache conditions
                 are violated. - collapsing conditions: - possible to
                 collapse several coherency conditions on same object
                 into one - load balancing: central site can partially
                 off-load enforcement of consistency to clients
                 describes a probabilistic performance model and
                 simulation results - simulation parameters are network
                 traffic, query processing time, update installation
                 time etc. conclusions: - quasi-caching can potentially
                 improve performance and availability - problems if: -
                 selection and consistency constraints are complex -
                 large number of updates at central site - open issues:
                 - how much data to cache - how does choice of when to
                 propagate updates affect performance etc.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Management; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cache coherency; data sharing; design; information
                 retrieval systems; management; performance",
  subject =      "{\bf H.3.5}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Online Information Services, Data bank
                 sharing. {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf D.4.7}: Software, OPERATING
                 SYSTEMS, Organization and Design, Distributed systems.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems.",
}

@Article{Kifer:1990:CTQ,
  author =       "Michael Kifer and Eliezer L. Lozinskii",
  title =        "On Compile-Time Query Optimization In Deductive
                 Databases By Means of Static Filtering",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "385--426",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 081 178",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p385-kifer/p385-kifer.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p385-kifer/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/87121.html",
  abstract =     "We extend the query optimization techniques known as
                 algebraic manipulations with relational expressions
                 [48] to work with deductive databases. In particular,
                 we propose a method for moving data-independent
                 selections and projections into recursive axioms, which
                 extends all other known techniques for performing that
                 task [2, 3, 9, 18, 20]. We also show that, in a
                 well-defined sense, our algorithm is optimal among the
                 algorithms that propagate data-independent selections
                 through recursion.",
  acknowledgement = ack-nhfb,
  affiliation =  "State Univ. of New York, Stony Brook, NY, USA",
  annote =       "Discusses algebraic optimizations for logic
                 programs.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Management; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; dataflow; deductive databases; design;
                 filtering; fixpoint; graph representation; inference;
                 management; performance; projection; recursive rules;
                 selection; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.3.3}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Search and Retrieval, Selection process. {\bf I.2.3}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Deduction and Theorem Proving, Deduction. {\bf I.2.8}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Problem Solving, Control Methods, and Search, Graph and
                 tree search strategies.",
}

@Article{Agrawal:1990:DTC,
  author =       "Rakesh Agrawal and Shaul Dar and H. V. Jagadish",
  title =        "Direct Transitive Closure Algorithms: Design and
                 Performance Evaluation",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "427--458",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68Q25 68R05)",
  MRnumber =     "91h:68029",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p427-agrawal/p427-agrawal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p427-agrawal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/88888.html",
  abstract =     "We present new algorithms for computing transitive
                 closure of large database relations. Unlike iterative
                 algorithms, such as the seminaive and logarithmic
                 algorithms, the termination of our algorithms does not
                 depend on the length of paths in the underlying graph
                 (hence the name {\em direct\/} algorithms). Besides
                 reachability computations, the proposed algorithms can
                 also be used for solving path problems. We discuss
                 issues related to the efficient implementation of these
                 algorithms, and present experimental results that show
                 the direct algorithms perform uniformly better than the
                 iterative algorithms. A side benefit of this work is
                 that we have proposed a new methodology for evaluating
                 the performance of recursive queries.",
  acknowledgement = ack-nhfb,
  affiliation =  "AT\&T Bell Labs., Murray Hill, NJ, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "deductive databases; design; experimentation;
                 performance; query processing; recursive query
                 processing TODS, algorithms; transitive closure",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems. {\bf I.2.8}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Problem Solving, Control
                 Methods, and Search. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems.",
}

@Article{Zhang:1990:NSC,
  author =       "Weining Zhang and Clement T. Yu and Daniel Troy",
  title =        "Necessary and Sufficient Conditions to Linearize
                 Doubly Recursive Programs in Logic Databases",
  journal =      j-TODS,
  volume =       "15",
  number =       "3",
  pages =        "459--482",
  month =        sep,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68T15)",
  MRnumber =     "91h:68036",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-3/p459-zhang/p459-zhang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-3/p459-zhang/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/89237.html",
  abstract =     "Linearization of nonlinear recursive programs is an
                 important issue in logic databases for both practical
                 and theoretical reasons. If a nonlinear recursive
                 program can be transformed into an equivalent linear
                 recursive program, then it may be computed more
                 efficiently than when the transformation is not
                 possible. We provide a set of necessary and sufficient
                 conditions for a simple doubly recursive program to be
                 equivalent to a simple linear recursive program. The
                 necessary and sufficient conditions can be verified
                 effectively.",
  acknowledgement = ack-nhfb,
  acmcrnumber =  "9211-0888",
  affiliation =  "Illinois Univ., Chicago, IL, USA",
  annote =       "extends authors' previous results",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CTYU TODS, algorithms; design; logic database;
                 theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving, Logic programming.",
}

@Article{Can:1990:CEC,
  author =       "Fazli Can and Esen A. Ozkarahan",
  title =        "Concepts and Effectiveness of the
                 Cover-Coefficient-Based Clustering Methodology for Text
                 Databases",
  journal =      j-TODS,
  volume =       "15",
  number =       "4",
  pages =        "483--517",
  month =        dec,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-4/p483-can/p483-can.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-4/p483-can/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/99938.html",
  abstract =     "A new algorithm for document clustering is introduced.
                 The base concept of the algorithm, the cover
                 coefficient (CC) concept, provides a means of
                 estimating the number of clusters within a document
                 database and related indexing and clustering
                 analytically. The CC concept is used also to identify
                 the cluster seeds and to form clusters with these
                 seeds. It is shown that the complexity of the
                 clustering process is very low. The retrieval
                 experiments show that the information-retrieval
                 effectiveness of the algorithm is compatible with a
                 very demanding complete linkage clustering method that
                 is known to have good retrieval performance. The
                 experiments also show that the algorithm is 15.1 to
                 63.5 (with an average of 47.5) percent better than four
                 other clustering algorithms in cluster-based
                 information retrieval. The experiments have validated
                 the indexing-clustering relationships and the
                 complexity of the algorithm and have shown improvements
                 in retrieval effectiveness. In the experiments two
                 document databases are used: TODS214 and INSPEC. The
                 latter is a common database with 12,684 documents.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Syst. Anal., Miami Univ., Oxford, OH, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cluster validity; clustering-indexing relationships;
                 cover coefficient; decoupling coefficient; design;
                 document retrieval; experimentation; Inf. retrieval
                 TODS, algorithms; retrieval effectiveness",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Clustering. {\bf H.3.1}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Content Analysis and
                 Indexing, Indexing methods. {\bf H.3.6}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Library
                 Automation, Large text archives. {\bf I.7.0}: Computing
                 Methodologies, TEXT PROCESSING, General.",
}

@Article{Nakano:1990:TOR,
  author =       "Ryohei Nakano",
  title =        "Translation with Optimization from Relational Calculus
                 to Relational Algebra Having Aggregate Functions",
  journal =      j-TODS,
  volume =       "15",
  number =       "4",
  pages =        "518--557",
  month =        dec,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "1 093 243",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-4/p518-nakano/p518-nakano.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-4/p518-nakano/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/99943.html",
  abstract =     "Most of the previous translations of relational
                 calculus to relational algebra aimed at proving that
                 the two languages have the equivalent expressive power,
                 thereby generating very complicated relational algebra
                 expressions, especially when aggregate functions are
                 introduced. This paper presents a rule-based
                 translation method from relational calculus expressions
                 having both aggregate functions and null values to
                 optimized relational algebra expressions. Thus, logical
                 optimization is carried out through translation. The
                 translation method comprises two parts: the
                 translational of the relational calculus kernel and the
                 translation of aggregate functions. The former uses the
                 familiar step-wise rewriting strategy, while the latter
                 adopts a two-phase rewriting strategy via standard
                 aggregate expressions. Each translation proceeds by
                 applying a heuristic rewriting rule in preference to a
                 basic rewriting rule. After introducing SQL-type null
                 values, their impact on the translation is thoroughly
                 investigated, resulting in several extensions of the
                 translation. A translation experiment with many queries
                 shows that the proposed translation method generates
                 optimized relational algebra expressions. It is shown
                 that heuristic rewriting rules play an essential role
                 in the optimization. The correctness of the present
                 translation is also shown.\par

                 \ldots{} aggregate expressions. Each translation
                 proceeds by applying a heuristic rewriting rule in
                 preference to a basic rewriting rule. After introducing
                 SQL-type null values, their impact on the translation
                 is thoroughly investigated, resulting in several
                 extensions of the translation. A translation experiment
                 with many queries shows that the proposed translation
                 method generates optimized relational",
  acknowledgement = ack-nhfb,
  affiliation =  "Knowledge Syst. Lab., NTT Commun. and Inf. Process.
                 Lab., Kanagawa, Japan",
  annote =       "a rule-based translation method from expressions
                 having aggregate functions being a two-phase rewriting
                 strategy; experiment with many queries shows that
                 heuristic rules are essential in optimization; the
                 translation will from the front end of a database
                 machine, MACH, developed by the author.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; languages; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing.",
}

@Article{Jagadish:1990:CTM,
  author =       "H. V. Jagadish",
  title =        "A Compression Technique to Materialize Transitive
                 Closure",
  journal =      j-TODS,
  volume =       "15",
  number =       "4",
  pages =        "558--598",
  month =        dec,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "1 093 244",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-4/p558-jagadish/p558-jagadish.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-4/p558-jagadish/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/99944.html",
  abstract =     "An important feature of database support for expert
                 systems is the ability of the database to answer
                 queries regarding the existence of a path from one node
                 to another in the directed graph underlying some
                 database relation. Given just the database relation,
                 answering such a query is time-consuming, but given the
                 transitive closure of the database relation a table
                 look-up suffices. We present an indexing scheme that
                 permits the storage of the pre-computed transitive
                 closure of a database relation in a compressed form.
                 The existence of a specified tuple in the closure can
                 be determined from this compressed store by a single
                 look-up followed by an index comparison. We show how to
                 add nodes and arcs to the compressed closure
                 incrementally. We also suggest how this compression
                 technique can be used to reduce the effort required to
                 compute the transitive closure.",
  acknowledgement = ack-nhfb,
  affiliation =  "AT\&T Bell Lab., Murray Hill, NJ, USA",
  annote =       "an indexing scheme that permits the storage of the
                 pre-computed transitive closure",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "performance; recursive query processing TODS,
                 algorithms",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf I.2.1}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Applications
                 and Expert Systems. {\bf I.2.8}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Problem
                 Solving, Control Methods, and Search.",
}

@Article{Omiecinski:1990:PAR,
  author =       "Edward Omiecinski and Peter Scheuermann",
  title =        "A Parallel Algorithm for Record Clustering",
  journal =      j-TODS,
  volume =       "15",
  number =       "4",
  pages =        "599--624",
  month =        dec,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (68Q25)",
  MRnumber =     "1 093 245",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-4/p599-omiecinski/p599-omiecinski.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-4/p599-omiecinski/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/99947.html",
  abstract =     "We present an efficient heuristic algorithm for record
                 clustering that can run on a SIMD machine. We introduce
                 the P-tree, and its associated numbering scheme, which
                 in the split phase allows each processor independently
                 to compute the unique cluster number of a record
                 satisfying an arbitrary query. We show that by
                 restricting ourselves in the merge phase to combining
                 only sibling clusters, we obtain a parallel algorithm
                 whose speedup ratio is optimal in the number of
                 processors used. Finally, we report on experiments
                 showing that our method produces substantial savings in
                 an environment with relatively little overlap among the
                 queries.",
  acknowledgement = ack-nhfb,
  affiliation =  "Sch. of Inf. and Comput. Sci., Georgia Inst. of
                 Technol., Atlanta, GA, USA",
  annote =       "for SIMD machine",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; experimentation; performance; Physical
                 database design TODS, algorithms",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Clustering. {\bf C.1.2}: Computer Systems Organization,
                 PROCESSOR ARCHITECTURES, Multiple Data Stream
                 Architectures (Multiprocessors),
                 Single-instruction-stream, multiple-data-stream
                 processors (SIMD). {\bf F.2.2}: Theory of Computation,
                 ANALYSIS OF ALGORITHMS AND PROBLEM COMPLEXITY,
                 Nonnumerical Algorithms and Problems, Sequencing and
                 scheduling. {\bf H.3.2}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Storage.
                 {\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design.",
}

@Article{Lomet:1990:HTM,
  author =       "David B. Lomet and Betty Salzberg",
  title =        "The {hB-Tree}: a Multiattribute Indexing Method with
                 Good Guaranteed Performance",
  journal =      j-TODS,
  volume =       "15",
  number =       "4",
  pages =        "625--658",
  month =        dec,
  year =         "1990",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1990-15-4/p625-lomet/p625-lomet.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1990-15-4/p625-lomet/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/99949.html",
  abstract =     "A new multiattribute index structure called the
                 hB-tree is introduced. It is derived from the
                 K-D-B-tree of Robinson [15] but has additional
                 desirable properties. The hB-tree internode search and
                 growth processes are precisely analogous to the
                 corresponding processes in B-trees [1]. The intranode
                 processes are unique. A k-d tree is used as the
                 structure within nodes for very efficient searching.
                 Node splitting requires that this k-d tree be split.
                 This produces nodes which no longer represent
                 brick-like regions in k-space, but that can be
                 characterized as holey bricks, bricks in which
                 subregions have been extracted. We present results that
                 guarantee hB-tree users decent storage utilization,
                 reasonable size index terms, and good search and insert
                 performance. These results guarantee that the hB-tree
                 copes well with arbitrary distributions of keys.",
  acknowledgement = ack-nhfb,
  annote =       "multiattributable index; node splitting produces nodes
                 that can be characterized as holey bricks",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; multi dimensional range queries TODS;
                 performance",
  subject =      "{\bf E.1}: Data, DATA STRUCTURES, Trees. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Access methods. {\bf H.3.2}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Storage, File organization.",
}

@Article{Abiteboul:1991:RBL,
  author =       "Serge Abiteboul and St{\'e}phane Grumbach",
  title =        "A Rule-Based Language with Functions and Sets",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "1--30",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68Q45 (68N17 68P15)",
  MRnumber =     "92a:68067",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p1-abiteboul/p1-abiteboul.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p1-abiteboul/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103141.html",
  abstract =     "A logic based language for manipulating complex
                 objects constructed using set and tuple constructors is
                 introduced. A key feature of the COL language is the
                 use of base and derived data functions. Under some
                 stratification restrictions, the semantics of programs
                 is given by a minimal and justified model that can be
                 computed using a finite sequence of fixpoints. The
                 language is extended using external functions and
                 predicates. An implementation of COL in a functional
                 language is briefly discussed.",
  acknowledgement = ack-nhfb,
  affiliation =  "INRIA, Le Chesnay, France",
  annote =       "the COL language uses base and derived data
                 functions",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "complex objects; deductive databases; deductive
                 knowledge TODS, design; fixpoint semantics; knowledge
                 bases; languages; object-oriented databases; rule
                 based; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving. {\bf F.4.1}: Theory of Computation,
                 MATHEMATICAL LOGIC AND FORMAL LANGUAGES, Mathematical
                 Logic, Logic programming.",
}

@Article{Mendelzon:1991:FDH,
  author =       "Alberto O. Mendelzon and Peter T. Wood",
  title =        "Functional dependencies in {Horn} clause queries",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "31--55",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (68N17)",
  MRnumber =     "92b:68028",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p31-mendelzon/p31-mendelzon.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p31-mendelzon/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103142.html",
  abstract =     "When a database query is expressed as a set of Horn
                 clauses whose execution is by top-down resolution of
                 goals, there is a need to improve the backtracking
                 behavior of the interpreter. Rather than putting on the
                 programmer the onus of using extra-logical operators
                 such as {\em cut\/} to improve performance, we show
                 that some uses of the cut can be automated by inferring
                 them from functional dependencies. This requires some
                 knowledge of which variables are guaranteed to be bound
                 at query execution time; we give a method for deriving
                 such information using data flow analysis.",
  acknowledgement = ack-nhfb,
  affiliation =  "Comput. Syst. Res. Inst., Toronto Univ., Ont.,
                 Canada",
  annote =       "some uses of the cut can be automated by inferring
                 them; this requires knowledge of which variables are
                 bound at execution time",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data flow analysis; design; functional dependency;
                 logic programming; performance; relational database;
                 theory; theory deductive knowledge TODS, algorithms",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving, Deduction. {\bf I.2.3}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving, Logic programming.",
}

@Article{Meghini:1991:COF,
  author =       "C. Meghini and C. Thanos",
  title =        "The Complexity of Operations on a Fragmented
                 Relation",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "56--87",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68Q25)",
  MRnumber =     "92g:68036",
  MRreviewer =   "K. Marguerite Hafen",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p56-meghini/p56-meghini.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p56-meghini/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103143.html",
  abstract =     "Data fragmentation is an important aspect of
                 distributed database design, in which portions of
                 relations, tailored to the specific needs of local
                 applications, are defined to be further allocated to
                 the sites of the computer network supporting the
                 database system. In this paper we present a theory of
                 fragmentation with overlapping fragments to study the
                 complexity of the problems involved in checking the
                 completeness of a fragmentation schema and in querying
                 and updating a fragmented relation. We analyze these
                 problems from the complexity viewpoint and present
                 sound and complete algorithms for their solution.",
  acknowledgement = ack-nhfb,
  affiliation =  "Istituto di Elaborazione della Inf., CNR, Pisa,
                 Italy",
  annote =       "Proposes a two-step process to reconstruct first
                 minimal horizontal, then vertical covers. When
                 fragments overlap, optimization becomes intractible.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory;
                 Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; completeness of fragmentation schemas;
                 design; NP-hardness; performance; query optimization;
                 relation fragmentation; theory; updates; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing. {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems.
                 {\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf F.1.3}:
                 Theory of Computation, COMPUTATION BY ABSTRACT DEVICES,
                 Complexity Classes.",
}

@Article{Rabitti:1991:MAN,
  author =       "Fausto Rabitti and Elisa Bertino and Won Kim and
                 Darrell Woelk",
  title =        "A Model of Authorization for Next-Generation Database
                 Systems",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "88--131",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p88-rabitti/p88-rabitti.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p88-rabitti/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103144.html",
  abstract =     "The conventional models of authorization have been
                 designed for database systems supporting the
                 hierarchical, network, and relational models of data.
                 However, these models are not adequate for
                 next-generation database systems that support richer
                 data models that include object-oriented concepts and
                 semantic data modeling concepts. Rabitti, Woelk, and
                 Kim [14] presented a preliminary model of authorization
                 for use as the basis of an authorization mechanism in
                 such database systems. In this paper we present a
                 fuller model of authorization that fills a few major
                 gaps that the conventional models of authorization
                 cannot fill for next-generation database systems. We
                 also further formalize the notion of implicit
                 authorization and refine the application of the notion
                 of implicit authorization to object-oriented and
                 semantic modeling concepts. We also describe a user
                 interface for using the model of authorization and
                 consider key issues in implementing the authorization
                 model.",
  acknowledgement = ack-nhfb,
  affiliation =  "Microelectronics and Comput. Technol. Corp., CNR,
                 Pisa, Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Security; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "object-oriented database; security; security access
                 control TODS, design; semantic database; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf D.1.5}: Software,
                 PROGRAMMING TECHNIQUES, Object-oriented Programming.
                 {\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection.",
}

@Article{Weikum:1991:PRS,
  author =       "Gerhard Weikum",
  title =        "Principles and Realization Strategies of Multilevel
                 Transaction Management",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "132--180",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p132-weikum/p132-weikum.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p132-weikum/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103145.html",
  abstract =     "One of the demands of database system transaction
                 management is to achieve a high degree of concurrency
                 by taking into consideration the semantics of
                 high-level operations. On the other hand, the
                 implementation of such operations must pay attention to
                 conflicts on the storage representation levels below.
                 To meet these requirements in a layered architecture,
                 we propose a multilevel transaction management
                 utilizing layer-specific semantics. Based on the
                 theoretical notion of multilevel serializability, a
                 family of concurrency control strategies is developed.
                 Suitable recovery protocols are investigated for
                 aborting single transactions and for restarting the
                 system after a crash. The choice of levels involved in
                 a multilevel transaction strategy reveals an inherent
                 trade-off between increased concurrency and growing
                 recovery costs. A series of measurements has been
                 performed in order to compare several strategies.
                 Preliminary results indicate considerable performance
                 gains of the multilevel transaction approach.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., ETH Zurich, Switzerland",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Management; Performance;
                 Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "atomicity persistence concurrency control; concurrency
                 control recovery atomicity nested transactions TODS,
                 algorithms; design; management; multilevel
                 transactions; performance; persistence; reliability;
                 serializability",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf D.4.1}: Software,
                 OPERATING SYSTEMS, Process Management, Concurrency.
                 {\bf H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration, Logging and recovery. {\bf
                 D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Synchronization.",
}

@Article{Wolfson:1991:MPR,
  author =       "Ouri Wolfson and Amir Milo",
  title =        "The Multicast Policy and its Relationship to
                 Replicated Data Placement",
  journal =      j-TODS,
  volume =       "16",
  number =       "1",
  pages =        "181--205",
  month =        mar,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68M10)",
  MRnumber =     "92a:68042",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See corrigendum in \cite{Wolfson:1991:CMP}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-1/p181-wolfson/p181-wolfson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-1/p181-wolfson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103146.html",
  abstract =     "In this paper we consider the communication complexity
                 of maintaining the replicas of a logical data-item, in
                 a database distributed over a computer network. We
                 propose a new method, called the minimum spanning tree
                 write, by which a processor in the network should
                 multicast a write of a logical data-item, to all the
                 processors that store replicas of the items. Then we
                 show that the minimum spanning tree write is optimal
                 from the communication cost point of view. We also
                 demonstrate that the method by which a write is
                 multicast to all the replicas of a data-item affects
                 the optimal replication scheme of the item, i.e., at
                 which processors in the network the replicas should be
                 located. Therefore, next we consider the problem of
                 determining an optimal replication scheme for a data
                 item, assuming that each processor employs the minimum
                 spanning tree write at run-time. The problem for
                 general networks is shown NP-Complete, but we provide
                 efficient algorithms to obtain an optimal allocation
                 scheme for three common types of network topologies.
                 They are completely-connected, tree, and ring networks.
                 For these topologies, efficient algorithms are also
                 provided for the case in which reliability
                 considerations dictate a minimum number of replicas.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Columbia Univ., New York, NY,
                 USA",
  annote =       "mimumun spanning tree write and multicast to store
                 replicas",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "181 TODS, algorithms; complexity; computer network;
                 file allocation; message passing; NP-Complete;
                 performance; theory",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf C.4}: Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS. {\bf C.2.1}:
                 Computer Systems Organization, COMPUTER-COMMUNICATION
                 NETWORKS, Network Architecture and Design. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Distributed systems.",
}

@Article{Casanova:1991:STM,
  author =       "M. A. Casanova and A. L. Furtado and L. Tucherman",
  title =        "A Software Tool for Modular Database Design",
  journal =      j-TODS,
  volume =       "16",
  number =       "2",
  pages =        "209--234",
  month =        jun,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-2/p209-casanova/p209-casanova.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-2/p209-casanova/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103711.html",
  abstract =     "A modularization discipline for database schemas is
                 first described. The discipline incorporates both a
                 strategy for enforcing integrity constraints and a
                 tactic for organizing large sets of database
                 structures, integrity constraints, and operations. A
                 software tool that helps the development and
                 maintenance of database schemas modularized according
                 to the discipline is then presented. It offers a
                 user-friendly interface that guides the designer
                 through the various stages of the creation of a new
                 module or through the process of changing objects of
                 existing modules. The tool incorporates, in a
                 declarative style, a description of the design and
                 redesign rules behind the modularization discipline,
                 hence facilitating the incremental addition of new
                 expertise about database design.",
  acknowledgement = ack-nhfb,
  affiliation =  "Rio Sci. Center, IBM Brazil, Rio de Janeiro, Brazil",
  annote =       "first modularize, then use the tool to develop and
                 maintain schemas; functions declared to enforce
                 constraints",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data types; consistency preservation; design;
                 encapsulation; integrity constraints; languages;
                 logical database design; modular design; module
                 constructors",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf D.2.2}:
                 Software, SOFTWARE ENGINEERING, Tools and Techniques.
                 {\bf D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Modules, packages. {\bf
                 H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration.",
}

@Article{VanGelder:1991:STR,
  author =       "Allen {Van Gelder} and Rodney W. Topor",
  title =        "Safety and Translation of Relational Calculus
                 Queries",
  journal =      j-TODS,
  volume =       "16",
  number =       "2",
  pages =        "235--278",
  month =        jun,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "92c:68037",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-2/p235-van_gelder/p235-van_gelder.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-2/p235-van_gelder/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103712.html",
  abstract =     "Not all queries in relational calculus can be answered
                 sensibly when disjunction, negation, and universal
                 quantification are allowed. The class of relation
                 calculus queries or formulas that have sensible answers
                 is called the {\em domain independent\/} class which is
                 known to be undecidable. Subsequent research has
                 focused on identifying large decidable subclasses of
                 domain independent formulas. In this paper we
                 investigate the properties of two such classes: the
                 {\em evaluable\/} formulas and the {\em allowed\/}
                 formulas. Although both classes have been defined
                 before, we give simplified definitions, present short
                 proofs of their main properties, and describe a method
                 to incorporate equality.\par

                 Although evaluable queries have sensible answers, it is
                 not straightforward to compute them efficiently or
                 correctly. We introduce {\em relational algebra normal
                 form\/} for formulas from which form the correct
                 translation into relational algebra is trivial. We give
                 algorithms to transform an evaluable formula into an
                 equivalent {\em allowed\/} formula and from there into
                 relational algebra normal form. Our algorithms avoid
                 use of the so-called {\em Dom\/} relation, consisting
                 of all constants appearing in the database or the
                 query.\par

                 Finally, we describe a restriction under which every
                 domain independent formula is evaluable and argue that
                 the class of evaluable formulas is the largest
                 decidable subclass of the domain independent formulas
                 that can be efficiently recognized.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Santa Cruz, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "allowed formulas; domain independence; evaluable
                 formulas; existential normal; query translation;
                 relational algebra; relational calculus",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf F.4.1}: Theory of Computation,
                 MATHEMATICAL LOGIC AND FORMAL LANGUAGES, Mathematical
                 Logic, Model theory.",
}

@Article{Shasha:1991:OEQ,
  author =       "Dennis Shasha and Tsong-Li L. Wang",
  title =        "Optimizing Equijoin Queries in Distributed Databases
                 where Relations are Hash-Partitioned",
  journal =      j-TODS,
  volume =       "16",
  number =       "2",
  pages =        "279--308",
  day =          "1",
  month =        jun,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "92c:68036",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib; UnCover
                 library database",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-2/p279-shasha/p279-shasha.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-2/p279-shasha/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103713.html",
  abstract =     "Consider the class of distributed database systems
                 consisting of a set of nodes connected by a high
                 bandwidth network. Each node consists of a processor, a
                 random access memory, and a slower but much larger
                 memory such as a disk. There is no shared memory among
                 the nodes. The data are horizontally partitioned often
                 using a hash function. Such a description characterizes
                 many parallel or distributed database systems that have
                 recently been proposed, both commercial and academic.
                 We study the optimization problem that arises when the
                 query processor must repartition the relations and
                 intermediate results participating in a multijoin
                 query. Using estimates of the sizes of intermediate
                 relations, we show (1) optimum solutions for closed
                 chain queries; (2) the NP-completeness of the
                 optimization problem for star, tree, and general graph
                 queries; and (3) effective heuristics for these hard
                 cases.\par

                 Our general approach and many of our results extend to
                 other attribute partitioning schemes, for example,
                 sort-partitioning on attributes, and to partitioned
                 object databases.",
  acknowledgement = ack-nhfb,
  affiliation =  "Courant Inst. of Math. Sci., New York Univ., NY, USA",
  annote =       "No shared memory so that processor must repartition
                 the relations and intermediate results in a multijoin
                 query for lost hashkeys (not dynamic optimization);
                 optimum solutions for closed chain queries,
                 NP-completeness of star, tree, and general graph
                 queries and effective heuristics.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; equijoin; hashing; NP-complete problems;
                 performance; relational data models; spanning trees;
                 systems; theory",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Distributed
                 systems. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf C.4}:
                 Computer Systems Organization, PERFORMANCE OF
                 SYSTEMS.",
}

@Article{Cesarini:1991:DHM,
  author =       "F. Cesarini and G. Soda",
  title =        "A Dynamic Hash Method with Signature",
  journal =      j-TODS,
  volume =       "16",
  number =       "2",
  pages =        "309--337",
  day =          "1",
  month =        jun,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib; UnCover
                 library database",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-2/p309-cesarini/p309-cesarini.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-2/p309-cesarini/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103714.html",
  abstract =     "We present a dynamic external hash method that allows
                 retrieval of a record by only one access to mass
                 storage while maintaining a high load factor. The hash
                 function is based on generalized spiral storage. Both
                 primary and overflow records are allocated to the same
                 file, and file expansion depends on being able to
                 allocate every overflow chain to one bucket. An in-core
                 index, built by means of a signature function,
                 discriminates between primary and overflow records and
                 assures one access to storage in the case of either
                 successful or unsuccessful searching. Simulation
                 results confirm the good expected performance.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dipartimento di Sistemi e Inf., Florence Univ.,
                 Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; dynamic hashing; external hashing;
                 generalized spiral storage; performance; signature
                 functions",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval, Retrieval models.
                 {\bf E.2}: Data, DATA STORAGE REPRESENTATIONS,
                 Hash-table representations. {\bf H.3.3}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Search and Retrieval. {\bf E.5}: Data, FILES.",
}

@Article{King:1991:MRB,
  author =       "Richard P. King and Nagui Halim and H{\'e}ctor
                 Garc{\'\i}a-Molina and Christos A. Polyzois",
  title =        "Management of a Remote Backup Copy for Disaster
                 Recovery",
  journal =      j-TODS,
  volume =       "16",
  number =       "2",
  pages =        "338--368",
  month =        jun,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-2/p338-king/p338-king.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-2/p338-king/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/103715.html",
  abstract =     "A remote backup database system tracks the state of a
                 primary system, taking over transaction processing when
                 disaster hits the primary site. The primary and backup
                 sites are physically isolated so that failures at one
                 site are unlikely to propagate to the other. For
                 correctness, the execution schedule at the backup must
                 be equivalent to that at the primary. When the primary
                 and backup sites contain a single processor, it is easy
                 to achieve this property. However, this is harder to do
                 when each site contains multiple processors and sites
                 are connected via multiple communication lines. We
                 present an efficient transaction processing mechanism
                 for multiprocessor systems that guarantees this and
                 other important properties. We also present a database
                 initialization algorithm that copies the database to a
                 backup site while transactions are being processed.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM T.J. Watson Research Center, Yorktown Heights, NY,
                 USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; database initialization; hot spare; hot
                 standby; reliability; remote backup",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf D.4.5}: Software, OPERATING SYSTEMS,
                 Reliability, Backup procedures. {\bf H.2.7}:
                 Information Systems, DATABASE MANAGEMENT, Database
                 Administration, Logging and recovery.",
}

@Article{Gogolla:1991:TSV,
  author =       "Martin Gogolla and Uwe Hohenstein",
  title =        "Towards a Semantic View of an Extended
                 Entity-Relationship Model",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "369--416",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68Q55 (68P15 68Q60)",
  MRnumber =     "1 131 140",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 Distributed/gesturing.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p369-gogolla/p369-gogolla.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p369-gogolla/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111200.html",
  abstract =     "Nearly all query languages discussed recently for the
                 Entity-Relationship (ER) model do not possess a formal
                 semantics. Languages are often defined by means of
                 examples only. The reason for this phenomenon is the
                 essential gap between features of query languages and
                 theoretical foundations like algebras and calculi.
                 Known languages offer arithmetic capabilities and allow
                 for aggregates, but algebras and calculi defined for ER
                 models do not.\par

                 This paper introduces an extended ER model
                 concentrating nearly all concepts of known so-called
                 semantic data models in a few syntactical constructs.
                 Moreover, we provide our extended ER model with a
                 formal mathematical semantics. On this basis a
                 well-founded calculus is developed taking into account
                 data operations on arbitrary user-defined data types
                 and aggregate functions. We pay special attention to
                 arithmetic operations, as well as multivalued terms
                 allowing nested queries, in a uniform and consistent
                 manner. We prove our calculus only allows the
                 formulation of safe terms and queries yielding a finite
                 result, and to be (at least) as expressive as the
                 relational calculi.",
  acknowledgement = ack-nhfb,
  affiliation =  "Tech. Univ. Braunschweig, Germany",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data type; aggregate function; calculus;
                 design; entity-relationship model; formal semantics;
                 languages; relational completeness; safeness; semantic
                 data model; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf F.3.1}: Theory of
                 Computation, LOGICS AND MEANINGS OF PROGRAMS,
                 Specifying and Verifying and Reasoning about Programs,
                 Specification techniques. {\bf F.3.2}: Theory of
                 Computation, LOGICS AND MEANINGS OF PROGRAMS, Semantics
                 of Programming Languages. {\bf D.3.1}: Software,
                 PROGRAMMING LANGUAGES, Formal Definitions and Theory,
                 Semantics. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Query languages.",
}

@Article{Sciore:1991:UAS,
  author =       "Edward Sciore",
  title =        "Using Annotations to Support Multiple Kinds of
                 Versioning in an Object- Oriented Database System",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "417--438",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p417-sciore/p417-sciore.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p417-sciore/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111205.html",
  abstract =     "The concept of {\em annotation\/} from object-oriented
                 languages is adapted to object-oriented databases. It
                 is shown how annotations can be used to model
                 activities such as constraint checking, default values,
                 and triggers. Annotations also are an appropriate way
                 to model different versioning concepts. This paper
                 discusses three kinds of versioning---histories,
                 revisions, and alternatives---and demonstrates how each
                 one can be modeled effectively using annotations. The
                 use of annotations also allows other kinds of
                 versioning to be defined extensibly, and arbitrary
                 combinations of versions can be handled easily.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Boston Coll., Chestnut Hill,
                 MA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "configuration management; design; languages;
                 object-oriented databases; versions",
  subject =      "{\bf D.3.2}: Software, PROGRAMMING LANGUAGES, Language
                 Classifications, Object-oriented languages. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf D.1.5}: Software, PROGRAMMING
                 TECHNIQUES, Object-oriented Programming.",
}

@Article{Karabeg:1991:SRC,
  author =       "Dino Karabeg and Victor Vianu",
  title =        "Simplification Rules and Complete Axiomatization for
                 Relational Update Transactions",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "439--475",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68P20)",
  MRnumber =     "92g:68033",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p439-karabeg/p439-karabeg.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p439-karabeg/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111208.html",
  abstract =     "Relational update transactions consisting of line
                 programs of inserts, deletes, and modifications are
                 studied with respect to equivalence and simplification.
                 A sound and complete set of axioms for proving
                 transaction equivalence is exhibited. The axioms yield
                 a set of simplification rules that can be used to
                 optimize efficiently a large class of transactions of
                 practical interest. The simplification rules are
                 particularly well suited to a dynamic environment where
                 transactions are presented in an on-line fashion, and
                 where the time available for optimization may consist
                 of arbitrarily short and sparse intervals.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., San Diego, La Jolla, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design. {\bf F.2.2}: Theory of Computation, ANALYSIS OF
                 ALGORITHMS AND PROBLEM COMPLEXITY, Nonnumerical
                 Algorithms and Problems. {\bf G.2.2}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Graph
                 algorithms.",
}

@Article{Yu:1991:RTR,
  author =       "Philip S. Yu and Avraham Leff and Yann-Hang Lee",
  title =        "On Robust Transaction Routing and Load Sharing",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "476--512",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p476-yu/p476-yu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p476-yu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111210.html",
  abstract =     "In this paper we examine the issue of robust
                 transaction routing in a locally distributed database
                 environment where transaction characteristics such as
                 reference locality imply that certain processing
                 systems can be identified as being more suitable than
                 others for a given transaction class. A response time
                 based routing strategy can strike a balance between
                 indiscriminate sharing of the load and routing based
                 only on transaction affinity. Since response time
                 estimates depend on workload and system parameters that
                 may not be readily available, it is important to
                 examine the robustness of routing decisions to
                 information accuracy. We find that a strategy which
                 strictly tries to minimize the response time of
                 incoming transactions is sensitive to the accuracy of
                 certain parameter values. On the other hand, naive
                 strategies, that simply ignore the parameters in making
                 routing decisions, have even worse performance. Three
                 alternative strategies are therefore examined:
                 threshold, discriminatory, and adaptive. Instead of
                 just optimizing an incoming transaction's response
                 time, the first two strategies pursue a strategy that
                 is somewhat more oriented towards global optimization.
                 This is achieved by being more restrictive on either
                 the condition or the candidate for balancing the load.
                 The third strategy, while trying to minimize the
                 response time of individual incoming transactions,
                 employs a feedback process to adaptively adjust future
                 response time estimates. It monitors the discrepancy
                 between the actual and estimated response times and
                 introduces a correction factor based on regression
                 analysis. All three strategies are shown to be robust
                 with respect to the accuracy of workload and system
                 parameters used in the response time estimation.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Thomas J. Watson Res. Center, Yorktown Heights,
                 NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; distributed database; load balancing;
                 performance analysis; performance, PSYU TODS;
                 transaction routing",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Routing and layout. {\bf D.4.8}: Software, OPERATING
                 SYSTEMS, Performance. {\bf D.4.1}: Software, OPERATING
                 SYSTEMS, Process Management. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Distributed
                 systems.",
}

@Article{Negri:1991:FSS,
  author =       "M. Negri and G. Pelagatti and L. Sbattella",
  title =        "Formal Semantics of {SQL} Queries",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "513--534",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (03B50 03B70)",
  MRnumber =     "92i:68033",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p513-negri/p513-negri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p513-negri/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111212.html",
  abstract =     "The semantics of SQL queries is formally defined by
                 stating a set of rules that determine a syntax-driven
                 translation of an SQL query to a formal model. The
                 target model, called Extended Three Valued Predicate
                 Calculus (E3VPC), is largely based on a set of
                 well-known mathematical concepts. The rules which allow
                 the transformation of a general E3VPC expression to a
                 Canonical Form, which can be manipulated using
                 traditional, two-valued predicate calculus are also
                 given; in this way, problems like equivalence analysis
                 of SQL queries are completely solved. Finally, the fact
                 that reasoning about the equivalence of SQL queries
                 using two-valued predicate calculus, without taking
                 care of the real SQL semantics can lead to errors is
                 shown, and the reasons for this are analyzed.",
  acknowledgement = ack-nhfb,
  affiliation =  "Brescia Univ., Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; query equivalence; query semantics;
                 SQL; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, SQL. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Query
                 formulation. {\bf F.3.2}: Theory of Computation, LOGICS
                 AND MEANINGS OF PROGRAMS, Semantics of Programming
                 Languages.",
}

@Article{Roussopoulos:1991:IAM,
  author =       "Nicholas Roussopoulos",
  title =        "An Incremental Access Method for {ViewCache}: Concept,
                 Algorithms, and Cost Analysis",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "535--563",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p535-roussopoulos/p535-roussopoulos.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p535-roussopoulos/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111215.html",
  abstract =     "A {\em ViewCache\/} is a stored collection of pointers
                 pointing to records of underlying relations needed to
                 materialize a view. This paper presents an {\em
                 Incremental Access Method (IAM)\/} that amortizes the
                 maintenance cost of ViewCaches over a long time period
                 or indefinitely. Amortization is based on {\em
                 deferred\/} and other update propagation strategies. A
                 deferred update strategy allows a ViewCache to remain
                 outdated until a query needs to selectively or
                 exhaustively materialize the view. At that point, an
                 incremental update of the ViewCache is performed. This
                 paper defines a set of conditions under which
                 incremental access to the ViewCache is cost effective.
                 The decision criteria are based on some dynamically
                 maintained cost parameters, which provide accurate
                 information but require inexpensive
                 bookkeeping.\par

                 The IAM capitalizes on the ViewCache storage
                 organization for performing the update and the
                 materialization of the ViewCaches in an interleaved
                 mode using one-pass algorithms. Compared to the
                 standard technique for supporting views that requires
                 reexecution of the definition of the view, the IAM
                 offers significant performance advantages. We will show
                 that under favorable conditions, most of which depend
                 on the size of the incremental update logs between
                 consecutive accesses of the views, the incremental
                 access method outperforms query modification.
                 Performance gains are higher for multilevel ViewCaches
                 because all the I/O and CPU for handling intermediate
                 results are avoided.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Maryland Univ., College Park,
                 MD, USA",
  annote =       "replicated data management",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; performance; Relational Precomputation TODS,
                 algorithms; terms",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.2.3}:
                 Information Systems, DATABASE MANAGEMENT, Languages,
                 Query languages.",
}

@Article{Mukkamala:1991:NEC,
  author =       "Ravi Mukkamala and Sushil Jajodia",
  title =        "A Note on Estimating the Cardinality of the Projection
                 of a Database Relation",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "564--566",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-3/p564-mukkamala/p564-mukkamala.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-3/p564-mukkamala/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/111218.html",
  abstract =     "The paper by Ahad et al. [1] derives an analytical
                 expression to estimate the cardinality of the
                 projection of a database relation. In this note, we
                 propose to show that this expression is in error even
                 when all the parameters are assumed to be constant. We
                 derive the correct formula for this expression.",
  acknowledgement = ack-nhfb,
  affiliation =  "Old Dominion Univ., Norfolk, VA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "block access estimation; design; performance; query
                 cost-estimation; relational databases",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design.",
}

@Article{Wolfson:1991:CMP,
  author =       "O. Wolfson and A. Milo",
  title =        "Corrigendum: {``The Multicast Policy and its
                 Relationship to Replicated Data Placement'' [ACM Trans.
                 Database Systems {\bf 16} (1991), no. 1, 181--205, by
                 O. Wolfson and A. Milo] (MR 92a:68042)}",
  journal =      j-TODS,
  volume =       "16",
  number =       "3",
  pages =        "567--567",
  month =        sep,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68M10)",
  MRnumber =     "1 131 143",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Wolfson:1991:MPR}.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hernandez:1991:CTM,
  author =       "H{\'e}ctor J. Hern{\'a}ndez and Edward P. F. Chan",
  title =        "Constant-Time-Maintainable {BCNF} Database Schemes",
  journal =      j-TODS,
  volume =       "16",
  number =       "4",
  pages =        "571--599",
  month =        dec,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-4/p571-hernandez/p571-hernandez.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-4/p571-hernandez/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/115301.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., New Mexico State Univ., Las
                 Cruces, NM, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; boundedness; constraint enforcement; data
                 dependencies; design; query processing; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Normal forms. {\bf H.2.1}: Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Data models. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing.",
}

@Article{Hou:1991:SEA,
  author =       "Wen-Chi Hou and Gultekin {\"O}zsoyo{\u{g}}lu",
  title =        "Statistical Estimators for Aggregate Relational
                 Algebra Queries",
  journal =      j-TODS,
  volume =       "16",
  number =       "4",
  pages =        "600--654",
  month =        dec,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-4/p600-hou/p600-hou.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-4/p600-hou/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/115300.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. and Eng. Sci., Case Western Reserve
                 Univ., Cleveland, OH, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Management; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; management; performance; relational
                 algebra; sampling; selectivity; simple random sampling;
                 statistical estimators; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.3}: Mathematics of
                 Computing, PROBABILITY AND STATISTICS, Statistical
                 computing. {\bf G.2.m}: Mathematics of Computing,
                 DISCRETE MATHEMATICS, Miscellaneous. {\bf H.2.3}:
                 Information Systems, DATABASE MANAGEMENT, Languages,
                 Query languages.",
}

@Article{Negri:1991:DJN,
  author =       "M. Negri and G. Pelagatti",
  title =        "Distributive Join: a New Algorithm for Joining
                 Relations",
  journal =      j-TODS,
  volume =       "16",
  number =       "4",
  pages =        "655--669",
  month =        dec,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-4/p655-negri/p655-negri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-4/p655-negri/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/115299.html",
  abstract =     "This paper describes a new algorithm for performing
                 joins in the absence of access paths. This algorithm is
                 shown to perform better than the merging scan
                 algorithm, which can be considered the paradigm of join
                 algorithms. Finally this algorithm is compared with
                 another recent sub-sort-merge algorithm.",
  acknowledgement = ack-nhfb,
  affiliation =  "Politecnico di Milano, Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Algorithm, Performance, buffer, hashing, join, merging
                 scan, nested scan, sort, algorithms; buffer; design;
                 experimentation; hashing; join; merging scan; nested
                 scan; performance; sort",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Sorting and searching. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods.",
}

@Article{Moerkotte:1991:RCC,
  author =       "Guido Moerkotte and Peter C. Lockemann",
  title =        "Reactive Consistency Control in Deductive Databases",
  journal =      j-TODS,
  volume =       "16",
  number =       "4",
  pages =        "670--702",
  month =        dec,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-4/p670-moerkotte/p670-moerkotte.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-4/p670-moerkotte/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/115298.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Inst. fuer Programmstrukturen und Datenorganisation,
                 Karlsruhe Univ., Germany",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance;
                 Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; experimentation; performance;
                 theory",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Data description
                 languages (DDL). {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing. {\bf
                 I.2.3}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Deduction and Theorem Proving,
                 Deduction.",
}

@Article{Carey:1991:CDT,
  author =       "Michael J. Carey and Miron Livny",
  title =        "Conflict Detection Tradeoffs for Replicated Data",
  journal =      j-TODS,
  volume =       "16",
  number =       "4",
  pages =        "703--746",
  month =        dec,
  year =         "1991",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1991-16-4/p703-carey/p703-carey.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1991-16-4/p703-carey/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/115289.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Wisconsin Univ., Madison, WI, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Measurement;
                 Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; concurrency control; experimentation;
                 measurement; performance; replicated data",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Concurrency. {\bf D.4.8}:
                 Software, OPERATING SYSTEMS, Performance, Modeling and
                 prediction. {\bf D.4.8}: Software, OPERATING SYSTEMS,
                 Performance, Simulation. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases.",
}

@Article{Cattell:1992:OOB,
  author =       "R. G. G. Cattell and J. Skeen",
  title =        "Object Operations Benchmark",
  journal =      j-TODS,
  volume =       "17",
  number =       "1",
  pages =        "1--31",
  month =        mar,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-1/p1-cattell/p1-cattell.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-1/p1-cattell/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128766.html",
  abstract =     "Performance is a major issue in the acceptance of
                 object-oriented and relational database systems aimed
                 at engineering applications such as computer-aided
                 software engineering (CASE) and computer-aided design
                 (CAD). Because traditional database systems benchmarks
                 are inappropriate to measure performance for operations
                 on engineering objects, we designed a new benchmark
                 Object Operations version 1 (OO1) to focus on important
                 characteristics of these applications. OO1 is descended
                 from an earlier benchmark for simple database
                 operations and is based on several years experience
                 with that benchmark. In this paper we describe the OO1
                 benchmark and results we obtained running it on a
                 variety of database systems. We provide a careful
                 specification of the benchmark, show how it can be
                 implemented on database systems, and present evidence
                 that more than an order of magnitude difference in
                 performance can result from a DBMS implementation quite
                 different from current products; minimizing overhead
                 per database call, offloading database server
                 functionality to workstations, taking advantage of
                 large main memories, and using link-based methods.",
  acknowledgement = ack-nhfb,
  affiliation =  "Sun Microsyst., Mountain View, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Languages;
                 Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; CAD; CASE; client-server architecture;
                 design; engineering database benchmark;
                 experimentation; hypermodel; languages; measurement;
                 object operations benchmark; object-oriented DBMS's;
                 performance; relation of DBMS's; workstations",
  subject =      "{\bf K.6.2}: Computing Milieux, MANAGEMENT OF
                 COMPUTING AND INFORMATION SYSTEMS, Installation
                 Management, Benchmarks. {\bf D.1.5}: Software,
                 PROGRAMMING TECHNIQUES, Object-oriented Programming.
                 {\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 description languages (DDL). {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML). {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Database
                 (persistent) programming languages. {\bf H.2.8}:
                 Information Systems, DATABASE MANAGEMENT, Database
                 applications. {\bf K.6.2}: Computing Milieux,
                 MANAGEMENT OF COMPUTING AND INFORMATION SYSTEMS,
                 Installation Management, Performance and usage
                 measurement.",
}

@Article{Weddell:1992:RAF,
  author =       "Grant E. Weddell",
  title =        "Reasoning About Functional Dependencies Generalized
                 for Semantic Data Models",
  journal =      j-TODS,
  volume =       "17",
  number =       "1",
  pages =        "32--64",
  month =        mar,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68Q55)",
  MRnumber =     "1 161 053",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-1/p32-weddell/p32-weddell.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-1/p32-weddell/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128767.html",
  abstract =     "We propose a more general form of functional
                 dependency for semantic data models that derives from
                 their common feature in which the separate notions of
                 {\em domain\/} and {\em relation\/} in the relational
                 model are combined into a single notion of {\em class}.
                 This usually results in a richer terminological
                 component for their query languages, whereby terms may
                 navigate through any number of properties, including
                 none. We prove the richer expressiveness of this more
                 general functional dependency, and exhibit a sound and
                 complete set of inference axioms. Although the general
                 problem of decidability of their logical implication
                 remains open at this time, we present decision
                 procedures for cases in which the dependencies included
                 in a schema correspond to keys, or in which the schema
                 itself is acyclic. The theory is then extended to
                 include a form of conjunctive query. Of particular
                 significance is that the query becomes an additional
                 source of functional dependency. Finally, we outline
                 several applications of the theory to various problems
                 in physical design and in query optimization. The
                 applications derive from an ability to predict when a
                 query can have at most one solution.",
  acknowledgement = ack-nhfb,
  affiliation =  "Waterloo Univ., Ont., Canada",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; constraint theory; functional
                 dependencies; query optimization; semantic data models;
                 theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing.",
}

@Article{Paredaens:1992:CNA,
  author =       "Jan Paredaens and Dirk {Van Gucht}",
  title =        "Converting Nested Algebra Expressions into Flat
                 Algebra Expressions",
  journal =      j-TODS,
  volume =       "17",
  number =       "1",
  pages =        "65--93",
  month =        mar,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P10",
  MRnumber =     "93c:68018",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-1/p65-paredaens/p65-paredaens.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-1/p65-paredaens/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128768.html",
  abstract =     "Nested relations generalize ordinary flat relations by
                 allowing tuple values to be either atomic or set
                 valued. The nested algebra is a generalization of the
                 flat relational algebra to manipulate nested relations.
                 In this paper we study the expressive power of the
                 nested algebra relative to its operation on flat
                 relational databases. We show that the flat relational
                 algebra is rich enough to extract the same ``flat
                 information'' from a flat database as the nested
                 algebra does. Theoretically, this result implies that
                 recursive queries such as the transitive closure of a
                 binary relation cannot be expressed in the nested
                 algebra. Practically, this result is relevant to (flat)
                 relational query optimization.",
  acknowledgement = ack-nhfb,
  affiliation =  "Antwerp Univ., Belgium",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algebraic query transformation; algorithms; languages;
                 nested algebra; nested calculus; nested relations;
                 relational databases; theory, van Gucht relational data
                 model TODS",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.3.3}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Search and Retrieval, Query formulation. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models.",
}

@Article{Mohan:1992:ATR,
  author =       "C. Mohan and Don Haderle and Bruce Lindsay and Hamid
                 Pirahesh and Peter Schwarz",
  title =        "{ARIES}: a Transaction Recovery Method Supporting
                 Fine-Granularity Locking and Partial Rollbacks Using
                 Write-Ahead Logging",
  journal =      j-TODS,
  volume =       "17",
  number =       "1",
  pages =        "94--162",
  month =        mar,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: IBM Almaden Res. Ctr, Res. R.
                 No. RJ-6649, Jan. 1989, 45 pp.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-1/p94-mohan/p94-mohan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-1/p94-mohan/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128770.html",
  abstract =     "DB2$^{\rm TM}$, IMS, and Tandem$^{\rm TM}$ systems.
                 ARIES is applicable not only to database management
                 systems but also to persistent object-oriented
                 languages, recoverable file systems and
                 transaction-based operating systems. ARIES has been
                 implemented, to varying degrees, in IBM's OS/2$^{\rm
                 TM}$ Extended Edition Database Manager, DB2,
                 Workstation Data Save Facility/VM, Starburst and
                 QuickSilver, and in the University of Wisconsin's
                 EXODUS and Gamma database machine.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Almaden Res. Center, San Jose, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; buffer management; design; latching;
                 locking; performance; reliability; space management;
                 write-ahead logging",
  subject =      "{\bf H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration, Logging and recovery. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Recovery and restart. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Concurrency.
                 {\bf E.5}: Data, FILES, Backup/recovery. {\bf D.4.5}:
                 Software, OPERATING SYSTEMS, Reliability, Backup
                 procedures.",
}

@Article{Badrinath:1992:SBC,
  author =       "B. R. Badrinath and Krithi Ramamritham",
  title =        "Semantics-Based Concurrency Control: Beyond
                 Commutativity",
  journal =      j-TODS,
  volume =       "17",
  number =       "1",
  pages =        "163--199",
  month =        mar,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "93b:68019",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-1/p163-badrinath/p163-badrinath.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-1/p163-badrinath/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128771.html",
  abstract =     "The concurrency of transactions executing on atomic
                 data types can be enhanced through the use of semantic
                 information about operations defined on these types.
                 Hitherto, commutativity of operations has been
                 exploited to provide enhanced concurrency while
                 avoiding cascading aborts. We have identified a
                 property known as {\em recoverability\/} which can be
                 used to decrease the delay involved in processing
                 noncommuting operations while still avoiding cascading
                 aborts. When an invoked operation is {\em
                 recoverable\/} with respect to an uncommitted
                 operation, the invoked operation can be executed by
                 forcing a commit dependency between the invoked
                 operation and the uncommitted operation; the
                 transaction invoking the operation will not have to
                 wait for the uncommitted operation to abort or commit.
                 Further, this commit dependency only affects the order
                 in which the operations should commit, if both commit;
                 if either operation aborts, the other can still commit
                 thus avoiding cascading aborts. To ensure the
                 serializability of transactions, we force the
                 recoverability relationship between transactions to be
                 acyclic. Simulation studies, based on the model
                 presented by Agrawal et al. [1], indicate that using
                 recoverability, the turnaround time of transactions can
                 be reduced. Further, our studies show enhancement in
                 concurrency even when {\em resource constraints\/} are
                 taken into consideration. The magnitude of enhancement
                 is dependent on the resource contention; the lower the
                 resource contention, the higher the improvement.",
  acknowledgement = ack-nhfb,
  affiliation =  "Rutgers Univ., New Brunswick, NJ, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; concurrency control; performance; semantic
                 information",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.
                 {\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Recovery and restart. {\bf D.2.1}:
                 Software, SOFTWARE ENGINEERING,
                 Requirements/Specifications.",
}

@Article{Wang:1992:CTM,
  author =       "Ke Wang and Marc H. Graham",
  title =        "Constant-Time Maintainability: a Generalization of
                 Independence",
  journal =      j-TODS,
  volume =       "17",
  number =       "2",
  pages =        "201--246",
  month =        jun,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-2/p201-wang/p201-wang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-2/p201-wang/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128904.html",
  abstract =     "The {\em maintenance problem\/} of a database scheme
                 is the following decision problem: Given a consistent
                 database state $ \rho $ and a new tuple $u$ over some
                 relation scheme of $ \rho $, is the modified state $
                 \rho \cup u $ still consistent? A database scheme is
                 said to be {\em constant-time-maintainable(ctm)\/} if
                 there exists an algorithm that solves its maintenance
                 problem by making a fixed number of tuple retrievals.
                 We present a practically useful algorithm, called the
                 {\em canonical maintenance algorithm}, that solves the
                 maintenance problem of all ctm database schemes within
                 a ``not too large'' bound. A number of interesting
                 properties are shown for ctm database schemes, among
                 them that non-ctm database schemes are not maintainable
                 in less than a linear time in the state size. A test
                 method is given when only cover embedded functional
                 dependencies (fds) appear. When the given dependencies
                 consist of fds and the join dependency (jd) $ \bowtie
                 {\bf R} $ of the database scheme, testing whether a
                 database scheme is ctm is reduced to the case of cover
                 embedded fds. When dependency-preserving database
                 schemes with only equality-generating dependencies
                 (egds) are considered, it is shown that every ctm
                 database scheme has a set of dependencies that is
                 equivalent to a set of embedded fds, and thus, our test
                 method for the case of embedded fds can be applied. In
                 particular, this includes the important case of
                 lossless database schemes with only egds.",
  acknowledgement = ack-nhfb,
  affiliation =  "Chongqing Univ., Sichuan, China",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; chase; constraint enforcement; design;
                 functional dependency; independent database schemes;
                 join dependency; lossless join; relational database;
                 representative instance; tableau; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.0}:
                 Information Systems, DATABASE MANAGEMENT, General,
                 Security, integrity, and protection. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Normal forms. {\bf F.4.1}: Theory of
                 Computation, MATHEMATICAL LOGIC AND FORMAL LANGUAGES,
                 Mathematical Logic, Mechanical theorem proving.",
}

@Article{Becker:1992:RBO,
  author =       "Ludger Becker and Ralf Hartmut G{\"u}ting",
  title =        "Rule-Based Optimization and Query Processing in an
                 Extensible Geometric Database System",
  journal =      j-TODS,
  volume =       "17",
  number =       "2",
  pages =        "247--303",
  month =        jun,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-2/p247-becker/p247-becker.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-2/p247-becker/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128905.html",
  abstract =     "Gral is an extensible database system, based on the
                 formal concept of a many-sorted relational algebra.
                 Many-sorted algebra is used to define any application's
                 query language, its query execution language, and its
                 optimization rules. In this paper we describe Gral's
                 optimization component. It provides (1) a sophisticated
                 rule language --- rules are transformations of abstract
                 algebra expressions, (2) a general optimization
                 framework under which more specific optimization
                 algorithms can be implemented, and (3) several control
                 mechanisms for the application of rules. An
                 optimization algorithm can be specified as a series of
                 steps. Each step is defined by its own collection of
                 rules together with a selected control strategy.
                 \par

                 The general facilities are illustrated by the complete
                 design of an example optimizer --- in the form of a
                 rule file --- for a small nonstandard query language
                 and an associated execution language. The query
                 language includes selection, join, ordering, embedding
                 derived values, aggregate functions, and several
                 geometric operations. The example shows in particular
                 how the special processing techniques of a geometric
                 database systems, such as spatial join methods and
                 geometric index structures, can be integrated into
                 query processing and optimization of a relational
                 database system. A similar, though larger, optimizer is
                 fully functional within the geometric database system
                 implemented as a Gral prototype.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ. Gesamthochschule Siegen, Germany",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; extensibility; geometric query processing;
                 languages, Guting Gral TODS; many-sorted algebra;
                 optimization; relational algebra; rule-based
                 optimization",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Query languages. {\bf F.2.0}:
                 Theory of Computation, ANALYSIS OF ALGORITHMS AND
                 PROBLEM COMPLEXITY, General. {\bf E.2}: Data, DATA
                 STORAGE REPRESENTATIONS, Hash-table representations.
                 {\bf I.3.5}: Computing Methodologies, COMPUTER
                 GRAPHICS, Computational Geometry and Object Modeling,
                 Geometric algorithms, languages, and systems.",
}

@Article{Franaszek:1992:CCH,
  author =       "Peter A. Franaszek and John T. Robinson and Alexander
                 Thomasian",
  title =        "Concurrency Control for High Contention Environments",
  journal =      j-TODS,
  volume =       "17",
  number =       "2",
  pages =        "304--345",
  month =        jun,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-2/p304-franaszek/p304-franaszek.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-2/p304-franaszek/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128906.html",
  abstract =     "Future transaction processing systems may have
                 substantially higher levels of concurrency due to
                 reasons which include: (1) increasing disparity between
                 processor speeds and data access latencies, (2) large
                 numbers of processors, and (3) distributed databases.
                 Another influence is the trend towards longer or more
                 complex transactions. A possible consequence is
                 substantially more data contention, which could limit
                 total achievable throughput. In particular, it is known
                 that the usual locking method of concurrency control is
                 not well suited to environments where data contention
                 is a significant factor.\par

                 Here we consider a number of concurrency control
                 concepts and transaction scheduling techniques that are
                 applicable to high contention environments, and that do
                 not rely on database semantics to reduce contention.
                 These include {\em access invariance\/} and its
                 application to prefetching of data, approximations to
                 {\em essential blocking\/} such as {\em wait depth
                 limited\/} scheduling, and {\em phase dependent\/}
                 control. The performance of various concurrency control
                 methods based on these concepts are studied using
                 detailed simulation models. The results indicate that
                 the new techniques can offer substantial benefits for
                 systems with high levels of data contention.",
  acknowledgement = ack-nhfb,
  affiliation =  "Thomas J. Watson Res. Center, Yorktown Heights, NY,
                 USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; concurrency control; design; performance;
                 transaction processing",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf D.4.8}: Software,
                 OPERATING SYSTEMS, Performance. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design.",
}

@Article{Leng:1992:OWA,
  author =       "Chun-Wu Roger Leng and Dik Lun Lee",
  title =        "Optimal Weight Assignment for Signature Generation",
  journal =      j-TODS,
  volume =       "17",
  number =       "2",
  pages =        "346--373",
  month =        jun,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/bibdb.bib; Database/Graefe.bib;
                 Database/Wiederhold.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-2/p346-leng/p346-leng.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-2/p346-leng/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128907.html",
  abstract =     "Previous work on superimposed coding has been
                 characterized by two aspects. First, it is generally
                 assumed that signatures are generated from {\em
                 logical\/} text blocks of the same size; that is, each
                 block contains the same number of unique terms after
                 stopword and duplicate removal. We call this approach
                 the fixed-size block (FSB) method, since each text
                 block has the same size, as measured by the number of
                 unique terms contained in it. Second, with only a few
                 exceptions [6,7,8,9,17], most previous work has assumed
                 that each term in the text contributes the same number
                 of ones to the signature (i.e., the weight of the term
                 signatures is fixed). The main objective of this paper
                 is to derive an optimal weight assignment that assigns
                 weights to document terms according to their occurrence
                 and query frequencies in order to minimize the
                 false-drop probability. The optimal scheme can account
                 for both uniform and nonuniform occurrence and query
                 frequencies, and the signature generation method is
                 still based on hashing rather than on table lookup.
                 Furthermore, a new way of generating signatures, the
                 fixed-weight block (FWB) method, is introduced. FWB
                 controls the weight of {\em every\/} signature to a
                 constant, whereas in FSB, only the {\em expected\/}
                 signature weight is constant. We have shown that FWB
                 has a lower false-drop probability than that of the FSB
                 method, but its storage overhead is slightly higher.
                 Other advantages of FWB are that the optimal weight
                 assignment can be obtained analytically without making
                 unrealistic assumptions and that the formula for
                 computing the term signature weights is simple and
                 efficient.",
  acknowledgement = ack-nhfb,
  affiliation =  "Ohio State Univ., Columbus, OH, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access method; coding methods; design; document
                 retrieval; information retrieval; optimization;
                 performance; signature file; superimposed coding; text
                 retrieval",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Retrieval models. {\bf H.2.2}: Information Systems,
                 DATABASE MANAGEMENT, Physical Design, Access methods.
                 {\bf H.3.6}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Library Automation. {\bf I.7.1}:
                 Computing Methodologies, TEXT PROCESSING, Text
                 Editing.",
}

@Article{Tansel:1992:MRH,
  author =       "Abdullah U. Tansel and Lucy Garnett",
  title =        "On {M. A. Roth, H. F. Korth and A. Silberschatz:
                 ``Extended Algebra and Calculus for Nested Relational
                 Databases'' [ACM Trans. Database Systems {\bf 13}
                 (1988), no. 4, 389--417]}",
  journal =      j-TODS,
  volume =       "17",
  number =       "2",
  pages =        "374--383",
  month =        jun,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "1 167 047",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Roth:1988:EAC}.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-2/p374-tansel/p374-tansel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-2/p374-tansel/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/128908.html",
  abstract =     "We discuss the issues encountered in the extended
                 algebra and calculus languages for nested relations
                 defined by Roth, Korth, and Silberschatz.[4]. Their
                 equivalence proof between algebra and calculus fails
                 because of the keying problems and the use of extended
                 set operations. Extended set operations also have
                 unintended side effects. Furthermore, their calculus
                 seems to allow the generation of power sets, thus
                 making it more powerful than their algebra.",
  acknowledgement = ack-nhfb,
  affiliation =  "Bilkent Univ., Ankara, Turkey",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "equivalence of algebra and calculus; languages; nested
                 relations; relational algebra; relational calculus;
                 theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf F.4.1}: Theory of
                 Computation, MATHEMATICAL LOGIC AND FORMAL LANGUAGES,
                 Mathematical Logic. {\bf H.2.1}: Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Normal forms. {\bf
                 H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Data manipulation languages (DML).",
}

@Article{Bergamaschi:1992:TRC,
  author =       "Sonia Bergamaschi and Claudio Sartori",
  title =        "On Taxonomic Reasoning in Conceptual Design",
  journal =      j-TODS,
  volume =       "17",
  number =       "3",
  pages =        "385--422",
  month =        sep,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-3/p385-bergamaschi/p385-bergamaschi.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-3/p385-bergamaschi/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/132272.html",
  abstract =     "Taxonomic reasoning is a typical task performed by
                 many AI knowledge representation systems. In this
                 paper, the effectiveness of taxonomic reasoning
                 techniques as an active support to knowledge
                 acquisition and conceptual schema design is shown. The
                 idea developed is that by extending conceptual models
                 with {\em defined concepts\/} and giving them rigorous
                 logic semantics, it is possible to infer {\em isa\/}
                 relationships between concepts on the basis of their
                 descriptions. From a theoretical point of view, this
                 approach makes it possible to give a formal definition
                 for {\em consistency\/} and {\em minimality\/} of a
                 conceptual schema. From a pragmatic point of view it is
                 possible to develop an active environment that allows
                 automatic {\em classification\/} of a new concept in
                 the right position of a given taxonomy, ensuring the
                 consistency and minimality of a conceptual schema. A
                 formalism that includes the data semantics of models
                 giving prominence to type constructors (E/R, TAXIS,
                 GALILEO) and algorithms for taxonomic inferences are
                 presented: their soundness, completeness, and
                 tractability properties are proved. Finally, an
                 extended formalism and taxonomic inference algorithms
                 for models giving prominence to attributes (FDM, IFO)
                 are given.",
  acknowledgement = ack-nhfb,
  affiliation =  "Bologna Univ., Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; schema consistency; schema
                 minimality; semantic models; taxonomic reasoning;
                 theory; verification",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf I.2.4}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Knowledge
                 Representation Formalisms and Methods, Representation
                 languages. {\bf I.2.4}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Knowledge Representation
                 Formalisms and Methods, Frames and scripts.",
}

@Article{Markowitz:1992:REE,
  author =       "Victor M. Markowitz and Arie Shoshani",
  title =        "Representing Extended Entity-Relationship Structures
                 in Relational Databases: a Modular Approach",
  journal =      j-TODS,
  volume =       "17",
  number =       "3",
  pages =        "423--464",
  month =        sep,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-3/p423-markowitz/p423-markowitz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-3/p423-markowitz/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/132273.html",
  abstract =     "A common approach to database design is to describe
                 the structures and constraints of the database
                 application in terms of a semantic data model, and then
                 represent the resulting schema using the data model of
                 a commercial database management system. Often, in
                 practice, {\em Extended Entity-Relationship\/} (EER)
                 schemas are translated into equivalent relational
                 schemas. This translation involves different aspects:
                 representing the EER schema using relational
                 constructs, assigning names to relational attributes,
                 normalization, and merging relations. Considering these
                 aspects together, as is usually done in the design
                 methodologies proposed in the literature, is confusing
                 and leads to inaccurate results. We propose to treat
                 separately these aspects and split the translation into
                 four stages (modules) corresponding to the four aspects
                 mentioned above. We define criteria for both evaluating
                 the correctness of and characterizing the relationship
                 between alternative relational representations of EER
                 schemas.",
  acknowledgement = ack-nhfb,
  affiliation =  "Lawrence Berkeley Lab., CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; database design; design; extended
                 entity-relationship model; relational data model;
                 schema translation; semantic data model",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Normal
                 forms. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Data description languages
                 (DDL).",
}

@Article{Date:1992:SCG,
  author =       "C. J. Date and Ronald Fagin",
  title =        "Simple Conditions for Guaranteeing Higher Normal Forms
                 in Relational Databases",
  journal =      j-TODS,
  volume =       "17",
  number =       "3",
  pages =        "465--476",
  month =        sep,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-3/p465-date/p465-date.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-3/p465-date/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/132274.html",
  abstract =     "A key is {\em simple\/} if it consists of a single
                 attribute. It is shown that if a relation schema is in
                 third normal form and every key is simple, then it is
                 in projection-join normal form (sometimes called fifth
                 normal form), the ultimate normal form with respect to
                 projections and joins. Furthermore, it is shown that if
                 a relation schema is in Boyce-Codd normal form and {\em
                 some\/} key is simple, then it is in fourth normal form
                 (but not necessarily projection-join normal form).
                 These results give the database designer simple
                 sufficient conditions, defined in terms of functional
                 dependencies alone, that guarantee that the schema
                 being designed is automatically in higher normal
                 forms.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "5NF; BCNF; Boyce-Codd normal form; database design;
                 design; fifth normal form; fourth normal form 4NF;
                 functional dependency; join dependency; multivalued
                 dependency; normalization; PJ/NF; projection-join
                 normal form; relational database; simple key; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms.",
}

@Article{Hsu:1992:PEC,
  author =       "Meichun Hsu and Bin Zhang",
  title =        "Performance Evaluation of Cautious Waiting",
  journal =      j-TODS,
  volume =       "17",
  number =       "3",
  pages =        "477--512",
  month =        sep,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-3/p477-hsu/p477-hsu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-3/p477-hsu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/132275.html",
  abstract =     "We study a deadlock-free locking-based concurrency
                 control algorithm, called {\em cautious waiting}, which
                 allows for a limited form of waiting. The algorithm is
                 very simple to implement. We present an analytical
                 solution to its performance evaluation based on the
                 mean-value approach proposed by Tay et al. [18]. From
                 the modeling point of view, we are able to do away with
                 a major assumption used in Tay's previous work, and
                 therefore capture more accurately both the restart and
                 the blocking rates in the system. We show that to solve
                 for this model we only need to solve for the root of a
                 polynomial. The analytical tools developed enable us to
                 see that the cautious waiting algorithm manages to
                 achieve a {\em delicate\/} balance between restart and
                 blocking, and therefore is superior (i.e., has higher
                 throughput to {\em both\/} the no-waiting (i.e.,
                 immediate restart) and the general waiting algorithms)
                 under a wide range of system parameters. The study
                 substantiates the argument that balancing restart and
                 blocking is important in locking systems.",
  acknowledgement = ack-nhfb,
  affiliation =  "Digital Equipment Corp., Mountain View, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Management; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; cautious waiting; concurrency control;
                 management; performance, Concurrency control locking
                 TODS",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf D.4.1}: Software, OPERATING
                 SYSTEMS, Process Management, Concurrency.",
}

@Article{Abbott:1992:SRT,
  author =       "Robert K. Abbott and H{\'e}ctor Garc{\'\i}a-Molina",
  title =        "Scheduling Real-Time Transactions: a Performance
                 Evaluation",
  journal =      j-TODS,
  volume =       "17",
  number =       "3",
  pages =        "513--560",
  month =        sep,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-3/p513-abbott/p513-abbott.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-3/p513-abbott/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/132276.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Digital Equipment Corp., Littleton, MA, USA",
  annote =       "real-time",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; deadlines; locking protocols; performance;
                 real-time systems",
  subject =      "{\bf D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Scheduling. {\bf D.4.1}: Software,
                 OPERATING SYSTEMS, Process Management, Concurrency.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.",
}

@Article{Siegel:1992:MAR,
  author =       "Michael Siegel and Edward Sciore and Sharon Salveter",
  title =        "A Method for Automatic Rule Derivation to Support
                 Semantic Query Optimization",
  journal =      j-TODS,
  volume =       "17",
  number =       "4",
  pages =        "563--600",
  month =        dec,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (68T05)",
  MRnumber =     "1 197 198",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-4/p563-siegel/p563-siegel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-4/p563-siegel/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/146932.html",
  abstract =     "The use of inference rules to support intelligent data
                 processing is an increasingly important tool in many
                 areas of computer science. In database systems, rules
                 are used in semantic query optimization as a method for
                 reducing query processing costs. The savings is
                 dependent on the ability of experts to supply a set of
                 useful rules and the ability of the optimizer to
                 quickly find the appropriate transformations generated
                 by these rules. Unfortunately, the most useful rules
                 are not always those that would or could be specified
                 by an expert. This paper describes the architecture of
                 a system having two interrelated components: a combined
                 conventional/semantic query optimizer, and an automatic
                 rule deriver.\par

                 Our automatic rule derivation method uses intermediate
                 results from the optimization process to direct the
                 search for learning new rules. Unlike a system
                 employing only user-specified rules, a system with an
                 automatic capability can derive rules that may be true
                 only in the current state of the database and can
                 modify the rule set to reflect changes in the database
                 and its usage pattern.\par

                 This system has been implemented as an extension of the
                 EXODUS conventional query optimizer generator. We
                 describe the implementation, and show how semantic
                 query optimization is an extension of conventional
                 optimization in this context.",
  acknowledgement = ack-nhfb,
  affiliation =  "Boston Univ., MA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "integrity constraint; languages; learning;
                 performance; transformation heuristic",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf I.2.6}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Learning,
                 Knowledge acquisition.",
}

@Article{Kamel:1992:IDC,
  author =       "Nabil Kamel and Roger King",
  title =        "Intelligent Database Caching Through the Use of Page
                 Answers and Page Traces",
  journal =      j-TODS,
  volume =       "17",
  number =       "4",
  pages =        "601--646",
  month =        dec,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "1 197 199",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-4/p601-kamel/p601-kamel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-4/p601-kamel/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/146933.html",
  abstract =     "In this paper a new method to improve the utilization
                 of main memory systems is presented. The new method is
                 based on prestoring in main memory a number of query
                 answers, each evaluated out of a single memory page. To
                 this end, the ideas of page-answers and page-traces are
                 formally described and their properties analyzed. The
                 query model used here allows for selection, projection,
                 join, recursive queries as well as arbitrary
                 combinations. We also show how to apply the approach
                 under update traffic. This concept is especially useful
                 in managing the main memories of an important class of
                 applications. This class includes the evaluation of
                 triggers and alerters, performance improvement of
                 rule-based systems, integrity constraint checking, and
                 materialized views. These applications are
                 characterized by the existence at compile time of a
                 predetermined set of queries, by a slow but persistent
                 update traffic, and by their need to repetitively
                 reevaluate the query set. The new approach represents a
                 new type of intelligent database caching, which
                 contrasts with traditional caching primarily in that
                 the cache elements are derived data and as a
                 consequence, they overlap arbitrarily and do not have a
                 fixed length. The contents of the main memory cache are
                 selected based on the data distribution within the
                 database, the set of fixed queries to preprocess, and
                 the paging characteristics. Page-answers and
                 page-traces are used as the smallest indivisible units
                 in the cache. An efficient heuristic to select a near
                 optimal set of page-answers and page-traces to populate
                 the main memory has been developed, implemented, and
                 tested. Finally, quantitative measurements of
                 performance benefits are reported.",
  acknowledgement = ack-nhfb,
  affiliation =  "Florida Univ., Gainesville, FL, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; artificial intelligence; databases;
                 design; page access; performance",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval. {\bf
                 H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Access methods. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Query processing. {\bf H.3.1}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Content Analysis and
                 Indexing, Indexing methods. {\bf H.3.2}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Storage, Record classification. {\bf I.1.3}: Computing
                 Methodologies, ALGEBRAIC MANIPULATION, Languages and
                 Systems, Evaluation strategies. {\bf I.2.8}: Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Problem
                 Solving, Control Methods, and Search, Plan execution,
                 formation, generation.",
}

@Article{Maiocchi:1992:ADT,
  author =       "Roberto Maiocchi and Barbara Pernici and Federico
                 Barbic",
  title =        "Automatic Deduction of Temporal Information",
  journal =      j-TODS,
  volume =       "17",
  number =       "4",
  pages =        "647--688",
  month =        dec,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20 (03B70 68T27)",
  MRnumber =     "93h:68038",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-4/p647-maiocchi/p647-maiocchi.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-4/p647-maiocchi/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/146934.html",
  abstract =     "In many computer-based applications, temporal
                 information has to be stored, retrieved, and related to
                 other temporal information. Several time models have
                 been proposed to manage temporal knowledge in the
                 fields of conceptual modeling, database systems, and
                 artificial intelligence.\par

                 In this paper we present TSOS, a system for reasoning
                 about time that can be integrated as a time expert in
                 environments designed for broader problem-solving
                 domains. The main intended goal of TSOS is to allow a
                 user to infer further information on the temporal data
                 stored in the database through a set of deduction rules
                 handling various aspects of time. For this purpose,
                 TSOS provides the capability of answering queries about
                 the temporal specifications it has in its temporal
                 database.\par

                 Distinctive time-modeling features of TSOS are the
                 introduction of {\em temporal modalities}, i.e., the
                 possibility of specifying if a piece of information is
                 always true within a time interval, or if it is only
                 sometimes true, and the capability of answering about
                 the possibility and the necessity of the validity of
                 some information at a given time, the association of
                 temporal knowledge both to {\em instances of data\/}
                 and to {\em types of data}, and the development of a
                 {\em time calculus\/} for reasoning on temporal data.
                 Another relevant feature of TSOS is the capability to
                 reason about temporal data specified at different time
                 granularities.",
  acknowledgement = ack-nhfb,
  affiliation =  "Politecnico di Milano, Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; events; languages; meta-level temporal
                 assertions; propositions; temporal database; temporal
                 modalities; theory; time calculus",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf I.2.3}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem Proving,
                 Deduction.",
}

@Article{Agrawal:1992:GTQ,
  author =       "D. Agrawal and A. {El Abbadi}",
  title =        "The Generalized Tree Quorum Protocol: An Efficient
                 Approach for Managing Replicated Data",
  journal =      j-TODS,
  volume =       "17",
  number =       "4",
  pages =        "689--717",
  month =        dec,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68M10",
  MRnumber =     "1 197 201",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-4/p689-agrawal/p689-agrawal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-4/p689-agrawal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/146935.html",
  abstract =     "In this paper, we present a low-cost fault-tolerant
                 protocol for managing replicated data. We impose a
                 logical tree structure on the set of copies of an
                 object and develop a protocol that uses the information
                 available in the logical structure to reduce the
                 communication requirements for read and write
                 operations. The tree quorum protocol is a
                 generalization of the static voting protocol with two
                 degrees of freedom for choosing quorums. In general,
                 this results in significantly lower communication costs
                 for comparable data availability. The protocol exhibits
                 the property of graceful degradation, i.e.,
                 communication costs for executing operations are
                 minimal in a failure-free environment but may increase
                 as failures occur. This approach in designing
                 distributed systems is desirable since it provides
                 fault-tolerance without imposing unnecessary costs on
                 the failure-free mode of operations.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Santa Barbara, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Economics; Experimentation; Measurement;
                 Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; economics; experimentation; measurement;
                 performance",
  subject =      "{\bf C.2.2}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Network Protocols.
                 {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design. {\bf
                 G.2.2}: Mathematics of Computing, DISCRETE MATHEMATICS,
                 Graph Theory, Network problems. {\bf G.2.2}:
                 Mathematics of Computing, DISCRETE MATHEMATICS, Graph
                 Theory, Trees.",
}

@Article{Atzeni:1992:URD,
  author =       "Paolo Atzeni and Riccardo Torlone",
  title =        "Updating Relational Databases Through Weak Instance
                 Interfaces",
  journal =      j-TODS,
  volume =       "17",
  number =       "4",
  pages =        "718--745",
  month =        dec,
  year =         "1992",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P20",
  MRnumber =     "93h:68035",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1992-17-4/p718-atzeni/p718-atzeni.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1992-17-4/p718-atzeni/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/146936.html",
  abstract =     "The problem of updating databases through interfaces
                 based on the weak instance model is studied, thus
                 extending previous proposals that considered them only
                 from the query point of view. Insertions and deletions
                 of tuples are considered.\par

                 As a preliminary tool, a lattice on states is defined,
                 based on the information content of the various
                 states.\par

                 Potential results of an insertion are states that
                 contain at least the information in the original state
                 and that in the new tuple. Sometimes there is no
                 potential result, and in the other cases there may be
                 many of them. We argue that the insertion is
                 deterministic if the state that contains the
                 information common to all the potential results (the
                 greatest lower bound, in the lattice framework) is a
                 potential result itself. Effective characterizations
                 for the various cases exist.\par

                 A symmetric approach is followed for deletions, with
                 fewer cases, since there are always potential results;
                 determinism is characterized as a consequence.",
  acknowledgement = ack-nhfb,
  affiliation =  "Roma Univ., Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Schema
                 and subschema. {\bf H.2.m}: Information Systems,
                 DATABASE MANAGEMENT, Miscellaneous. {\bf H.5.2}:
                 Information Systems, INFORMATION INTERFACES AND
                 PRESENTATION, User Interfaces, Theory and methods.",
}

@Article{Ishikawa:1993:MLI,
  author =       "Hiroshi Ishikawa and Fumio Suzuki and Fumihiko
                 Kozakura and Akifumi Makinouchi and Mika Miyagishima
                 and Yoshio Izumida and Masaaki Aoshima and Yasuo
                 Yamane",
  title =        "The Model, Language, and Implementation of an
                 Object-Oriented Multimedia Knowledge Base Management
                 System",
  journal =      j-TODS,
  volume =       "18",
  number =       "1",
  pages =        "1--50",
  month =        mar,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-1/p1-ishikawa/p1-ishikawa.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-1/p1-ishikawa/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151285.html",
  abstract =     "New applications such as CAD, AI, and hypermedia
                 require direct representation and flexible use of
                 complex objects, behavioral knowledge, and multimedia
                 data. To this end, we have devised a knowledge base
                 management system called Jasmine. An object-oriented
                 approach in a programming language also seems promising
                 for use in Jasmine. Jasmine extends the current
                 object-oriented approach and provides the following
                 features. Our object model is based on functional data
                 models and well-established set theory. Attributes or
                 functions composing objects can represent both
                 structural and behavioral knowledge. The object model
                 can represent incomplete and generic knowledge. The
                 model can support the basic storage and operations of
                 multimedia data. The facets of attributes can flexibly
                 represent constraints and triggers. The object
                 manipulation language can support associative access of
                 objects. The structural and behavioral knowledge can be
                 uniformly treated to allow the user to specify complex
                 object operations in a compact manner. The user-defined
                 and system-defined attributes can be uniformly
                 specified to ease user customization of the language.
                 The classes and instances can be uniformly accessed.
                 Incomplete knowledge can be flexibly accessed. The
                 system has a layered architecture. Objects are stored
                 in nested relations provided by extensive DBMS as a
                 sublayer. User query of objects is compiled into
                 relational operations such as select and join, which
                 can be efficiently processed using hashing. The
                 behavioral knowledge is compiled into predicate and
                 manipulation function interfaces that can directly
                 access tuples in a buffer.",
  acknowledgement = ack-nhfb,
  affiliation =  "Fujitsu Labs., Ltd., Kawasaki, Japan",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages",
  subject =      "{\bf I.2.1}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Applications and Expert Systems. {\bf
                 I.2.4}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Knowledge Representation Formalisms and
                 Methods, Representations (procedural and rule-based).
                 {\bf H.4.2}: Information Systems, INFORMATION SYSTEMS
                 APPLICATIONS, Types of Systems. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Query languages. {\bf
                 H.5.1}: Information Systems, INFORMATION INTERFACES AND
                 PRESENTATION, Multimedia Information Systems.",
}

@Article{Johnson:1993:PCB,
  author =       "Theodore Johnson and Dennis Shasha",
  title =        "The Performance of Current {B-Tree} Algorithms",
  journal =      j-TODS,
  volume =       "18",
  number =       "1",
  pages =        "51--101",
  month =        mar,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "``Current'' in the title should be ``Concurrent''.",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-1/p51-johnson/p51-johnson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-1/p51-johnson/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151286.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Florida Univ., Gainesville, FL, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Measurement;
                 Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "B-trees; concurrent B-trees; concurrent data
                 structures; experimentation; measurement; performance;
                 performance of concurrent algorithms; TOC Concurrency
                 control simulations TODS, algorithms",
  subject =      "{\bf H.3.3}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Search and Retrieval,
                 Retrieval models. {\bf H.3.3}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Search
                 and Retrieval, Search process. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design. {\bf F.2.2}: Theory of
                 Computation, ANALYSIS OF ALGORITHMS AND PROBLEM
                 COMPLEXITY, Nonnumerical Algorithms and Problems,
                 Sorting and searching. {\bf I.6.6}: Computing
                 Methodologies, SIMULATION AND MODELING, Simulation
                 Output Analysis.",
}

@Article{Kumar:1993:CAT,
  author =       "Akhil Kumar and Arie Segev",
  title =        "Cost and Availability Tradeoffs in Replicated Data
                 Concurrency Control",
  journal =      j-TODS,
  volume =       "18",
  number =       "1",
  pages =        "102--131",
  month =        mar,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-1/p102-kumar/p102-kumar.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-1/p102-kumar/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151287.html",
  acknowledgement = ack-nhfb,
  affiliation =  "Cornell Univ., Ithaca, NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; availability; performance; replicated
                 database",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases.",
}

@Article{Abdel-Ghaffar:1993:ODA,
  author =       "Khaled A. S. Abdel-Ghaffar and Amr {El Abbadi}",
  title =        "Optimal Disk Allocation for Partial Match Queries",
  journal =      j-TODS,
  volume =       "18",
  number =       "1",
  pages =        "132--156",
  month =        mar,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-1/p132-abdel-ghaffar/p132-abdel-ghaffar.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-1/p132-abdel-ghaffar/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151288.html",
  abstract =     "The problem of disk allocation addresses the issue of
                 how to distribute a file on several disks in order to
                 maximize concurrent disk accesses in response to a
                 partial match query. In this paper a coding-theoretic
                 analysis of this problem is presented, and both
                 necessary and sufficient conditions for the existence
                 of strictly optimal allocation methods are provided.
                 Based on a class of optimal codes, known as maximum
                 distance separable codes, strictly optimal allocation
                 methods are constructed. Using the necessary conditions
                 proved, we argue that the standard definition of strict
                 optimality is too strong and cannot be attained, in
                 general. Hence, we reconsider the definition of
                 optimality. Instead of basing it on an abstract
                 definition that may not be attainable, we propose a new
                 definition based on the best possible allocation
                 method. Using coding theory, allocation methods that
                 are optimal according to our proposed criterion are
                 developed.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Davis, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Abbadi TODS, algorithms; Cartesian product files;
                 coding theory; design; multiple disk systems; partial
                 match queries; theory",
  subject =      "{\bf H.3.2}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Information Storage, File organization.
                 {\bf D.4.3}: Software, OPERATING SYSTEMS, File Systems
                 Management, File organization. {\bf E.5}: Data, FILES,
                 Organization/structure. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.1.1}: Information Systems, MODELS AND
                 PRINCIPLES, Systems and Information Theory, Information
                 theory. {\bf E.4}: Data, CODING AND INFORMATION
                 THEORY.",
}

@Article{Matsliach:1993:PAF,
  author =       "Gabriel Matsliach",
  title =        "Performance Analysis of File Organizations that Use
                 Multibucket Data Leaves with Partial Expansions",
  journal =      j-TODS,
  volume =       "18",
  number =       "1",
  pages =        "157--180",
  month =        mar,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-1/p157-matsliach/p157-matsliach.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-1/p157-matsliach/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151289.html",
  abstract =     "We present an exact performance analysis, under random
                 insertions, of file organizations that use multibucket
                 data leaves and perform partial expansions before
                 splitting. We evaluate the expected disk space
                 utilization of the file and show how the expected
                 search and insert costs can be estimated. The
                 analytical results are confirmed by simulations. The
                 analysis can be used to investigate both the dynamic
                 and the asymptotic behaviors.",
  acknowledgement = ack-nhfb,
  affiliation =  "Technion-Israel Inst. of Technol., Haifa, Israel",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bounded disorder files; multibucket data leaves;
                 partial expansion; performance; performance analysis;
                 search structures; theory; verification",
  subject =      "{\bf E.5}: Data, FILES, Organization/structure. {\bf
                 E.1}: Data, DATA STRUCTURES, Trees. {\bf H.3.2}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Storage, File organization. {\bf H.3.1}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Content Analysis and Indexing. {\bf H.3.3}: Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Information
                 Search and Retrieval.",
}

@Article{Chomicki:1993:FRI,
  author =       "Jan Chomicki and Tomasz Imieli{\'n}ski",
  title =        "Finite Representation of Infinite Query Answers",
  journal =      j-TODS,
  volume =       "18",
  number =       "2",
  pages =        "181--223",
  month =        jun,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-2/p181-chomicki/p181-chomicki.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-2/p181-chomicki/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151635.html",
  abstract =     "We define here a formal notion of finite
                 representation of infinite query answers in logic
                 programs. We apply this notion to Datalog$_{\rm nS}$
                 programs may be infinite and consequently queries may
                 have infinite answers.\par

                 We present a method to finitely represent infinite
                 least Herbrand models of Datalog$_{\rm nS}$ program
                 (and its underlying computational engine) can be
                 forgotten. Given a query to be evaluated, it is easy to
                 obtain from the relational specification finitely many
                 answer substitutions that represent infinitely many
                 answer substitutions to the query. The method involved
                 is a combination of a simple, unificationless,
                 computational mechanism (graph traversal, congruence
                 closure, or term rewriting) and standard relational
                 query evaluation methods. Second, a relational
                 specification is {\em effectively computable\/} and its
                 computation is no harder, in the sense of the
                 complexity class, than answering yes-no
                 queries.\par

                 Our method is applicable to every range-restricted
                 Datalog$_{\rm nS}$ program. We also show that for some
                 very simple non-Datalog$_{\rm nS}$ logic programs,
                 finite representations of query answers do not exist.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. and Inf. Sci., Kansas State Univ.,
                 Manhattan, KS, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; theory",
  subject =      "{\bf F.2.2}: Theory of Computation, ANALYSIS OF
                 ALGORITHMS AND PROBLEM COMPLEXITY, Nonnumerical
                 Algorithms and Problems. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Datalog. {\bf
                 I.2.3}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Deduction and Theorem Proving, Deduction.
                 {\bf F.4.1}: Theory of Computation, MATHEMATICAL LOGIC
                 AND FORMAL LANGUAGES, Mathematical Logic.",
}

@Article{Hou:1993:PTC,
  author =       "Wen-Chi Hou and Gultekin {\"O}zsoyo{\u{g}}lu",
  title =        "Processing Time-Constrained Aggregate Queries in
                 {CASE-DB}",
  journal =      j-TODS,
  volume =       "18",
  number =       "2",
  pages =        "224--261",
  month =        jun,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-2/p224-hou/p224-hou.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-2/p224-hou/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151636.html",
  abstract =     "In this paper, we present an algorithm to strictly
                 control the time to process an estimator for an
                 aggregate relational query. The algorithm implemented
                 in a prototype database management system, called
                 CASE-DB, iteratively samples from input relations, and
                 evaluates the associated estimator until the time quota
                 expires.\par

                 In order to estimate the time cost of a query, CASE-DB
                 uses adaptive time cost formulas. The formulas are
                 adaptive in that the parameters of the formulas can be
                 adjusted at runtime to better fit the characteristics
                 of a query. To control the use of time quota, CASE-DB
                 adopts the one-at-a-time-interval time control strategy
                 to make a tradeoff between the risks of overspending
                 and the overhead, finally, experimental evaluation of
                 the methodology is presented.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Southern Illinois Univ.,
                 Carbondale, IL, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; estimation; performance;
                 relational algebra; risk of overspending; sampling;
                 selectivity; theory; time constraints",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.3}: Mathematics of
                 Computing, PROBABILITY AND STATISTICS, Statistical
                 computing. {\bf H.2.8}: Information Systems, DATABASE
                 MANAGEMENT, Database applications. {\bf J.7}: Computer
                 Applications, COMPUTERS IN OTHER SYSTEMS, Real time.",
}

@Article{Drenick:1993:SQO,
  author =       "P. E. Drenick and E. J. Smith",
  title =        "Stochastic Query Optimization in Distributed
                 Databases",
  journal =      j-TODS,
  volume =       "18",
  number =       "2",
  pages =        "262--288",
  month =        jun,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-2/p262-drenick/p262-drenick.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-2/p262-drenick/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151637.html",
  abstract =     "Many algorithms have been devised for minimizing the
                 costs associated with obtaining the answer to a single,
                 isolated query in a distributed database system.
                 However, if more than one query may be processed by the
                 system at the same time and if the arrival times of the
                 queries are unknown, the determination of optimal
                 query-processing strategies becomes a stochastic
                 optimization problem. In order to cope with such
                 problems, a theoretical state-transition model is
                 presented that treats the system as one operating under
                 a stochastic load. Query-processing strategies may then
                 be distributed over the processors of a network as
                 probability distributions, in a manner which
                 accommodates many queries over time.\par

                 It is then shown that the model leads to the
                 determination of optimal query-processing strategies as
                 the solution of mathematical programming problems, and
                 analytical results for several examples are presented.
                 Furthermore, a divide-and-conquer approach is
                 introduced for decomposing stochastic query
                 optimization problems into distinct subproblems for
                 processing queries sequentially and in parallel.",
  acknowledgement = ack-nhfb,
  affiliation =  "Polytech. Univ., Farmingdale, NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; distributed query processing; performance;
                 state-transition model; stochastic query optimization;
                 theory",
  subject =      "{\bf G.1.6}: Mathematics of Computing, NUMERICAL
                 ANALYSIS, Optimization, Linear programming. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Retrieval
                 models. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Distributed systems. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Query processing. {\bf C.2.4}: Computer Systems
                 Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases.",
}

@Article{Rothermel:1993:OCP,
  author =       "Kurt Rothermel and Stefan Pappe",
  title =        "Open Commit Protocols Tolerating Commission Failures",
  journal =      j-TODS,
  volume =       "18",
  number =       "2",
  pages =        "289--332",
  month =        jun,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-2/p289-rothermel/p289-rothermel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-2/p289-rothermel/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151638.html",
  abstract =     "To ensure atomicity of transactions in distributed
                 systems so-called 2-phase commit (2PC) protocols have
                 been proposed. The basic assumption of these protocols
                 is that the processing nodes involved in transactions
                 are ``sane,'' i.e., they only fail with omission
                 failures, and nodes eventually recover from failures.
                 Unfortunately, this assumption is not realistic for
                 so-called Open Distributed Systems (ODSs), in which
                 nodes may have totally different reliability
                 characteristics. In ODSs, nodes can be classified into
                 trusted nodes (e.g., a banking server) and nontrusted
                 nodes (e.g., a home PC requesting a remote banking
                 service). While trusted nodes are assumed to be sane,
                 nontrusted nodes may fail permanently and even cause
                 commission failures to occur.\par

                 In this paper, we propose a family of 2PC protocols
                 that tolerate any number of omission failures at
                 trusted nodes and any number of commission and omission
                 failures at nontrusted nodes. The proposed protocols
                 ensure that (at least) the trusted nodes participating
                 in a transaction {\em eventually\/} terminate the
                 transaction in a {\em consistent\/} manner. Unlike
                 Byzantine commit protocols, our protocols do {\em
                 not\/} incorporate mechanisms for achieving Byzantine
                 agreement, which has advantages in terms of complexity:
                 Our protocols have the same or only a slightly higher
                 message complexity than traditional 2PC protocols.",
  acknowledgement = ack-nhfb,
  affiliation =  "Inst. of Parallel and Distributed High Performance
                 Syst., Stuttgart Univ., Germany",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; commit protocols; open systems;
                 performance; reliability",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Network operating systems. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf C.4}: Computer Systems Organization,
                 PERFORMANCE OF SYSTEMS, Reliability, availability, and
                 serviceability.",
}

@Article{Rahm:1993:EPE,
  author =       "Erhard Rahm",
  title =        "Empirical Performance Evaluation of Concurrency and
                 Coherency Control Protocols for Database Sharing
                 Systems",
  journal =      j-TODS,
  volume =       "18",
  number =       "2",
  pages =        "333--377",
  month =        jun,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; Database/Wiederhold.bib;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-2/p333-rahm/p333-rahm.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-2/p333-rahm/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/151639.html",
  abstract =     "Database Sharing (DB-sharing) refers to a general
                 approach for building a distributed high performance
                 transaction system. The nodes of a DB-sharing system
                 are locally coupled via a high-speed interconnect and
                 share a common database at the disk level. This is also
                 known as a ``shared disk'' approach. We compare
                 database sharing with the database partitioning (shared
                 nothing) approach and discuss the functional DBMS
                 components that require new and coordinated solutions
                 for DB-sharing. The performance of DB-sharing systems
                 critically depends on the protocols used for
                 concurrency and coherency control. The frequency of
                 communication required for these functions has to be
                 kept as low as possible in order to achieve high
                 transaction rates and short response times. A
                 trace-driven simulation system for DB-sharing complexes
                 has been developed that allows a realistic performance
                 comparison of four different concurrency and coherency
                 control protocols. We consider two locking and two
                 optimistic schemes which operate either under central
                 or distributed control. For coherency control, we
                 investigate so-called on-request and broadcast
                 invalidation schemes, and employ buffer-to-buffer
                 communication to exchange modified pages directly
                 between different nodes. The performance impact of
                 random routing versus affinity-based load distribution
                 and different communication costs is also examined. In
                 addition, we analyze potential performance bottlenecks
                 created by hot spot pages.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Kaiserslautern Univ., Germany",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; coherency control; concurrency control;
                 database partitioning; database sharing; design;
                 performance; performance analysis; shared disk; shared
                 nothing; theory; trace-driven simulation",
  subject =      "{\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf C.4}: Computer Systems Organization,
                 PERFORMANCE OF SYSTEMS. {\bf D.4.8}: Software,
                 OPERATING SYSTEMS, Performance, Simulation. {\bf
                 D.4.1}: Software, OPERATING SYSTEMS, Process
                 Management, Concurrency.",
}

@Article{Kuper:1993:LDM,
  author =       "Gabriel M. Kuper and Moshe Y. Vardi",
  title =        "The Logical Data Model",
  journal =      j-TODS,
  volume =       "18",
  number =       "3",
  pages =        "379--413",
  month =        sep,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "96c:68035",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-3/p379-kuper/p379-kuper.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-3/p379-kuper/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/155274.html",
  abstract =     "We propose an object-oriented data model that
                 generalizes the relational, hierarchical, and network
                 models. A database scheme in this model is a directed
                 graph, whose leaves represent data and whose internal
                 nodes represent connections among the data. Instances
                 are constructed from objects, which have separate names
                 and values. We define a logic for the model, and
                 describe a nonprocedural query language that is based
                 on the logic. We also describe an algebraic query
                 language and show that it is equivalent to the logical
                 language.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Thomas J. Watson Res. Center, Yorktown Heights,
                 NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Algebra; database schema; design; languages; logic;
                 relational database; theory; tuple calculus",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Schema and subschema.",
}

@Article{Levene:1993:SNE,
  author =       "Mark Levene and George Loizou",
  title =        "Semantics for Null Extended Nested Relations",
  journal =      j-TODS,
  volume =       "18",
  number =       "3",
  pages =        "414--459",
  month =        sep,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-3/p414-levene/p414-levene.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-3/p414-levene/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/155275.html",
  abstract =     "The nested relational model extends the flat
                 relational model by relaxing the first normal form
                 assumption in order to allow the modeling of complex
                 objects. Much of the previous work on the nested
                 relational model has concentrated on defining the data
                 structures and query language for the model. The work
                 done on integrity constraints in nested relations has
                 mainly focused on characterizing subclasses of nested
                 relations and defining normal forms for nested
                 relations with certain desirable properties.\par

                 In this paper we define the semantics of nested
                 relations, which may contain null values, in terms of
                 integrity constraints, called {\em null extended data
                 dependencies}, which extend functional dependencies and
                 join dependencies encountered in flat relational
                 database theory. We formalize incomplete information in
                 nested relations by allowing only one unmarked {\em
                 generic null value}, whose semantics we do not further
                 specify. The motivation for the choice of a generic
                 null is our desire to investigate only fundamental
                 semantics which are common to all unmarked null types.
                 This lead us to define a preorder on nested relations,
                 which allows us to measure the relative information
                 content of nested relations. We also define a
                 procedure, called the {\em extended chase procedure},
                 for testing satisfaction of null extended data
                 dependencies and for making inferences by using these
                 null extended data dependencies. The extended chase
                 procedure is shown to generalize the classical chase
                 procedure, which is of major importance in flat
                 relational database theory. As a consequence of our
                 approach we are able to capture the novel notion of
                 losslessness in nested relations, called herein {\em
                 null extended lossless decomposition}. Finally, we show
                 that the semantics of nested relations are a natural
                 extension of the semantics of flat relations.",
  acknowledgement = ack-nhfb,
  affiliation =  "Univ. Coll., London, UK",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; extended chase; languages; nested
                 relations; null extended algebra; null extended data
                 dependencies; nulls; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Normal forms.",
}

@Article{Agrawal:1993:COS,
  author =       "Divyakant Agrawal and Amr {El Abbadi} and Ambuj K.
                 Singh",
  title =        "Consistency and Orderability: Semantics-Based
                 Correctness Criteria for Databases",
  journal =      j-TODS,
  volume =       "18",
  number =       "3",
  pages =        "460--486",
  month =        sep,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-3/p460-agrawal/p460-agrawal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-3/p460-agrawal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/155276.html",
  abstract =     "The semantics of objects and transactions in database
                 systems are investigated. User-defined predicates
                 called {\em consistency assertions\/} are used to
                 specify user programs. Three new correctness criteria
                 are proposed. The first correctness criterion {\em
                 consistency\/} is based solely on the users'
                 specifications and admit nonserializable executions
                 that are acceptable to the users. Integrity constraints
                 of the database are maintained through consistency
                 assertions. The second correctness criterion {\em
                 orderability\/} is a generalization of view
                 serializability and represents a weak notion of
                 equivalence to a serial schedule. Finally, the third
                 correctness criterion {\em strong order-ability\/} is
                 introduced as a generalization of conflict
                 serializability. Unlike consistency, the notions of
                 orderability allow users to operate an isolation as
                 maintenance of the integrity constraint now becomes the
                 responsibility of the database system.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Santa Barbara, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Abbadi TODS, theory; concurrency control;
                 object-oriented databases; semantics; serializability
                 theory; verification",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Concurrency.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf D.2.4}: Software,
                 SOFTWARE ENGINEERING, Program Verification, Correctness
                 proofs. {\bf F.3.1}: Theory of Computation, LOGICS AND
                 MEANINGS OF PROGRAMS, Specifying and Verifying and
                 Reasoning about Programs, Assertions.",
}

@Article{Sagiv:1993:SQT,
  author =       "Yehoshua Sagiv and Oded Shmueli",
  title =        "Solving Queries by Tree Projections",
  journal =      j-TODS,
  volume =       "18",
  number =       "3",
  pages =        "487--511",
  month =        sep,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15",
  MRnumber =     "96c:68038",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-3/p487-sagiv/p487-sagiv.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-3/p487-sagiv/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/155277.html",
  abstract =     "Suppose a database schema {\bf D} is extended to {\bf
                 D'} by adding new relation schemas, and states for {\bf
                 D} are extended to states for {\bf D'} by applying
                 joins and projections to existing relations. It is
                 shown that certain desirable properties that {\bf D'}
                 has with respect to {\bf D}. These properties amount to
                 the ability to compute efficiently the join of all
                 relations in a state for {\bf D} from an extension of
                 this state over {\bf D'}. The equivalence is proved for
                 unrestricted (i.e., both finite and infinite)
                 databases. If {\bf D'} is obtained from {\bf D} by
                 adding a set of new relation schemas that form a tree
                 schema, then the equivalence also holds for finite
                 databases. In this case there is also a polynomial time
                 algorithm for testing the existence of a tree
                 projection of {\bf D'} with respect to {\bf D}.",
  acknowledgement = ack-nhfb,
  affiliation =  "Hebrew Univ., Jerusalem, Israel",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "acyclicity; algorithms; chase; database schema;
                 design; hypergraph; inclusion dependency; join;
                 monotone join expression; projection; qual graph;
                 relational database; semijoin; semijoin reduction;
                 tableau; theory; tree projection; tree schema",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Query processing.",
}

@Article{Ioannidis:1993:TCA,
  author =       "Yannis Ioannidis and Raghu Ramakrishnan and Linda
                 Winger",
  title =        "Transitive Closure Algorithms Based on Graph
                 Traversal",
  journal =      j-TODS,
  volume =       "18",
  number =       "3",
  pages =        "512--576",
  month =        sep,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-3/p512-ioannidis/p512-ioannidis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-3/p512-ioannidis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/155273.html",
  abstract =     "Several graph-based algorithms have been proposed in
                 the literature to compute the transitive closure of a
                 directed graph. We develop two new algorithms
                 (Basic\_TC and Gobal\_DFTC) and compare the performance
                 of their implementations in a disk-based environment
                 with a well-known graph-based algorithm proposed by
                 Schmitz. Our algorithms use depth-first search to
                 traverse a graph and a technique called {\em marking\/}
                 to avoid processing some of the arcs in the graph. They
                 compute the closure by processing nodes in reverse
                 topological order, building descendent sets by adding
                 the descendent sets of children. While the details of
                 these algorithms differ considerably, one important
                 difference among them is the time at which descendent
                 set additions are performed. Basic\_TC, results in
                 superior performance. The first reason is that early
                 additions result in larger descendent set sizes on the
                 average over the duration of the execution, thereby
                 causing more I/O; very often this turns out to more
                 than offset the gains of not having to fetch certain
                 sets again to add them. The second reason is that
                 information collected in the first pass can be used to
                 apply several optimizations in the second pass. To the
                 extent possible, we also adapt these algorithms to
                 perform path computations. Again, our performance
                 comparison confirms the trends seen in reachability
                 queries. Taken in conjunction with another performance
                 study our results indicate that all graph-based
                 algorithms significantly outperform other types of
                 algorithms such as Seminaive and Warren.",
  acknowledgement = ack-nhfb,
  affiliation =  "Wisconsin Univ., Madison, WI, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; depth-first search; node reachability;
                 path computations; performance; transitive closure",
  subject =      "{\bf D.3.4}: Software, PROGRAMMING LANGUAGES,
                 Processors. {\bf F.2.2}: Theory of Computation,
                 ANALYSIS OF ALGORITHMS AND PROBLEM COMPLEXITY,
                 Nonnumerical Algorithms and Problems, Computations on
                 discrete structures. {\bf D.3.3}: Software, PROGRAMMING
                 LANGUAGES, Language Constructs and Features, Recursion.
                 {\bf D.4.2}: Software, OPERATING SYSTEMS, Storage
                 Management, Main memory. {\bf D.4.2}: Software,
                 OPERATING SYSTEMS, Storage Management, Secondary
                 storage. {\bf D.4.2}: Software, OPERATING SYSTEMS,
                 Storage Management, Swapping. {\bf E.1}: Data, DATA
                 STRUCTURES, Graphs. {\bf E.1}: Data, DATA STRUCTURES,
                 Trees. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing.",
}

@Article{Thomasian:1993:TPL,
  author =       "Alexander Thomasian",
  title =        "Two-Phase Locking Performance and Its Thrashing
                 Behavior",
  journal =      j-TODS,
  volume =       "18",
  number =       "4",
  pages =        "579--625",
  month =        dec,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-4/p579-thomasian/p579-thomasian.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-4/p579-thomasian/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/169720.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; data contention; load control;
                 performance; theory; thrashing; two-phase locking",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Deadlock avoidance. {\bf D.4.8}:
                 Software, OPERATING SYSTEMS, Performance, Modeling and
                 prediction. {\bf D.4.8}: Software, OPERATING SYSTEMS,
                 Performance, Operational analysis. {\bf D.4.8}:
                 Software, OPERATING SYSTEMS, Performance, Simulation.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.",
}

@Article{Qian:1993:DSD,
  author =       "Xiaolei Qian",
  title =        "The Deductive Synthesis of Database Transactions",
  journal =      j-TODS,
  volume =       "18",
  number =       "4",
  pages =        "626--677",
  month =        dec,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-4/p626-qian/p626-qian.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-4/p626-qian/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/169716.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database programming; deductive tableau; design;
                 integrity constraints; search control; theory;
                 transaction logic; transaction synthesis;
                 verification",
  subject =      "{\bf I.2.2}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Automatic Programming, Program synthesis.
                 {\bf D.1.2}: Software, PROGRAMMING TECHNIQUES,
                 Automatic Programming. {\bf F.3.1}: Theory of
                 Computation, LOGICS AND MEANINGS OF PROGRAMS,
                 Specifying and Verifying and Reasoning about Programs,
                 Logics of programs. {\bf H.2.0}: Information Systems,
                 DATABASE MANAGEMENT, General, Security, integrity, and
                 protection. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Database (persistent)
                 programming languages. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing.",
}

@Article{Malvestuto:1993:USA,
  author =       "Francesco M. Malvestuto",
  title =        "A Universal-Scheme Approach to Statistical Databases
                 Containing Homogeneous Summary Tables",
  journal =      j-TODS,
  volume =       "18",
  number =       "4",
  pages =        "678--708",
  month =        dec,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-4/p678-malvestuto/p678-malvestuto.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-4/p678-malvestuto/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/169712.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bipartite graph; category relation; design;
                 query-answering system; statistical database; summary
                 table; theory; universal classification scheme",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.1.3}: Mathematics of
                 Computing, NUMERICAL ANALYSIS, Numerical Linear
                 Algebra, Linear systems (direct and iterative methods).
                 {\bf G.1.6}: Mathematics of Computing, NUMERICAL
                 ANALYSIS, Optimization, Integer programming. {\bf
                 G.1.6}: Mathematics of Computing, NUMERICAL ANALYSIS,
                 Optimization, Linear programming. {\bf G.2.2}:
                 Mathematics of Computing, DISCRETE MATHEMATICS, Graph
                 Theory, Graph algorithms. {\bf G.2.2}: Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Network
                 problems. {\bf G.2.2}: Mathematics of Computing,
                 DISCRETE MATHEMATICS, Graph Theory, Trees. {\bf H.2.1}:
                 Information Systems, DATABASE MANAGEMENT, Logical
                 Design, Data models. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Query languages.",
}

@Article{Ioannidis:1993:OHL,
  author =       "Yannis E. Ioannidis and Stavros Christodoulakis",
  title =        "Optimal Histograms for Limiting Worst-Case Error
                 Propagation in the Size of Join Results",
  journal =      j-TODS,
  volume =       "18",
  number =       "4",
  pages =        "709--748",
  month =        dec,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1993-18-4/p709-ioannidis/p709-ioannidis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1993-18-4/p709-ioannidis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/169708.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "histograms; join size estimation; performance; query
                 optimization; theory; vector majorization",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.1.0}: Mathematics of
                 Computing, NUMERICAL ANALYSIS, General, Error analysis.
                 {\bf H.1.1}: Information Systems, MODELS AND
                 PRINCIPLES, Systems and Information Theory.",
}

@Article{Anonymous:1993:AI,
  author =       "Anonymous",
  title =        "1993 Author Index",
  journal =      j-TODS,
  volume =       "18",
  number =       "4",
  pages =        "749--750",
  month =        dec,
  year =         "1993",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 12:59:37 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kim:1994:CS,
  author =       "Won Kim",
  title =        "Charter and Scope",
  journal =      j-TODS,
  volume =       "19",
  number =       "1",
  pages =        "1--??",
  month =        mar,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Dec 6 18:01:56 MST 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Olivier:1994:TSO,
  author =       "Martin S. Olivier and Sebastiaan H. {von Solms}",
  title =        "A Taxonomy for Secure Object-Oriented Databases",
  journal =      j-TODS,
  volume =       "19",
  number =       "1",
  pages =        "3--46",
  month =        mar,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib;
                 http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-1/p3-olivier/p3-olivier.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-1/p3-olivier/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/174640.html",
  abstract =     "This paper proposes a taxonomy for secure
                 object-oriented databases in order to clarify the
                 issues in modeling and implementing such databases. It
                 also indicates some implications of the various choices
                 one may make when designing such a database.\par

                 Most secure database models have been designed for
                 relational databases. The object-oriented database
                 model is more complex than the relational model. For
                 these reasons, models for secure object-oriented
                 databases are more complex than their relational
                 counterparts. Furthermore, since views of the
                 object-oriented model differ, each security model has
                 to make some assumptions about the object-oriented
                 model used for its particular database.\par

                 A number of models for secure object-oriented databases
                 have been proposed. These models differ in many
                 respects, because they focus on different aspects of
                 the security problem, or because they make different
                 assumptions about what constitutes a secure database or
                 because they make different assumptions about the
                 object-oriented model.\par

                 The taxonomy proposed in this paper may be used to
                 compare the various models: Models that focus on
                 specific issues may be positioned in the broader
                 context with the aid of the taxonomy. The taxonomy also
                 identifies the major aspects where security models may
                 differ and indicates some alternatives available to the
                 system designer for each such design choice. We show
                 some implications of using specific
                 alternatives.\par

                 Since differences between models for secure
                 object-oriented databases are often subtle, a formal
                 notation is necessary for a proper comparison. Such a
                 formal notation also facilitates the formal derivation
                 of restrictions that apply under specific conditions.
                 The formal approach also gives a clear indication about
                 the assumptions made by us---given as axioms---and the
                 consequences of those assumptions (and of design
                 choices made by the model designer)---given as
                 theorems.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; formal security models; information security;
                 multilevel secure databases; object-orientation;
                 security",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 D.4.6}: Software, OPERATING SYSTEMS, Security and
                 Protection. {\bf K.6.5}: Computing Milieux, MANAGEMENT
                 OF COMPUTING AND INFORMATION SYSTEMS, Security and
                 Protection. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design.",
}

@Article{Tendick:1994:MRP,
  author =       "Patrick Tendick and Norman Matloff",
  title =        "A Modified Random Perturbation Method for Database
                 Security",
  journal =      j-TODS,
  volume =       "19",
  number =       "1",
  pages =        "47--63",
  month =        mar,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-1/p47-tendick/p47-tendick.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-1/p47-tendick/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/174641.html",
  abstract =     "The random data perturbation (RDP) method of
                 preserving the privacy of individual records in a
                 statistical database is discussed. In particular, it is
                 shown that if confidential attributes are allowed as
                 query-defining variables, severe biases may result in
                 responses to queries. It is also shown that even if
                 query definition through confidential variables is {\em
                 not\/} allowed, biases can still occur in responses to
                 queries such as those involving proportions or counts.
                 In either case, serious distortions may occur in user
                 statistical analyses. A modified version of RDP is
                 presented, in the form of a query adjustment procedure
                 and specialized perturbation structure which will
                 produce unbiased results.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bias; correlation; design; noise addition; random
                 perturbation method; security",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 D.4.6}: Software, OPERATING SYSTEMS, Security and
                 Protection, Access controls. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing.",
}

@Article{Clifford:1994:CHR,
  author =       "James Clifford and Albert Croker and Alexander
                 Tuzhilin",
  title =        "On Completeness of Historical Relational Query
                 Languages",
  journal =      j-TODS,
  volume =       "19",
  number =       "1",
  pages =        "64--116",
  month =        mar,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-1/p64-clifford/p64-clifford.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-1/p64-clifford/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/174642.html",
  abstract =     "Numerous proposals for extending the relational data
                 model to incorporate the temporal dimension of data
                 have appeared in the past several years. These
                 proposals have differed considerably in the way that
                 the temporal dimension has been incorporated both into
                 the {\em structure\/} of the extended relations of
                 these temporal models and into the extended relational
                 {\em algebra\/} or {\em calculus\/} that they define.
                 Because of these differences, it has been difficult to
                 compare the proposed models and to make judgments as to
                 which of them might in some sense be equivalent or even
                 {\em better}. In this paper we define {\em temporally
                 grouped\/} and {\em temporally ungrouped\/} historical
                 data models and propose two notions of {\em historical
                 relational completeness}, analogous to Codd's notion of
                 relational completeness, one for each type of model. We
                 show that the temporally ungrouped models are less
                 expressive than the grouped models, but demonstrate a
                 technique for extending the ungrouped models with a
                 grouping mechanism to capture the additional semantic
                 power of temporal grouping. For the ungrouped models,
                 we define three different languages, a logic with
                 explicit reference to time, a temporal logic, and a
                 temporal algebra, and motivate our choice for the first
                 of these as the basis for completeness for these
                 models. For the grouped models, we define a many-sorted
                 logic with variables over ordinary values, historical
                 values, and times. Finally, we demonstrate the
                 equivalence of this grouped calculus and the ungrouped
                 calculus extended with a grouping mechanism. We believe
                 the classification of historical data models into
                 grouped and ungrouped models provides a useful
                 framework for the comparison of models in the
                 literature, and furthermore, the exposition of
                 equivalent languages for each type provides reasonable
                 standards for common, and minimal, notions of
                 historical relational completeness.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "completeness; historical databases; languages; query
                 languages; relational model; temporal databases;
                 temporal grouping; temporal logic; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models.",
}

@Article{Salem:1994:AL,
  author =       "Kenneth Salem and H{\'e}ctor Garc{\'\i}a-Molina and
                 Jeannie Shands",
  title =        "Altruistic Locking",
  journal =      j-TODS,
  volume =       "19",
  number =       "1",
  pages =        "117--165",
  month =        mar,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "Database/Graefe.bib; http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-1/p117-salem/p117-salem.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-1/p117-salem/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/174639.html",
  abstract =     "Long-lived transactions (LLTs) hold on to database
                 resources for relatively long periods of time,
                 significantly delaying the completion of shorter and
                 more common transactions. To alleviate this problem we
                 propose an extension to two-phase locking, called
                 altruistic locking, whereby LLTs can release their
                 locks early. Transactions that access this released
                 data are said to run in the wake of the LLT and must
                 follow special locking rules. Like two-phase locking,
                 altruistic locking is easy to implement and guarantees
                 serializability.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; atomicity; locking; performance;
                 scheduling; serializability; theory, concurrency
                 control ``wake'' of a single transaction TODS",
  subject =      "{\bf H.2.2}: Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Deadlock avoidance. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Transaction processing.",
}

@Article{Rosenthal:1994:TTR,
  author =       "Arnon Rosenthal and David Reiner",
  title =        "Tools and Transformations --- Rigorous and Otherwise
                 --- for Practical Database Design",
  journal =      j-TODS,
  volume =       "19",
  number =       "2",
  pages =        "167--211",
  month =        jun,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-2/p167-rosenthal/p167-rosenthal.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-2/p167-rosenthal/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/176568.html",
  abstract =     "We describe the tools and theory of a comprehensive
                 system for database design, and show how they work
                 together to support multiple conceptual and logical
                 design processes. The Database Design and Evaluation
                 Workbench (DDEW) system uses a rigorous,
                 information-content-preserving approach to schema
                 transformation, but combines it with heuristics, guess
                 work, and user interactions. The main contribution lies
                 in illustrating how theory was adapted to a practical
                 system, and how the consistency and power of a design
                 system can be increased by use of theory.\par

                 First, we explain why a design system needs multiple
                 data models, and how implementation over a unified
                 underlying model reduces redundancy and inconsistency.
                 Second, we present a core set of small but fundamental
                 algorithms that rearrange a schema without changing its
                 information content. From these reusable components, we
                 easily built larger tools and transformations that were
                 still formally justified. Third, we describe heuristic
                 tools that attempt to improve a schema, often by adding
                 missing information. In these tools, unreliable
                 techniques such as normalization and relationship
                 inference are bolstered by system-guided user
                 interactions to remove errors. We present a rigorous
                 criterion for identifying unnecessary relationships,
                 and discuss an interactive view integrator. Last, we
                 examine the relevance of database theory to building
                 these practically motivated tools and contrast the
                 paradigms of system builders with those of
                 theoreticians.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "applications of database theory; computer-aided
                 software engineering; data model translation; database
                 design; database equivalence; design; design
                 heuristics; entity-relationship model; heuristics;
                 normalization; theory; view integration",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf D.2.2}:
                 Software, SOFTWARE ENGINEERING, Tools and Techniques,
                 Programmer workbench. {\bf D.2.2}: Software, SOFTWARE
                 ENGINEERING, Tools and Techniques, Software libraries.
                 {\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Normal
                 forms.",
}

@Article{Bright:1994:ARS,
  author =       "M. W. Bright and A. R. Hurson and S. Pakzad",
  title =        "Automated Resolution of Semantic Heterogeneity in
                 Multidatabases",
  journal =      j-TODS,
  volume =       "19",
  number =       "2",
  pages =        "212--253",
  month =        jun,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-2/p212-bright/p212-bright.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-2/p212-bright/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/176569.html",
  abstract =     "A multidatabase system provides integrated access to
                 heterogeneous, autonomous local databases in a
                 distributed system. An important problem in current
                 multidatabase systems is identification of semantically
                 similar data in different local databases. The Summary
                 Schemas Model (SSM) is proposed as an extension to
                 multidatabase systems to aid in semantic
                 identification. The SSM uses a global data structure to
                 abstract the information available in a multidatabase
                 system. This abstracted form allows users to use their
                 own terms (imprecise queries) when accessing data
                 rather than being forced to use system-specified terms.
                 The system uses the global data structure to match the
                 user's terms to the semantically closest available
                 system terms. A simulation of the SSM is presented to
                 compare imprecise-query processing with corresponding
                 query-processing costs in a standard multidatabase
                 system. The costs and benefits of the SSM are
                 discussed, and future research directions are
                 presented.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; federated database; imprecise queries;
                 multidatabase; performance; schemas; semantic
                 heterogeneity; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 H.2.5}: Information Systems, DATABASE MANAGEMENT,
                 Heterogeneous Databases.",
}

@Article{Sciore:1994:USV,
  author =       "Edward Sciore and Michael Siegel and Arnon Rosenthal",
  title =        "Using Semantic Values to Facilitate Interoperability
                 Among Heterogeneous Information Systems",
  journal =      j-TODS,
  volume =       "19",
  number =       "2",
  pages =        "254--290",
  month =        jun,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-2/p254-sciore/p254-sciore.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-2/p254-sciore/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/176570.html",
  abstract =     "Large organizations need to exchange information among
                 many separately developed systems. In order for this
                 exchange to be useful, the individual systems must
                 agree on the meaning of their exchanged data. That is,
                 the organization must ensure {\em semantic
                 interoperability}. This paper provides a theory of {\em
                 semantic values\/} as a unit of exchange that
                 facilitates semantic interoperability between
                 heterogeneous information systems. We show how semantic
                 values can either be stored explicitly or be defined by
                 {\em environments}. A system architecture is presented
                 that allows autonomous components to share semantic
                 values. The key component in this architecture is
                 called the {\em context mediator}, whose job is to
                 identify and construct the semantic values being sent,
                 to determine when the exchange is meaningful, and to
                 convert the semantic values to the form required by the
                 receiver.\par

                 Our theory is then applied to the relational model. We
                 provide an interpretation of standard SQL queries in
                 which context conversions and manipulations are
                 transparent to the user. We also introduce an extension
                 of SQL, called Context-SQL (C-SQL), in which the {\em
                 context\/} of a semantic value can be explicitly
                 accessed and updated. Finally, we describe the
                 implementation of a prototype context mediator for a
                 relational C-SQL system.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Management",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; languages; management",
  subject =      "{\bf H.2.5}: Information Systems, DATABASE MANAGEMENT,
                 Heterogeneous Databases, Data translation. {\bf H.2.3}:
                 Information Systems, DATABASE MANAGEMENT, Languages,
                 Query languages. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Distributed systems.",
}

@Article{Subrahmanian:1994:AKB,
  author =       "V. S. Subrahmanian",
  title =        "Amalgamating Knowledge Bases",
  journal =      j-TODS,
  volume =       "19",
  number =       "2",
  pages =        "291--331",
  month =        jun,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-2/p291-subrahmanian/p291-subrahmanian.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-2/p291-subrahmanian/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/176571.html",
  abstract =     "The integration of knowledge for multiple sources is
                 an important aspect of automated reasoning systems.
                 When different knowledge bases are used to store
                 knowledge provided by multiple sources, we are faced
                 with the problem of integrating multiple knowledge
                 bases: Under these circumstances, we are also
                 confronted with the prospect of inconsistency. In this
                 paper we present a uniform theoretical framework, based
                 on annotated logics, for amalgamating multiple
                 knowledge bases when these knowledge bases (possibly)
                 contain inconsistencies, uncertainties, and
                 nonmonotonic modes of negation. We show that annotated
                 logics may be used, with some modifications, to mediate
                 between different knowledge bases. The multiple
                 knowledge bases are amalgamated by a transformation of
                 the individual knowledge bases into new annotated logic
                 programs, together with the addition of a new axiom
                 scheme. We characterize the declarative semantics of
                 such amalgamated knowledge bases and study how the
                 semantics of the amalgam is related to the semantics of
                 the individual knowledge bases being combined.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "amalgamated knowledge bases; annotated logics;
                 languages",
  subject =      "{\bf I.2.4}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Knowledge Representation Formalisms and
                 Methods, Representations (procedural and rule-based).
                 {\bf H.2.5}: Information Systems, DATABASE MANAGEMENT,
                 Heterogeneous Databases, Data translation. {\bf I.2.4}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Knowledge Representation Formalisms and Methods,
                 Representation languages.",
}

@Article{Yan:1994:ISS,
  author =       "Tak W. Yan and H{\'e}ctor Garc{\'\i}a-Molina",
  title =        "Index Structures for Selective Dissemination of
                 Information Under the {Boolean} Model",
  journal =      j-TODS,
  volume =       "19",
  number =       "2",
  pages =        "332--364",
  month =        jun,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-2/p332-yan/p332-yan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-2/p332-yan/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/176573.html",
  abstract =     "The number, size, and user population of bibliographic
                 and full-text document databases are rapidly growing.
                 With a high document arrival rate, it becomes essential
                 for users of such databases to have access to the very
                 latest documents; yet the high document arrival rate
                 also makes it difficult for users to keep themselves
                 updated. It is desirable to allow users to submit
                 profiles, i.e., queries that are constantly evaluated,
                 so that they will be automatically informed of new
                 additions that may be of interest. Such service is
                 traditionally called Selective Dissemination of
                 Information (SDI).\par

                 The high document arrival rate, the huge number of
                 users, and the timeliness requirement of the service
                 pose a challenge in achieving efficient SDL. In this
                 article, we propose several index structures for
                 indexing profiles and algorithms that efficiently match
                 documents against large number of profiles. We also
                 present analysis and simulation results to compare
                 their performance under different scenarios.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; performance",
  subject =      "{\bf H.3.1}: Information Systems, INFORMATION STORAGE
                 AND RETRIEVAL, Content Analysis and Indexing, Indexing
                 methods. {\bf H.2.2}: Information Systems, DATABASE
                 MANAGEMENT, Physical Design, Access methods. {\bf
                 H.3.4}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Systems and Software, Current awareness
                 systems (selective dissemination of information ---
                 SDI).",
}

@Article{Ceri:1994:AGP,
  author =       "Stefano Ceri and Piero Fraternali and Stefano
                 Paraboschi and Letizia Tanca",
  title =        "Automatic Generation of Production Rules for Integrity
                 Maintenance",
  journal =      j-TODS,
  volume =       "19",
  number =       "3",
  pages =        "367--422",
  month =        sep,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-3/p367-ceri/p367-ceri.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-3/p367-ceri/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/185828.html",
  abstract =     "In this article we present an approach to integrity
                 maintenance, consisting of automatically generating
                 production rules for integrity enforcement. Constraints
                 are expressed as particular formulas of Domain
                 Relational Calculus; they are automatically translated
                 into a set of repair actions, encoded as production
                 rules of an active database system. Production rules
                 may be redundant (they enforce the same constraint in
                 different ways) and conflicting (because repairing one
                 constraint may cause the violation of another
                 constraint). Thus, it is necessary to develop
                 techniques for analyzing the properties of the set of
                 active rules and for ensuring that any computation of
                 production rules after any incorrect transaction
                 terminates and produces a consistent database
                 state.\par

                 Along these guidelines, we describe a specific
                 architecture for constraint definition and enforcement.
                 The components of the architecture include a {\em Rule
                 Generator}, for producing all possible repair actions,
                 and a {\em Rule Analyzer and Selector}, for producing a
                 collection of production rules such that their
                 execution after an incorrect transaction always
                 terminates in a consistent state (possibly by rolling
                 back the transaction); moreover, the needs of
                 applications are modeled, so that integrity-enforcing
                 rules reach the final state that better represents the
                 original intentions of the transaction's supplier.
                 Specific input from the designer can also drive the
                 process and integrate or modify the rules generated
                 automatically by the method. Experimental results of a
                 prototype implementation of the proposed architecture
                 are also described.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dipartimento di Elettronica, Politecnico di Milano,
                 Italy",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Experimentation; Management",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "automatic generation of production rules;
                 experimentation; management",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 I.2.2}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Automatic Programming.",
}

@Article{Polyzois:1994:ERB,
  author =       "Christos A. Polyzois and H{\'e}ctor
                 Garc{\'\i}a-Molina",
  title =        "Evaluation of Remote Backup Algorithms for
                 Transaction-Processing Systems",
  journal =      j-TODS,
  volume =       "19",
  number =       "3",
  pages =        "423--449",
  month =        sep,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-3/p423-polyzois/p423-polyzois.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-3/p423-polyzois/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/185836.html",
  abstract =     "A remote backup is a copy of a primary database
                 maintained at a geographically separate location and is
                 used to increase data availability. Remote backup
                 systems are typically log-based and can be classified
                 into 2-safe and 1-safe, depending on whether
                 transactions commit at both sites simultaneously or
                 first commit at the primary and are later propagated to
                 the backup. We have built an experimental database
                 system on which we evaluated the performance of the
                 epoch and the dependency reconstruction algorithms, two
                 1-safe algorithms we have developed. We compared the
                 1-safe with the 2-safe approach under various
                 conditions.",
  acknowledgement = ack-nhfb,
  affiliation =  "IBM Thomas J. Watson Res. Center, Yorktown Heights,
                 NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; disaster recovery; hot spare; hot standby;
                 performance; reliability; remote backup",
  subject =      "{\bf H.2.7}: Information Systems, DATABASE MANAGEMENT,
                 Database Administration, Logging and recovery. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed applications. {\bf D.4.5}: Software,
                 OPERATING SYSTEMS, Reliability, Backup procedures. {\bf
                 D.4.5}: Software, OPERATING SYSTEMS, Reliability,
                 Fault-tolerance.",
}

@Article{Chrysanthis:1994:SET,
  author =       "Panos K. Chrysanthis and Krithi Ramamritham",
  title =        "Synthesis of Extended Transaction Models Using
                 {ACTA}",
  journal =      j-TODS,
  volume =       "19",
  number =       "3",
  pages =        "450--491",
  month =        sep,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-3/p450-chrysanthis/p450-chrysanthis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-3/p450-chrysanthis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/185843.html",
  abstract =     "ACTA is a comprehensive transaction framework that
                 facilitates the formal description of properties of
                 extended transaction models. Specifically, using ACTA,
                 one can specify and reason about (1) the effects of
                 transactions on objects and (2) the interactions
                 between transactions. This article presents ACTA {\em
                 as a tool for the synthesis of extended transaction
                 models}, one which supports the development and
                 analysis of new extended transaction models in a
                 systematic manner. Here, this is demonstrated by
                 deriving new transaction definitions (1) by modifying
                 the specifications of existing transaction models, (2)
                 by combining the specifications of existing models, and
                 (3) by starting from first principles. To exemplify the
                 first, new models are synthesized from {\em atomic
                 transactions\/} and {\em join transactions}. To
                 illustrate the second, we synthesize a model that
                 combines aspect of the {\em nested\/}- and {\em
                 split-transaction\/} models. We demonstrate the latter
                 by deriving the specification of an {\em
                 open-nested-transaction\/} model from high-level
                 requirements.",
  acknowledgement = ack-nhfb,
  affiliation =  "Pittsburgh Univ., PA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Reliability; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; correctness criteria; design;
                 reliability; semantics; serializability theory; theory;
                 transaction models; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed databases. {\bf
                 D.2.4}: Software, SOFTWARE ENGINEERING, Program
                 Verification, Correctness proofs. {\bf D.3.3}:
                 Software, PROGRAMMING LANGUAGES, Language Constructs
                 and Features, Abstract data types. {\bf D.4.1}:
                 Software, OPERATING SYSTEMS, Process Management,
                 Concurrency. {\bf F.3.1}: Theory of Computation, LOGICS
                 AND MEANINGS OF PROGRAMS, Specifying and Verifying and
                 Reasoning about Programs, Assertions. {\bf H.2.4}:
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Distributed systems.",
}

@Article{Korth:1994:FAC,
  author =       "Henry F. Korth and Greg Speegle",
  title =        "Formal Aspects of Concurrency Control in Long-Duration
                 Transaction Systems Using the {NT\slash PV} model",
  journal =      j-TODS,
  volume =       "19",
  number =       "3",
  pages =        "492--535",
  month =        sep,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-3/p492-korth/p492-korth.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-3/p492-korth/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/185854.html",
  abstract =     "In the typical database system, an execution is
                 correct if it is equivalent to some serial execution.
                 This criterion, called serializability, is unacceptable
                 for new database applications which require
                 long-duration transactions. We present a new
                 transaction model which allows correctness criteria
                 more suitable for these applications. This model
                 combines three enhancements to the standard model:
                 nested transactions, explicit predicates, and multiple
                 versions. These features yield the name of the new
                 model, nested transactions with predicates and
                 versions, or NT/PV.\par

                 The modular nature of the NT/PV model allows a
                 straightforward representation of simple systems. It
                 also provides a formal framework for describing complex
                 interactions. The most complex interactions the model
                 allows can be captured by a protocol which exploits all
                 of the semantics available to the NT/PV model. An
                 example of these interactions is shown in a CASE
                 application. The example shows how a system based on
                 the NT/PV model is superior to both standard database
                 techniques and unrestricted systems in both correctness
                 and performance.",
  acknowledgement = ack-nhfb,
  affiliation =  "Matsushita Inf. Technol. Lab., Panasonic Technol.
                 Inc., Princeton, NJ, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control protocol; semantic information;
                 theory; transaction processing",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Transaction processing.",
}

@Article{Goldman:1994:QCN,
  author =       "Kenneth J. Goldman and Nancy Lynch",
  title =        "Quorum Consensus in Nested-Transaction Systems",
  journal =      j-TODS,
  volume =       "19",
  number =       "4",
  pages =        "537--585",
  month =        dec,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-4/p537-goldman/p537-goldman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-4/p537-goldman/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/195666.html",
  abstract =     "Gifford's Quorum Consensus algorithm for data
                 replication is studied in the context of nested
                 transactions and transaction failures (aborts), and a
                 fully developed reconfiguration strategy is presented.
                 A formal description of the algorithm is presented
                 using the Input/Output automaton model for
                 nested-transaction systems due to Lynch and Merritt. In
                 this description, the algorithm itself is described in
                 terms of nested transactions. The formal description is
                 used to construct a complete proof of correctness that
                 uses standard assertional techniques, is based on a
                 natural correctness condition, and takes advantage of
                 modularity that arises from describing the algorithm as
                 nested transactions. The proof is accomplished
                 hierarchically, showing that a fully replicated
                 reconfigurable system ``simulates'' an intermediate
                 replicated system, and that the intermediate system
                 simulates an unreplicated system. The presentation and
                 proof treat issues of data replication entirely
                 separately from issues of concurrency control and
                 recovery.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Washington Univ., St. Louis,
                 MO, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; concurrency control; data replication;
                 hierarchical proofs; I/O automata; nested transactions;
                 quorum consensus; theory; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4}: Information Systems,
                 DATABASE MANAGEMENT, Systems, Distributed systems.",
}

@Article{Krishnakumar:1994:BIT,
  author =       "Narayanan Krishnakumar and Arthur J. Bernstein",
  title =        "Bounded Ignorance: a Technique for Increasing
                 Concurrency in a Replicated System",
  journal =      j-TODS,
  volume =       "19",
  number =       "4",
  pages =        "586--625",
  month =        dec,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-4/p586-krishnakumar/p586-krishnakumar.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-4/p586-krishnakumar/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/195670.html",
  abstract =     "Databases are replicated to improve performance and
                 availability. The notion of correctness that has
                 commonly been adopted for concurrent access by
                 transactions to shared, possibly replicated, data is
                 serializability. However, serializability may be
                 impractical in high-performance applications since it
                 imposes too stringent a restriction on concurrency.
                 When serializability is relaxed, the integrity
                 constraints describing the data may be violated. By
                 allowing bounded violations of the integrity
                 constraints, however, we are able to increase the
                 concurrency of transactions that execute in a
                 replicated environment. In this article, we introduce
                 the notion of an {\em N-ignorant\/} transaction, which
                 is a transaction that may be ignorant of the results of
                 at most $N$ prior transactions, which is a transaction
                 that may be ignorant of the results of at most $N$
                 prior transactions. A system in which all transactions
                 are {\em N-ignorant\/} can have an $N$ + 1-fold
                 increase in concurrency over serializable systems, at
                 the expense of bounded violations of its integrity
                 constraints. We present algorithms for implementing
                 replicated databases in {\em N-ignorant\/} systems. We
                 then provide constructive methods for calculating the
                 reachable states in such systems, given the value of
                 $N$, so that one may assess the maximum liability that
                 is incurred in allowing constraint violation. Finally,
                 we generalize the notion of {\em N-ignorance\/} to a
                 matrix of ignorance for the purpose of higher
                 concurrency.",
  acknowledgement = ack-nhfb,
  affiliation =  "State Univ. of New York, Stony Brook, NY, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; concurrency control; integrity
                 constraints; performance; reachability analysis;
                 replication; serializability; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf C.2.4}: Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed applications. {\bf
                 C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Concurrency.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed systems.",
}

@Article{Winslett:1994:FQL,
  author =       "Marianne Winslett and Kenneth Smith and Xiaolei Qian",
  title =        "Formal Query Languages for Secure Relational
                 Databases",
  journal =      j-TODS,
  volume =       "19",
  number =       "4",
  pages =        "626--662",
  month =        dec,
  year =         "1994",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1994-19-4/p626-winslett/p626-winslett.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1994-19-4/p626-winslett/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/195675.html",
  abstract =     "The addition of stringent security specifications to
                 the list of requirements for an application poses many
                 new problems in DBMS design and implementation, as well
                 as database design, use, and maintenance. Tight
                 security requirements, such as those that result in
                 silent masking of withholding of true information from
                 a user or the introduction of false information into
                 query answers, also raise fundamental questions about
                 the meaning of the database and the semantics of
                 accompanying query languages. In this paper, we propose
                 a belief-based semantics for secure databases, which
                 provides a semantics for databases that can ``lie''
                 about the state of the world, or about their knowledge
                 about the state of the world, in order to preserve
                 security. This kind of semantics can be used as a
                 helpful retrofit for the proposals for a ``multilevel
                 secure'' database model (a particularly stringent form
                 of security), and may be useful for less restrictive
                 security policies as well. We also propose a family of
                 query languages for multilevel secure relational
                 database applications, and base the semantics of those
                 languages on our semantics for secure databases. Our
                 query languages are free of the semantic problems
                 associated with use of ordinary SQL in a multilevel
                 secure context, and should be easy for users to
                 understand and employ.",
  acknowledgement = ack-nhfb,
  affiliation =  "Dept. of Comput. Sci., Illinois Univ., Urbana, IL,
                 USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "formal security models; information security;
                 multilevel secure databases; security",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 K.6.5}: Computing Milieux, MANAGEMENT OF COMPUTING AND
                 INFORMATION SYSTEMS, Security and Protection. {\bf
                 H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages.",
}

@Article{Kim:1995:CS,
  author =       "Won Kim",
  title =        "Charter and scope",
  journal =      j-TODS,
  volume =       "20",
  number =       "1",
  pages =        "1--2",
  month =        mar,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 10 13:00:12 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Aiken:1995:SAT,
  author =       "Alexander Aiken and Joseph M. Hellerstein and Jennifer
                 Widom",
  title =        "Static Analysis Techniques for Predicting the Behavior
                 of Active Database Rules",
  journal =      j-TODS,
  volume =       "20",
  number =       "1",
  pages =        "3--41",
  month =        mar,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-1/p3-aiken/p3-aiken.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-1/p3-aiken/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/202107.html",
  abstract =     "This article gives methods for statically analyzing
                 sets of active database rules to determine if the rules
                 are (1) guaranteed to terminate, (2) guaranteed to
                 produce a unique final database state, and (3)
                 guaranteed to produce a unique stream of observable
                 actions. If the analysis determines that one of these
                 properties is not guaranteed, it isolates the rules
                 responsible for the problem and determines criteria
                 that, if satisfied, guarantee the property. The
                 analysis methods are presented in the context of the
                 {\em Starburst Rule System}.",
  acknowledgement = ack-nhfb,
  affiliation =  "California Univ., Berkeley, CA, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Management; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active database systems; algorithms; confluence;
                 database rule processing; design; management; static
                 analysis; termination; verification",
  subject =      "{\bf H.2.m}: Information Systems, DATABASE MANAGEMENT,
                 Miscellaneous. {\bf D.2.4}: Software, SOFTWARE
                 ENGINEERING, Program Verification, Validation. {\bf
                 I.2.5}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Programming Languages and Software.",
}

@Article{Chen:1995:DUR,
  author =       "Weidong Chen",
  title =        "Declarative Updates of Relational Databases",
  journal =      j-TODS,
  volume =       "20",
  number =       "1",
  pages =        "42--70",
  month =        mar,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-1/p42-chen/p42-chen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-1/p42-chen/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/202110.html",
  abstract =     "This article presents a declarative language, called
                 {\em update calculus}, of relational database updates.
                 A formula in update calculus involves conditions for
                 the current database, as well as assertions about a new
                 database. Logical connectives and quantifiers become
                 constructors of complex updates, offering flexible
                 specifications of database transformations. Update
                 calculus can express all nondeterministic database
                 transformations that are polynomial time.\par

                 For set-at-a-time evaluation of updates, we present a
                 corresponding {\em update algebra}. Existing techniques
                 of query processing can be incorporated into update
                 evaluation. We show that updates in update calculus can
                 be translated into expressions in update algebra and
                 vice versa.",
  acknowledgement = ack-nhfb,
  affiliation =  "Southern Methodist Univ., Dallas, TX, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database languages; database updates; expressive
                 power; languages; theory; update algebra; update
                 calculus",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Data manipulation languages (DML).",
}

@Article{Jagadish:1995:IDM,
  author =       "H. V. Jagadish",
  title =        "The {INCINERATE} data model",
  journal =      j-TODS,
  volume =       "20",
  number =       "1",
  pages =        "71--110",
  month =        mar,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-1/p71-jagadish/p71-jagadish.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-1/p71-jagadish/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/202113.html",
  abstract =     "In this article, we present an extended relational
                 algebra with universally or existentially quantified
                 classes as attribute values. The proposed extension can
                 greatly enhance the expressive power of relational
                 systems, and significantly reduce the size of a
                 database, at small additional computational cost. We
                 also show how the proposed extensions can be built on
                 top of a standard relational database system.",
  acknowledgement = ack-nhfb,
  affiliation =  "AT\&T Bell Labs., Murray Hill, NJ, USA",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models.",
}

@Article{Baekgaard:1995:ICN,
  author =       "Lars Baekgaard and Leo Mark",
  title =        "Incremental Computation of Nested Relational Query
                 Expressions",
  journal =      j-TODS,
  volume =       "20",
  number =       "2",
  pages =        "111--148",
  month =        jun,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-2/p111-baekgaard/p111-baekgaard.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-2/p111-baekgaard/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/210198.html",
  abstract =     "Efficient algorithms for incrementally computing
                 nested query expressions do not exist. Nested query
                 expressions are query expressions in which
                 selection/join predicates contain subqueries. In order
                 to respond to this problem, we propose a two-step
                 strategy for incrementally computing nested query
                 expressions. In step (1), the query expression is
                 transformed into an equivalent unnested flat query
                 expression. In step (2), the flat query expression is
                 incrementally computed. To support step (1), we have
                 developed a very concise algebra-to-algebra
                 transformation algorithm, and we have formally proved
                 its correctness. The flat query expressions resulting
                 from the transformation make intensive use of the
                 relational set-difference operator. To support step
                 (2), we present and analyze an efficient algorithm for
                 incrementally computing set differences based on view
                 pointer caches. When combined with existing incremental
                 algorithms for SPJ queries, our incremental
                 set-difference algorithm can be used to compute the
                 unnested flat query expressions efficiently. It is
                 important to notice that without our incremental
                 set-difference algorithm the existing incremental
                 algorithms for SPJ queries are useless for any query
                 involving the set-difference operator, including
                 queries that are not the result of unnesting nested
                 queries.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; incremental computation; nested query
                 expressions; performance; set differences; unnesting;
                 view pointer caches",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.2}: Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Access
                 methods. {\bf H.2.3}: Information Systems, DATABASE
                 MANAGEMENT, Languages, Query languages. {\bf H.3.3}:
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval, Search process. {\bf
                 H.3.3}: Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval, Query
                 formulation. {\bf E.5}: Data, FILES,
                 Sorting/searching.",
}

@Article{Chomicki:1995:ECT,
  author =       "Jan Chomicki",
  title =        "Efficient Checking of Temporal Integrity Constraints
                 Using Bounded History Encoding",
  journal =      j-TODS,
  volume =       "20",
  number =       "2",
  pages =        "149--186",
  month =        jun,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-2/p149-chomicki/p149-chomicki.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-2/p149-chomicki/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/210200.html",
  abstract =     "We present an efficient implementation method for
                 temporal integrity constraints formulated in Past
                 Temporal Logic. Although the constraints can refer to
                 past states of the database, their checking does not
                 require that the entire database history be stored.
                 Instead, every database state is extended with
                 auxiliary relations that contain the historical
                 information necessary for checking constraints.
                 Auxiliary relations can be implemented as materialized
                 relational views.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active databases; algorithms; database integrity;
                 integrity constraints; real-time databases; temporal
                 databases; temporal logic; theory; triggers",
  subject =      "{\bf H.2.0}: Information Systems, DATABASE MANAGEMENT,
                 General, Security, integrity, and protection. {\bf
                 F.4.1}: Theory of Computation, MATHEMATICAL LOGIC AND
                 FORMAL LANGUAGES, Mathematical Logic. {\bf H.2.3}:
                 Information Systems, DATABASE MANAGEMENT, Languages,
                 Data description languages (DDL).",
}

@Article{Graefe:1995:FAU,
  author =       "Goetz Graefe and Richard L. Cole",
  title =        "Fast Algorithms for Universal Quantification in large
                 Databases",
  journal =      j-TODS,
  volume =       "20",
  number =       "2",
  pages =        "187--236",
  month =        jun,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-2/p187-graefe/p187-graefe.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-2/p187-graefe/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/210202.html",
  abstract =     "Universal quantification is not supported directly in
                 most database systems despite the fact that it adds
                 significant power to a system's query processing and
                 inference capabilities, in particular for the analysis
                 of many-to-many relationships and of set-valued
                 attributes. One of the main reasons for this omission
                 has been that universal quantification algorithms and
                 their performance have not been explored for large
                 databases. In this article, we describe and compare
                 three known algorithms and one recently proposed
                 algorithm for relational division, the algebra operator
                 that embodies universal quantification. For each
                 algorithm, we investigate the performance effects of
                 explicit duplicate removal and referential integrity
                 enforcement, variants for inputs larger than memory,
                 and parallel execution strategies. Analytical and
                 experimental performance comparisons illustrate the
                 substantial differences among the algorithms. Moreover,
                 comparisons demonstrate that the recently proposed
                 division algorithm evaluates a universal quantification
                 predicate over two relations as fast as hash (semi-)
                 join evaluates an existential quantification predicate
                 over the same relations. Thus, existential and
                 universal quantification can be supported with equal
                 efficiency by adding the recently proposed algorithm to
                 a query evaluation system. A second result of our study
                 is that universal quantification should be expressed
                 directly in a database query language, because most
                 query optimizers do not recognize the rather indirect
                 formulations available in SQL as relational division
                 and therefore produce very poor evaluation plans for
                 many universal quantification queries.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; experimentation",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf E.5}: Data, FILES.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML).",
}

@Article{Chen:1995:QED,
  author =       "Weidong Chen",
  title =        "Query Evaluation in Deductive Databases with
                 Alternating Fixpoint Semantics",
  journal =      j-TODS,
  volume =       "20",
  number =       "3",
  pages =        "239--287",
  month =        sep,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  MRclass =      "68P15 (68P20)",
  MRnumber =     "96g:68024",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-3/p239-chen/p239-chen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-3/p239-chen/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/211416.html",
  abstract =     "First-order formulas allow natural descriptions of
                 queries and rules. Van Gelder's alternating fixpoint
                 semantics extends the well-founded semantics of normal
                 logic programs to general logic programs with arbitrary
                 first-order formulas in rule bodies. However, an
                 implementation of general logic programs through the
                 standard translation into normal logic programs does
                 not preserve the alternating fixpoint semantics. This
                 paper presents a direct method for goal-oriented query
                 evaluation of general logic programs. Every general
                 logic program is first transformed into a normal form
                 where the body of each rule is either an existential
                 conjunction of literals or a universal disjunction of
                 literals. Techniques of memoing and loop checking are
                 incorporated so that termination and polynomial-time
                 data complexity are guaranteed for deductive databases
                 (or function-free programs). Results of the soundness
                 and search space completeness are established.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "alternating fixpoint; deductive databases; negation;
                 predicate logic; query evaluations; SLG resolution;
                 theory; verification",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing. {\bf I.2.3}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem Proving,
                 Logic programming.",
}

@Article{Ioannidis:1995:CCQ,
  author =       "Yannis E. Ioannidis and Raghu Ramakrishnan",
  title =        "Containment of Conjunctive Queries: Beyond Relations
                 as Sets",
  journal =      j-TODS,
  volume =       "20",
  number =       "3",
  pages =        "288--324",
  month =        sep,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-3/p288-ioannidis/p288-ioannidis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-3/p288-ioannidis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/211419.html",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; conjunctive queries; equivalence;
                 languages; multisets; query containment; query
                 optimization; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf F.0}: Theory of
                 Computation, GENERAL. {\bf H.2.3}: Information Systems,
                 DATABASE MANAGEMENT, Languages, Query languages. {\bf
                 H.1.1}: Information Systems, MODELS AND PRINCIPLES,
                 Systems and Information Theory.",
}

@Article{Shasha:1995:TCA,
  author =       "Dennis Shasha and Fran{\c{c}}ois Llirbat and Eric
                 Simon and Patrick Valduriez",
  title =        "Transaction Chopping: Algorithms and Performance
                 Studies",
  journal =      j-TODS,
  volume =       "20",
  number =       "3",
  pages =        "325--363",
  month =        sep,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-3/p325-shasha/p325-shasha.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-3/p325-shasha/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/211427.html",
  abstract =     "Chopping transactions into pieces is good for
                 performance but may lead to nonserializable executions.
                 Many researchers have reacted to this fact by either
                 inventing new concurrency-control mechanisms, weakening
                 serializability, or both. We adopt a different
                 approach. We assume a user who\par

                 ---has access only to user-level tools such as (1)
                 choosing isolation degrees 1ndash;4, (2) the ability to
                 execute a portion of a transaction using multiversion
                 read consistency, and (3) the ability to reorder the
                 instructions in transaction programs; and\par

                 ---knows the set of transactions that may run during a
                 certain interval (users are likely to have such
                 knowledge for on-line or real-time transactional
                 applications).\par

                 Given this information, our algorithm finds the finest
                 chopping of a set of transactions TranSet with the
                 following property: {\em If the pieces of the chopping
                 execute serializably, then TranSet executes
                 serializably}. This permits users to obtain more
                 concurrency while preserving correctness. Besides
                 obtaining more intertransaction concurrency, chopping
                 transactions in this way can enhance intratransaction
                 parallelism.\par

                 The algorithm is inexpensive, running in $ O(n x(e +
                 m)) $ time, once conflicts are identified, using a
                 naive implementation, where $n$ is the number of
                 concurrent transactions in the interval $e$ is the
                 number of edges in the conflict graph among the
                 transactions, and $m$ is the maximum number of accesses
                 of any transaction. This makes it feasible to add as a
                 tuning knob to real systems.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; locking; multidatabase; performance;
                 serializability; tuning",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf D.4.8}: Software,
                 OPERATING SYSTEMS, Performance, Simulation. {\bf
                 H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf I.6.8}: Computing
                 Methodologies, SIMULATION AND MODELING, Types of
                 Simulation, Discrete event.",
}

@Article{Chen:1995:EML,
  author =       "I.-Min A. Chen and Richard Hull and Dennis McLeod",
  title =        "An Execution Model for Limited Ambiguity Rules and its
                 Application to Derived Data Update",
  journal =      j-TODS,
  volume =       "20",
  number =       "4",
  pages =        "365--413",
  month =        dec,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-4/p365-chen/p365-chen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-4/p365-chen/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/219039.html",
  abstract =     "A novel execution model for rule application in active
                 databases is developed and applied to the problem of
                 updating derived data in a database represented using a
                 semantic, object-based database model. The execution
                 model is based on the use of ``limited ambiguity
                 rules'' (LARs), which permit disjunction in rule
                 actions. The execution model essentially performs a
                 breadth-first exploration of alternative extensions of
                 a user-requested update. Given an object-based database
                 schema, both integrity constraints and specifications
                 of derived classes and attributes are compiled into a
                 family of limited ambiguity rules. A theoretical
                 analysis shows that the approach is sound: the
                 execution model returns all valid ``completions'' of a
                 user-requested update, or terminates with an
                 appropriate error notification. The complexity of the
                 approach in connection with derived data update is
                 considered.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Languages; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active database systems; algorithms; deltas on
                 database states; derived data; design; languages;
                 limited ambiguity rules; management; semantic data
                 models; theory; update propagation",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.m}: Information
                 Systems, DATABASE MANAGEMENT, Miscellaneous.",
}

@Article{Fraternali:1995:SAD,
  author =       "Piero Fraternali and Letizia Tanca",
  title =        "A Structured Approach for the Definition of the
                 Semantics of Active Databases",
  journal =      j-TODS,
  volume =       "20",
  number =       "4",
  pages =        "414--471",
  month =        dec,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-4/p414-fraternali/p414-fraternali.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-4/p414-fraternali/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/219042.html",
  abstract =     "Active DBMSs couple database technology with
                 rule-based programming to achieve the capability of
                 reaction to database (and possibly external) stimuli,
                 called {\em events}. The reactive capabilities of
                 active databases are useful for a wide spectrum of
                 applications, including security, view materialization,
                 integrity checking and enforcement, or heterogeneous
                 database integration, which makes this technology very
                 promising for the near future. An active database
                 system consists of a (passive) database and a set of
                 {\em active rules\/}; the most popular form of active
                 rule is the so-called {\em event-condition-action\/}
                 (ECA) rule, which specifies an action to be executed
                 upon the occurrence of one or more events, provided
                 that a condition holds. Several active database systems
                 and prototypes have been designed and partially or
                 completely implemented. Unfortunately, they have been
                 designed in a totally independent way, without the
                 support of a common theory dictating the semantics of
                 ECA rules, and thus often show different behaviors for
                 rules with a similar form. In this article we consider
                 a number of different possible options in the behavior
                 of an active DBMS, based on a broad analysis of some of
                 the best known implemented systems and prototypes. We
                 encode these options in a user-readable form, called
                 {\em Extended ECA}. A rule from any existing system can
                 be rewritten in this formalism making all the semantic
                 choices apparent. Then an EECA rule can be
                 automatically translated into an internal (less
                 readable) format, based on a logical style, which is
                 called {\em core\/} format: the execution semantics of
                 core rules is specified as the fixpoint of a simple
                 transformation involving core rules. As an important
                 premise to this research, a semantics for database
                 updates and transactions has also been established,
                 with respect to a notion of state that comprises both
                 data and events. The article also presents an extensive
                 bibliography on the subject of active databases.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active database systems; database rule processing;
                 events; fixpoint semantics; languages; rules;
                 semantics; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf F.3.2}: Theory of Computation,
                 LOGICS AND MEANINGS OF PROGRAMS, Semantics of
                 Programming Languages, Operational semantics. {\bf
                 H.2.m}: Information Systems, DATABASE MANAGEMENT,
                 Miscellaneous. {\bf I.2.5}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Programming Languages and
                 Software.",
}

@Article{Srivastava:1995:SOD,
  author =       "Divesh Srivastava and S. Sudarshan and Raghu
                 Ramakrishnan and Jeffrey F. Naughton",
  title =        "Space Optimization in Deductive Databases",
  journal =      j-TODS,
  volume =       "20",
  number =       "4",
  pages =        "472--516",
  month =        dec,
  year =         "1995",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1995-20-4/p472-srivastava/p472-srivastava.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1995-20-4/p472-srivastava/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/219056.html",
  abstract =     "In the bottom-up evaluation of logic programs and
                 recursively defined views on databases, all generated
                 facts are usually assumed to be stored until the end of
                 the evaluation. Discarding facts during the evaluation,
                 however, can considerably improve the efficiency of the
                 evaluation: the space needed to evaluate the program,
                 the I/O costs, the costs of maintaining and accessing
                 indices, and the cost of eliminating duplicates may all
                 be reduced. Given an evaluation method that is sound,
                 complete, and does not repeat derivation steps, we
                 consider how facts can be discarded during the
                 evaluation without compromising these properties. We
                 show that every such space optimization method has
                 certain components, the first to ensure soundness and
                 completeness, the second to avoid redundancy (i.e.,
                 repetition of derivations), and the third to reduce
                 ``fact lifetimes'' (i.e., the time period for which
                 each fact must be retained during evaluation). We
                 present new techniques based on providing bounds on the
                 number of derivations and uses of facts, and using
                 monotonicity constraints for each of the first two
                 components, and provide novel synchronization
                 techniques for the third component of a space
                 optimization method. We describe how techniques for
                 each of the three components can be combined in
                 practice to obtain a space optimization method for a
                 program. Our results are also of importance in
                 applications such as sequence querying, and in active
                 databases where triggers are defined over multiple
                 ``events.''",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; bottom-up query evaluation deductive
                 database systems; discarding facts; languages; logic
                 programming; theory",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages.",
}

@Article{Baralis:1996:MTA,
  author =       "Elena Baralis and Stefano Ceri and Stefano
                 Paraboschi",
  title =        "Modularization Techniques for Active Rules Design",
  journal =      j-TODS,
  volume =       "21",
  number =       "1",
  pages =        "1--29",
  month =        mar,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-1/p1-baralis/p1-baralis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-1/p1-baralis/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/227605.html",
  abstract =     "Active database systems can be used to establish and
                 enforce data management policies. A large amount of the
                 semantics that normally needs to be coded in
                 application programs can be abstracted and assigned to
                 active rules. This trend is sometimes called
                 ``knowledge independence'' a nice consequence of
                 achieving full knowledge independence is that data
                 management policies can then effectively evolve just by
                 modifying rules instead of application programs. Active
                 rules, however, may be quite complex to understand and
                 manage: rules react to arbitrary event sequences, they
                 trigger each other, and sometimes the outcome of rule
                 processing may depend on the order in which events
                 occur or rules are scheduled. Although reasoning on a
                 large collection of rules is very difficult, the task
                 becomes more manageable when the rules are few.
                 Therefore, we are convinced that modularization,
                 similar to what happens in any software development
                 process, is the key principle for designing active
                 rules; however, this important notion has not been
                 addressed so far. This article introduces a
                 modularization technique for active rules called
                 stratification; it presents a theory of stratification
                 and indicates how stratification can be practically
                 applied. The emphasis of this article is on providing a
                 solution to a very concrete and practical problem;
                 therefore, our approach is illustrated by several
                 examples.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active database systems; database rule processing;
                 design; modularization; static analysis; termination",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Schema and subschema. {\bf D.2.1}:
                 Software, SOFTWARE ENGINEERING,
                 Requirements/Specifications, Methodologies. {\bf
                 D.2.2}: Software, SOFTWARE ENGINEERING, Tools and
                 Techniques, Modules and interfaces. {\bf H.2.8}:
                 Information Systems, DATABASE MANAGEMENT, Database
                 applications.",
}

@Article{Buneman:1996:PTI,
  author =       "Peter Buneman and Atsushi Ohori",
  title =        "Polymorphism and Type Inference in Database
                 Programming",
  journal =      j-TODS,
  volume =       "21",
  number =       "1",
  pages =        "30--76",
  month =        mar,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-1/p30-buneman/p30-buneman.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-1/p30-buneman/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/227609.html",
  abstract =     "In order to find a static type system that adequately
                 supports database languages, we need to express the
                 most general type of a program that involves database
                 operations. This can be achieved through an extension
                 to the type system of ML that captures the polymorphic
                 nation of field selection, together with a techniques
                 that generalizes relational operators to arbitrary data
                 structures. The combination provides a statically typed
                 language in which generalized relational databases may
                 be cleanly represented as typed structures. As in ML
                 types are inferred, which relieves the programmer of
                 making the type assertions that may be required in a
                 complex database environment.\par

                 These extensions may also be used to provide static
                 polymorphic typechecking in object-oriented languages
                 and databases. A problem that arises with
                 object-oriented databases is the apparent need for
                 dynamic typechecking when dealing queries on
                 heterogeneous collections of objects. An extension of
                 the type system needed for generalized relational
                 operations can also be used for manipulating
                 collections of dynamically typed values in a statically
                 typed language. A prototype language based on these
                 ideas has been implemented. While it lacks a proper
                 treatment of persistent data, it demonstrates that a
                 wide variety of database structures can be cleanly
                 represented in a polymorphic programming language.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "generalized relational algebra; inheritance;
                 object-oriented databases; polymorphism; record
                 calculus; theory; type inference",
  subject =      "{\bf D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features, Data types and structures.
                 {\bf D.3.1}: Software, PROGRAMMING LANGUAGES, Formal
                 Definitions and Theory. {\bf D.3.2}: Software,
                 PROGRAMMING LANGUAGES, Language Classifications,
                 Applicative languages. {\bf D.3.3}: Software,
                 PROGRAMMING LANGUAGES, Language Constructs and
                 Features, Abstract data types. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models. {\bf H.2.1}: Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Schema and subschema. {\bf
                 H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Database (persistent) programming languages.
                 {\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages.",
}

@Article{Mok:1996:NFP,
  author =       "Wai Yin Mok and Yiu-Kai Ng and David W. Embley",
  title =        "A Normal Form for Precisely Characterizing Redundancy
                 in Nested Relations",
  journal =      j-TODS,
  volume =       "21",
  number =       "1",
  pages =        "77--106",
  month =        mar,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-1/p77-mok/p77-mok.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-1/p77-mok/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/227612.html",
  abstract =     "We give a straightforward definition for redundancy in
                 individual nested relations and define a new normal
                 form that precisely characterizes redundancy for nested
                 relations. We base our definition of redundancy on an
                 arbitrary set of functional and multivalued
                 dependencies, and show that our definition of nested
                 normal form generalizes standard relational
                 normalization theory. In addition, we give a condition
                 that can prevent an unwanted structural anomaly in
                 nested relations, namely, embedded nested relations
                 with at most one tuple. Like other normal forms, our
                 nested normal form can serve as a guide for database
                 design.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data redundancy; database design; design; functional
                 and multivalued dependencies; nested normal form;
                 nested relations; normalization theory; scheme trees;
                 theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms. {\bf H.2.1}: Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models.",
}

@Article{Mumick:1996:MC,
  author =       "Inderpal Singh Mumick and Sheldon J. Finkelstein and
                 Hamid Pirahesh and Raghu Ramakrishnan",
  title =        "Magic conditions",
  journal =      j-TODS,
  volume =       "21",
  number =       "1",
  pages =        "107--155",
  month =        mar,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-1/p107-mumick/p107-mumick.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-1/p107-mumick/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/227624.html",
  abstract =     "Much recent work has focused on the bottom-up
                 evaluation of Datalog programs [Bancilhon and
                 Ramakrishnan 1988]. One approach, called magic-sets, is
                 based on rewriting a logic program so that bottom-up
                 fixpoint evaluation of the program avoids generation of
                 irrelevant facts [Bancilhon et al. 1986; Beeri and
                 Ramakrishnan 1987; Ramakrishnan 1991]. It was widely
                 believed for some time that the principal application
                 of the magic-sets technique is to restrict computation
                 in recursive queries using equijoin predicates. We
                 extend the magic-sets transformation to use predicates
                 other than equality ($ X > 10 $, for example) in
                 restricting computation. The resulting {\em ground
                 magic-sets transformation\/} is an important step in
                 developing an extended magic-sets transformation that
                 has practical utility in ``real'' relational databases,
                 not only for recursive queries, but for nonrecursive
                 queries as well [Mumick et al. 1990b; Mumick 1991].",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Measurement; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; bottom-up evaluation; constraint logic
                 programming; constraints; deductive databases; magic
                 sets; measurement; query optimization; relational
                 databases; Starburst; theory; verification",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing.",
}

@Article{Liu:1996:BBS,
  author =       "Ling Liu and Robert Meersman",
  title =        "The Building Blocks for Specifying Communication
                 Behavior of Complex Objects: An Activity-Driven
                 Approach",
  journal =      j-TODS,
  volume =       "21",
  number =       "2",
  pages =        "157--207",
  month =        jun,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-2/p157-liu/p157-liu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-2/p157-liu/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232622.html",
  abstract =     "Communication behavior represents dynamic evolution
                 and cooperation of a group of objects in accomplishing
                 a task. It is an important feature in object-oriented
                 systems. We propose the concept of activity as a basic
                 building block for declarative specification of
                 communication behavior in object-oriented database
                 systems, including the temporal ordering of message
                 exchanges within object communication and the
                 behavioral relationships between activity executions.
                 We formally introduce two kinds of activity composition
                 mechanisms: {\em activity specialization\/} and
                 activity {\em aggregation\/} for abstract
                 implementation of communication behavior. The former is
                 suited for behavioral refinement of existing activities
                 into specialized activities. The latter is used for
                 behavioral composition of simpler activities into
                 complex activities, and ultimately, into the envisaged
                 database system. We use first-order temporal logic as
                 an underlying formalism for specification of
                 communication constraints. The well known
                 Air-traffic-control case is used as a running example
                 to highlight the underlying concepts, to illustrate the
                 usefulness, and to assess the effectiveness of the
                 activity model for declarative specification of
                 communication behavior in the relevant universe of
                 discourse. We also propose a methodological framework
                 for integrating activity schema with entity schema in
                 an object-oriented design environment.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "activity aggregation; activity patterns; activity
                 specialization; communication behavior; design;
                 first-order temporal logic; languages; object-oriented
                 databases; synchronization schemes; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf D.3.1}: Software, PROGRAMMING
                 LANGUAGES, Formal Definitions and Theory, Semantics.
                 {\bf F.3.1}: Theory of Computation, LOGICS AND MEANINGS
                 OF PROGRAMS, Specifying and Verifying and Reasoning
                 about Programs, Specification techniques. {\bf H.2.3}:
                 Information Systems, DATABASE MANAGEMENT, Languages.
                 {\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems. {\bf C.2.4}: Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases.",
}

@Article{Ross:1996:TRE,
  author =       "Kenneth A. Ross",
  title =        "Tail Recursion Elimination in Deductive Databases",
  journal =      j-TODS,
  volume =       "21",
  number =       "2",
  pages =        "208--237",
  month =        jun,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-2/p208-ross/p208-ross.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-2/p208-ross/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232628.html",
  abstract =     "We consider an optimization technique for deductive
                 and relational databases. The optimization technique is
                 an extension of the magic templates rewriting, and it
                 can improve the performance of query evaluation by not
                 materializing the extension of intermediate views.
                 Standard relational techniques, such as unfolding
                 embedded view definitions, do not apply to recursively
                 defined views, and so alternative techniques are
                 necessary. We demonstrate the correctness of our
                 rewriting. We define a class of ``nonrepeating'' view
                 definitions, and show that for certain queries our
                 rewriting performs at least as well as magic templates
                 on nonrepeating views, and often much better. A
                 syntactically recognizable property, called ``weak
                 right-linearity'', is proposed. Weak right-linearity is
                 a sufficient condition for nonrepetition and is more
                 general than right-linearity. Our technique gives the
                 same benefits as right-linear evaluation of
                 right-linear views, while applying to a significantly
                 more general class of views.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; deductive databases; magic sets; query
                 optimization; tail recursion",
  subject =      "{\bf H.2.4}: Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf I.2.3}: Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem Proving,
                 Logic programming.",
}

@Article{Bell:1996:IDD,
  author =       "Colin Bell and Anil Nerode and Raymond T. Ng and V. S.
                 Subrahmanian",
  title =        "Implementing Deductive Databases by Mixed Integer
                 Programming",
  journal =      j-TODS,
  volume =       "21",
  number =       "2",
  pages =        "238--269",
  month =        jun,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-2/p238-bell/p238-bell.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-2/p238-bell/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232691.html",
  abstract =     "Existing and past generations of Prolog compilers have
                 left deduction to run-time and this may account for the
                 poor run-time performance of existing Prolog systems.
                 Our work tries to minimize run-time deduction by
                 shifting the deductive process to compile-time. In
                 addition, we offer an alternative inferencing procedure
                 based on translating logic to mixed integer
                 programming. This makes available for research and
                 implementation in deductive databases, all the
                 theorems, algorithms, and software packages developed
                 by the operations research community over the past 50
                 years. The method keeps the same query language as for
                 disjunctive deductive databases, only the inferencing
                 procedure changes. The language is purely declarative,
                 independent of the order of rules in the program, and
                 independent of the order in which literals occur in
                 clause bodies. The technique avoids Prolog's problem of
                 infinite looping. It saves run-time by doing primary
                 inferencing at compile-time. Furthermore, it is
                 incremental in nature. The first half of this article
                 translates disjunctive clauses, integrity constraints,
                 and database facts into Boolean equations, and develops
                 procedures to use mixed integer programming methods to
                 compute equations, and develops procedures to use mixed
                 integer programming methods to compute equations, and
                 develops procedures to use mixed integer programming
                 methods to compute equations, and develops procedures
                 to use mixed integer programming methods to
                 compute\par

                 ---least models of definite deductive databases,
                 and\par

                 ---minimal models and the Generalized Closed World
                 Assumption of disjunctive databases.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; minimal models; negation and disjunction in
                 deductive databases; theory",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages. {\bf F.4.1}: Theory of Computation,
                 MATHEMATICAL LOGIC AND FORMAL LANGUAGES, Mathematical
                 Logic. {\bf I.2.4}: Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Knowledge Representation Formalisms and
                 Methods.",
}

@Article{Guo:1996:SSI,
  author =       "Sha Guo and Wei Sun and Mark A. Weiss",
  title =        "Solving Satisfiability and Implication Problems in
                 Database Systems",
  journal =      j-TODS,
  volume =       "21",
  number =       "2",
  pages =        "270--293",
  month =        jun,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-2/p270-guo/p270-guo.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-2/p270-guo/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232692.html",
  abstract =     "Satisfiability, implication, and equivalence problems
                 involving conjunctive inequalities are important and
                 widely encountered database problems that need to be
                 efficiently and effectively processed. In this article
                 we consider two popular types of arithmetic
                 inequalities, ({\em X{\bf op\/}Y}) and ({\em X {\bf
                 op\/} C}), where $X$ and $Y$ are attributes, $C$ is a
                 constant of the domain or $X$, and {\bf op}[epsilon]
                 $<$, $ < = $, $=$, ***, $>$, $ > = $. These
                 inequalities are most frequently used in a database
                 system, inasmuch as the former type of inequality
                 represents a 0-join, and the latter is a selection. We
                 study the satisfiability and implication problems under
                 the integer domain and the real domain, as well as
                 under two different operator sets ({$<$, $ < = $, =, $
                 > = $, $>$} and {$<$, $ < = $, =, ***, $ > = $, $>$}).
                 Our results show that solutions under different domains
                 and/or different operator sets are quite different. Out
                 of these eight cases, excluding two cases that had been
                 shown to be NP-hard, we either report the first
                 necessary and sufficient conditions for these problems
                 as well as their efficient algorithms with complexity
                 analysis (for four cases), or provide an improved
                 algorithm (for two cases). These iff conditions and
                 algorithms are essential to database designers,
                 practitioners, and researchers. These algorithms have
                 been implemented and an experimental study comparing
                 the proposed algorithms and those previously known is
                 conducted. Our experiments show that the proposed
                 algorithms are more efficient than previously known
                 algorithms even for small input. The C++ code can be
                 obtained by an anonymous ftp from
                 \path=archive.fiu.edu=.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; deduction; equivalence; implication;
                 languages; reasoning; satisfiabilty; theory",
  subject =      "{\bf F.2.2}: Theory of Computation, ANALYSIS OF
                 ALGORITHMS AND PROBLEM COMPLEXITY, Nonnumerical
                 Algorithms and Problems, Complexity of proof
                 procedures. {\bf H.2.4}: Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing. {\bf I.2.3}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Deduction and Theorem Proving, Deduction. {\bf I.1.2}:
                 Computing Methodologies, ALGEBRAIC MANIPULATION,
                 Algorithms, Analysis of algorithms.",
}

@Article{Ciaccia:1996:DKB,
  author =       "Paolo Ciaccia and Paolo Tiberio and Pavel Zezula",
  title =        "Declustering of Key-Based Partitioned Signature
                 Files",
  journal =      j-TODS,
  volume =       "21",
  number =       "3",
  pages =        "295--338",
  month =        sep,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-3/p295-ciaccia/p295-ciaccia.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-3/p295-ciaccia/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232755.html",
  abstract =     "Access methods based on signature files can largely
                 benefit from possibilities offered by parallel
                 environments. To this end, an effective {\em
                 declustering\/} strategy that would distribute
                 signatures over a set of parallel independent disks has
                 to be combined with a synergic {\em clustering\/} which
                 is employed to avoid searching the whole signature file
                 while executing a query. This article proposes two
                 parallel signature file organizations, Hamming Filter
                 ({\em HF\/}) and Hamming$^+$ Filter ($ H^+F $ ), whose
                 common declustering strategy is based on {\em error
                 correcting codes}, and where clustering is achieved by
                 organizing signatures into fixed-size buckets, each
                 containing signatures sharing the same {\em key\/}
                 value. {\em HF\/} allocates signatures on disks in a
                 static way and works well if a correct relationship
                 holds between the parameters of the code and the size
                 of the file. $ H^+F $ is a generalization of $ H F $
                 suitable to manage highly dynamic files. It uses a
                 dynamic declustering, obtained through a {\em
                 sequence\/} of codes, and organizes a smooth migration
                 of signatures between disks so that high performance
                 levels are retained regardless of current file size.
                 Theoretical analysis characterizes the best-case,
                 expected, and worst-case behaviors of these
                 organizations. Analytical results are verified by
                 experiments on prototype systems.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "design; error correcting codes; information retrieval;
                 parallel independent disks; partial match queries;
                 performance; performance evaluation; superimposed
                 coding",
  subject =      "{\bf E.4}: Data, CODING AND INFORMATION THEORY. {\bf
                 E.5}: Data, FILES, Organization/structure. {\bf H.2.2}:
                 Information Systems, DATABASE MANAGEMENT, Physical
                 Design, Access methods. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing. {\bf H.3.2}: Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Information Storage,
                 File organization.",
}

@Article{Dey:1996:PRM,
  author =       "Debabrata Dey and Sumit Sarkar",
  title =        "A Probabilistic Relational Model and Algebra",
  journal =      j-TODS,
  volume =       "21",
  number =       "3",
  pages =        "339--369",
  month =        sep,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-3/p339-dey/p339-dey.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-3/p339-dey/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232796.html",
  abstract =     "Although the relational model for databases provides a
                 great range of advantages over other data models, it
                 lacks a comprehensive way to handle incomplete and
                 uncertain data. Uncertainty in data values, however, is
                 pervasive in all real-world environments and has
                 received much attention in the literature. Several
                 methods have been proposed for incorporating uncertain
                 data into relational databases. However, the current
                 approaches have many shortcomings and have not
                 established an acceptable extension of the relational
                 model. In this paper, we propose a consistent extension
                 of the relational model. We present a revised
                 relational structure and extend the relational algebra.
                 The extended algebra is shown to be closed, a
                 consistent extension of the conventional relational
                 algebra, and reducible to the latter.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data incompleteness; data uncertainty; languages;
                 probabilistic relation; probability calculus;
                 relational algebra; relational model; theory",
  subject =      "{\bf H.2.1}: Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf F.4.3}: Theory of
                 Computation, MATHEMATICAL LOGIC AND FORMAL LANGUAGES,
                 Formal Languages, Algebraic language theory. {\bf G.3}:
                 Mathematics of Computing, PROBABILITY AND STATISTICS,
                 Statistical computing. {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 manipulation languages (DML). {\bf H.2.3}: Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.8}: Information Systems, DATABASE
                 MANAGEMENT, Database applications. {\bf I.2.3}:
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Deduction and Theorem Proving, Uncertainty, ``fuzzy,''
                 and probabilistic reasoning.",
}

@Article{Ghandeharizadeh:1996:HED,
  author =       "Shahram Ghandeharizadeh and Richard Hull and Dean
                 Jacobs",
  title =        "{Heraclitus}: Elevating Deltas to be First-Class
                 Citizens in a Database Programming Language",
  journal =      j-TODS,
  volume =       "21",
  number =       "3",
  pages =        "370--426",
  month =        sep,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-3/p370-ghandeharizadeh/p370-ghandeharizadeh.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-3/p370-ghandeharizadeh/;
                 http://www.acm.org/pubs/toc/Abstracts/tods/232801.html",
  abstract =     "Traditional database systems provide a user with the
                 ability to query and manipulate one database state,
                 namely the current database state. However, in several
                 emerging applications, the ability to analyze
                 ``what-if'' scenarios in order to reason about the
                 impact of an update (before committing that update) is
                 of paramount importance. Example applications include
                 hypothetical database access, active database
                 management systems, and version management, to name a
                 few. The central thesis of the Heraclitus paradigm is
                 to provide flexible support for applications such as
                 these by elevating {\em deltas}, which represent
                 updates proposed against the current database state, to
                 be first-class citizens. Heraclitus[Alg,C] is a
                 database programming language that extends C to
                 incorporate the relational algebra and deltas.
                 Operators are provided that enable the programmer to
                 explicitly construct, combine, and access deltas. Most
                 interesting is the when operator, that supports
                 hypothetical access to a delta: the expression {\bf E}
                 when [sigma] yields the value that side effect free
                 expression E would have if the value of delta
                 expression [sigma] were applied to the current database
                 state. This article presents a broad overview of the
                 philosophy underlying the Heraclitus paradigm, and
                 describes the design and prototype implementation of
                 Heraclitus[Alg, C]. A model-independent formalism for
                 the Heraclitus paradigm is also presented. To
                 illustrate the utility of Heraclitus, the article
                 presents an in-depth discussion of how Heraclitus[Alg,
                 C] can be used to specify, and thereby implement, a
                 wide range of execution models for rule application in
                 active databases; this includes both prominent
                 execution models presented in the literature, and more
                 recent ``customized'' execution models with novel
                 features.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active databases; deltas; design; execution model for
                 rule application; hypothetical access; hypothetical
                 database state; languages",
  subject =      "{\bf H.2.3}: Information Systems, DATABASE MANAGEMENT,
                 Languages, Database (persistent) programming languages.
                 {\bf D.3.3}: Software, PROGRAMMING LANGUAGES, Language
                 Constructs and Features. {\bf H.2.4}: Information
                 Systems, DATABASE MANAGEMENT, Systems, Query
                 processing. {\bf H.2.m}: Information Systems, DATABASE
                 MANAGEMENT, Miscellaneous.",
}

@Article{Kuo:1996:MVD,
  author =       "Dean Kuo",
  title =        "Model and verification of a data manager based on
                 {ARIES}",
  journal =      j-TODS,
  volume =       "21",
  number =       "4",
  pages =        "427--479",
  month =        dec,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-4/p427-kuo/p427-kuo.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-4/p427-kuo/",
  abstract =     "In this article, we model and verify a data manager
                 whose algorithm is based on ARIES. The work uses the
                 I/O automata method as the formal model and the
                 definition of correctness is defined on the interface
                 between the scheduler and the data manager.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Reliability; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "ARIES; I/O automata; reliability; system failures;
                 theory; verification",
  subject =      "{\bf F.3.1} Theory of Computation, LOGICS AND MEANINGS
                 OF PROGRAMS, Specifying and Verifying and Reasoning
                 about Programs, Assertions. {\bf H.2.2} Information
                 Systems, DATABASE MANAGEMENT, Physical Design, Recovery
                 and restart. {\bf H.2.4} Information Systems, DATABASE
                 MANAGEMENT, Systems, Transaction processing.",
}

@Article{Litwin:1996:LSD,
  author =       "Witold Litwin and Marie-Anna Neimat and Donovan A.
                 Schneider",
  title =        "{LH*} --- a scalable, distributed data structure",
  journal =      j-TODS,
  volume =       "21",
  number =       "4",
  pages =        "480--525",
  month =        dec,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-4/p480-litwin/p480-litwin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-4/p480-litwin/",
  abstract =     "We present a scalable distributed data structure
                 called LH*. LH* generalizes Linear Hashing (LH) to
                 distributed RAM and disk files. An LH* file can be
                 created from records with primary keys, or objects with
                 OIDs, provided by any number of distributed and
                 autonomous clients. It does not require a central
                 directory, and grows gracefully, through splits of one
                 bucket at a time, to virtually any number of servers.
                 The number of messages per random insertion is one in
                 general, and three in the worst case, regardless of the
                 file size. The number of messages per key search is two
                 in general, and four in the worst case. The file
                 supports parallel operations, e.g., hash joins and
                 scans. Performing a parallel operation on a file of $M$
                 buckets costs at most 2 $M$ + 1 messages, and between 1
                 and $O$ (log 2 $M$ ) rounds of messages.\par

                 We first describe the basic LH* scheme where a
                 coordinator site manages abucket splits, and splits a
                 bucket every time a collision occurs. We show that the
                 average load factor of an LH* file is 65\%--70\%
                 regardless of file size, and bucket capacity. We then
                 enhance the scheme with load control, performed at no
                 additional message cost. The average load factor then
                 increases to 80--95\%. These values are about that of
                 LH, but the load factor for LH* varies more.\par

                 We nest define LH* schemes without a coordinator. We
                 show that insert and search costs are the same as for
                 the basic scheme. The splitting cost decreases on the
                 average, but becomes more variable, as cascading splits
                 are needed to prevent file overload. Next, we briefly
                 describe two variants of splitting policy, using
                 parallel splits and presplitting that should enhance
                 performance for high-performance applications.\par

                 All together, we show that LH* files can efficiently
                 scale to files that are orders of magnitude larger in
                 size than single-site files. LH* files that reside in
                 main memory may also be much faster than single-site
                 disk files. Finally, LH* files can be more efficient
                 than any distributed file with a centralized directory,
                 or a static parallel or distributed hash file.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; data structures; design; distributed
                 access methods; extensible hashing; linear hashing;
                 performance",
  subject =      "{\bf E.2} Data, DATA STORAGE REPRESENTATIONS,
                 Hash-table representations. {\bf E.1} Data, DATA
                 STRUCTURES. {\bf H.2.1} Information Systems, DATABASE
                 MANAGEMENT, Logical Design",
}

@Article{Raschid:1996:SUR,
  author =       "Louiqa Raschid and Jorge Lobo",
  title =        "Semantics for update rule programs and implementation
                 in a relational database management system",
  journal =      j-TODS,
  volume =       "21",
  number =       "4",
  pages =        "526--572",
  month =        dec,
  year =         "1996",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1996-21-4/p526-raschid/p526-raschid.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1996-21-4/p526-raschid/",
  abstract =     "In this paper, we present our research on defining a
                 correct semantics for a class of update rule (UR)
                 programs, and discuss implementing these programs in a
                 DBMS environment. Update rules execute by updating
                 relations in a database which may cause the further
                 execution of rules. A correct semantics must guarantee
                 that the execution of the rules will terminate and that
                 it will produce a minimal updated database. The class
                 of UR programs is syntactically identified, based upon
                 a concept that is similar to stratification. We extend
                 that strict definition of stratification and allow a
                 relaxed criterion for partitioning of the rules in the
                 UR program. This relaxation allows a limited degree of
                 nondeterminism in rule execution. We define an
                 execution semantics based upon a monotonic fixpoint
                 operator TUR, resulting in a set of fixpoints for UR.
                 The monotonicity of the operator is maintained by
                 explicitly representing the effect of asserting and
                 retracting tuples in the database. A declarative
                 semantics for the update rule program is obtained by
                 associating a normal logic program UR to represent the
                 UR program. We use the stable model semantics which
                 characterize a normal logic program by a set of minimal
                 models which are called stable models. We show the
                 equivalence between the set of fixpoints for UR and the
                 set of stable models for UR. We briefly discuss
                 implementing the fixpoint semantics of the UR program
                 in a DBMS environment. Relations that can be updated by
                 the rules are {\em updatable\/} relations and they are
                 extended with two flags. An update rule is represented
                 by a database query, which queries the updatable
                 relations as well as database relations, i.e., those
                 relations which are not update by rules. We describe an
                 algorithm to process the queries and compute a fixpoint
                 in the DBMS environment and obtain a final database.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Experimentation; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "constant maintenance; deductive database; fixpoint
                 semantics; rule-based systems; stable model semantics;
                 update rules",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems. {\bf F.4.1} Theory of Computation,
                 MATHEMATICAL LOGIC AND FORMAL LANGUAGES, Mathematical
                 Logic, Model theory. {\bf H.2.3} Information Systems,
                 DATABASE MANAGEMENT, Languages, Data description
                 languages (DDL). {\bf I.2.4} Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Knowledge Representation
                 Formalisms and Methods, Predicate logic. {\bf I.2.4}
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Knowledge Representation Formalisms and Methods,
                 Representations (procedural and rule-based). {\bf
                 H.2.7} Information Systems, DATABASE MANAGEMENT,
                 Database Administration",
}

@Article{Keen:1997:EEL,
  author =       "John S. Keen and William J. Dally",
  title =        "Extended ephemeral logging: log storage management for
                 applications with long lived transactions",
  journal =      j-TODS,
  volume =       "22",
  number =       "1",
  pages =        "1--42",
  month =        mar,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-1/p1-keen/p1-keen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-1/p1-keen/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; disk management; logging; long
                 transactions; OLTP; performance; reliability",
  subject =      "{\bf H.2.7} Information Systems, DATABASE MANAGEMENT,
                 Database Administration, Logging and recovery. {\bf
                 H.2.2} Information Systems, DATABASE MANAGEMENT,
                 Physical Design, Recovery and restart. {\bf H.2.4}
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Transaction processing",
}

@Article{Galindo-Legaria:1997:OSR,
  author =       "C{\'e}sar Galindo-Legaria and Arnon Rosenthal",
  title =        "Outerjoin simplification and reordering for query
                 optimization",
  journal =      j-TODS,
  volume =       "22",
  number =       "1",
  pages =        "43--74",
  month =        mar,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-1/p43-galindo-legaria/p43-galindo-legaria.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-1/p43-galindo-legaria/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; outerjoins; query optimization; query
                 reordering; theory; verification",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing. {\bf G.2.2} Mathematics of
                 Computing, DISCRETE MATHEMATICS, Graph Theory, Graph
                 algorithms",
}

@Article{Peters:1997:AMD,
  author =       "Randel J. Peters and M. Tamer {\"O}zsu",
  title =        "An axiomatic model of dynamic schema evolution in
                 objectbase systems",
  journal =      j-TODS,
  volume =       "22",
  number =       "1",
  pages =        "75--114",
  month =        mar,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-1/p75-peters/p75-peters.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-1/p75-peters/",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Management; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; dynamic schema evolution;
                 management; object database management systems;
                 theory",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.2.1} Information Systems,
                 DATABASE MANAGEMENT, Logical Design, Schema and
                 subschema",
}

@Article{Wang:1997:LDT,
  author =       "X. Sean Wang and Claudio Bettini and Alexander Brodsky
                 and Sushil Jajodia",
  title =        "Logical design for temporal databases with multiple
                 granularities",
  journal =      j-TODS,
  volume =       "22",
  number =       "2",
  pages =        "115--170",
  month =        jun,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-2/p115-wang/p115-wang.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-2/p115-wang/",
  abstract =     "The purpose of good database logical design is to
                 eliminate data redundancy and insertion and deletion
                 anomalies. In order to achieve this objective for
                 temporal databases, the notions of {\em temporal
                 types}, which formalize time granularities, and {\em
                 temporal functional dependencies\/} (TFDs) are
                 introduced. A temporal type is a monotonic mapping from
                 ticks of time (represented by positive integers) to
                 time sets (represented by subsets of reals) and is used
                 to capture various standard and user-defined calendars.
                 A TFD is a proper extension of the traditional
                 functional dependency and takes the form $ X**Y $
                 meaning that there is a unique value for $Y$ during one
                 tick of the temporal type [mu] for one particular $X$
                 value. An axiomatization for TFDs is given. Because a
                 finite set TFDs usually implies an infinite number of
                 TFDs, we introduce the notion of and give an
                 axiomatization for a {\em finite closure\/} to
                 effectively capture a finite set of implied TFDs that
                 are essential of the logical design. Temporal
                 normalization procedures with respect to TFDs are
                 given. Specifically, temporal Boyce-Codd normal form
                 (TBCNF) that avoids all data redundancies due to TFDs,
                 and temporal third normal form (T3NF) that allows
                 dependency preservation, are defined. Both normal forms
                 are proper extensions of their traditional
                 counterparts, BCNF and 3NF. Decomposition algorithms
                 are presented that give lossless TBCNF decompositions
                 and lossless, dependency-preserving, T3NF
                 decompositions.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; Boyce-Codd normal form; design;
                 granularity; normalization; temporal databases;
                 temporal modules; temporal relations; theory; third
                 normal form",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Normal forms. {\bf H.2.4} Information
                 Systems, DATABASE MANAGEMENT, Systems",
}

@Article{Clifford:1997:SND,
  author =       "James Clifford and Curtis Dyreson and Tom{\'a}s
                 Isakowitz and Christian S. Jensen and Richard T.
                 Snodgrass",
  title =        "On the semantics of ``now'' in databases",
  journal =      j-TODS,
  volume =       "22",
  number =       "2",
  pages =        "171--214",
  month =        jun,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-2/p171-clifford/p171-clifford.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-2/p171-clifford/",
  abstract =     "Although ``{\em now\/}'' is expressed in SQL and
                 CURRENT\_TIMESTAMP within queries, this value cannot be
                 stored in the database. However, this notion of an
                 ever-increasing current-time value has been reflected
                 in some temporal data models by inclusion of
                 database-resident variables, such as ``{\em now\/}''
                 ``{\em until-changed},'' ``**,'' ``@,'' and ``-''. Time
                 variables are very desirable, but their used also leads
                 to a new type of database, consisting of tuples with
                 variables, termed a {\em variable database}.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "indeterminacy; languages; Now; now-relative value;
                 performance; SQL; temporal query language; TSQL2",
  subject =      "{\bf H.2.3} Information Systems, DATABASE MANAGEMENT,
                 Languages, SQL. {\bf H.2.3} Information Systems,
                 DATABASE MANAGEMENT, Languages, Data description
                 languages (DDL). {\bf H.2.4} Information Systems,
                 DATABASE MANAGEMENT, Systems, Query processing",
}

@Article{Ammann:1997:AFM,
  author =       "Paul Ammann and Sushil Jajodia and Indrakshi Ray",
  title =        "Applying formal methods to semantic-based
                 decomposition of transactions",
  journal =      j-TODS,
  volume =       "22",
  number =       "2",
  pages =        "215--254",
  month =        jun,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-2/p215-ammann/p215-ammann.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-2/p215-ammann/",
  abstract =     "In some database applications the traditional approach
                 of serializability, in which transactions appear to
                 execute atomically and in isolation on a consistent
                 database state, fails to satisfy performance
                 requirements. Although many researchers have
                 investigated the process of decomposing transactions
                 into steps to increase concurrency, such research
                 typically focuses on providing algorithms necessary to
                 implement a decomposition supplied by the database
                 application developer and pays relatively little
                 attention to what constitutes a desirable decomposition
                 or how the developer should obtain one. We focus on the
                 decomposition itself. A decomposition generates proof
                 obligations whose discharge ensures desirable
                 properties with respect to the original collection of
                 transactions. We introduce the notion of semantic
                 histories to formulate and prove the necessary
                 properties, and the notion of successor sets to
                 describe efficiently the correct interleavings of
                 steps. The successor set constraints use information
                 about conflicts between steps so as to take full
                 advantage of conflict serializability at the level of
                 steps. We propose a mechanism based on two-phase
                 locking to generate correct stepwise serializable
                 histories.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "concurrency control; database management systems;
                 theory; transaction processing; verification",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf D.2.1} Software,
                 SOFTWARE ENGINEERING, Requirements/Specifications,
                 Methodologies. {\bf F.3.1} Theory of Computation,
                 LOGICS AND MEANINGS OF PROGRAMS, Specifying and
                 Verifying and Reasoning about Programs, Invariants.
                 {\bf F.3.1} Theory of Computation, LOGICS AND MEANINGS
                 OF PROGRAMS, Specifying and Verifying and Reasoning
                 about Programs, Pre- and post-conditions. {\bf F.3.1}
                 Theory of Computation, LOGICS AND MEANINGS OF PROGRAMS,
                 Specifying and Verifying and Reasoning about Programs,
                 Specification techniques",
}

@Article{Wolfson:1997:ADR,
  author =       "Ouri Wolfson and Sushil Jajodia and Yixiu Huang",
  title =        "An adaptive data replication algorithm",
  journal =      j-TODS,
  volume =       "22",
  number =       "2",
  pages =        "255--314",
  month =        jun,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-2/p255-wolfson/p255-wolfson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-2/p255-wolfson/",
  abstract =     "This article addresses the performance of distributed
                 database systems. Specifically, we present an algorithm
                 for dynamic replication of an object in distributed
                 systems. The algorithm is adaptive in the sense that it
                 changes the replication scheme of the object (i.e., the
                 set of processors at which the object is replicated) as
                 changes occur in the read-write pattern of the object
                 (i.e., the number of reads and writes issued by each
                 processor). The algorithm continuously moves the
                 replication scheme towards an optimal one. We show that
                 the algorithm can be combined with the concurrency
                 control and recovery mechanisms of ta distributed
                 database management system. The performance of the
                 algorithm is analyzed theoretically and experimentally.
                 On the way we provide a lower bound on the performance
                 of any dynamic replication algorithm.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; computer networks; dynamic data
                 allocation; file allocation; performance; replicated
                 data",
  subject =      "{\bf C.2.4} Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf H.2.4} Information Systems,
                 DATABASE MANAGEMENT, Systems, Distributed systems. {\bf
                 H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing. {\bf C.2.4} Computer
                 Systems Organization, COMPUTER-COMMUNICATION NETWORKS,
                 Distributed Systems, Distributed applications. {\bf
                 C.4} Computer Systems Organization, PERFORMANCE OF
                 SYSTEMS, Design studies. {\bf C.4} Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS, Measurement
                 techniques. {\bf C.4} Computer Systems Organization,
                 PERFORMANCE OF SYSTEMS, Modeling techniques. {\bf C.4}
                 Computer Systems Organization, PERFORMANCE OF SYSTEMS,
                 Reliability, availability, and serviceability",
}

@Article{Franklin:1997:TCS,
  author =       "Michael J. Franklin and Michael J. Carey and Miron
                 Livny",
  title =        "Transactional client-server cache consistency:
                 alternatives and performance",
  journal =      j-TODS,
  volume =       "22",
  number =       "3",
  pages =        "315--363",
  month =        sep,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-3/p315-franklin/p315-franklin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-3/p315-franklin/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1997-22-3/p315-franklin/",
  abstract =     "Client-server database systems based on a data
                 shipping model can exploit client memory resources by
                 caching copies of data items across transaction
                 boundaries. Caching reduces the need to obtain data
                 from servers or other sites on the network. In order to
                 ensure that such caching does not result in the
                 violation of transaction semantics, a transactional
                 cache consistency maintenance algorithm is required.
                 Many such algorithms have been proposed in the
                 literature and, as all provide the same functionality,
                 performance is a primary concern in choosing among
                 them. In this article we present a taxonomy that
                 describes the design space for transactional cache
                 consistency maintenance algorithms and show how
                 proposed algorithms relate to one another. We then
                 investigate the performance of six of these algorithms,
                 and use these results to examine the tradeoffs inherent
                 in the design choices identified in the taxonomy. The
                 results show that the interactions among dimensions of
                 the design space impact performance in many ways, and
                 that classifications of algorithms as simply
                 ``pessimistic'' or ``optimistic'' do not accurately
                 characterize the similarities and differences among the
                 many possible cache consistency algorithms.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; design; performance",
  subject =      "{\bf C.2.4} Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases. {\bf C.4} Computer Systems
                 Organization, PERFORMANCE OF SYSTEMS. {\bf D.4.8}
                 Software, OPERATING SYSTEMS, Performance. {\bf H.2.4}
                 Information Systems, DATABASE MANAGEMENT, Systems,
                 Concurrency.",
}

@Article{Eiter:1997:DD,
  author =       "Thomas Eiter and Georg Gottlob and Heikki Mannila",
  title =        "Disjunctive {Datalog}",
  journal =      j-TODS,
  volume =       "22",
  number =       "3",
  pages =        "364--418",
  month =        sep,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-3/p364-eiter/p364-eiter.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-3/p364-eiter/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1997-22-3/p364-eiter/",
  abstract =     "We consider disjunctive Datalog, a powerful database
                 query language based on disjunctive logic programming.
                 Briefly, disjunctive Datalog is a variant of Datalog
                 where disjunctions may appear in the rule heads;
                 advanced versions also allow for negation in the bodies
                 which can be handled according to a semantics for
                 negation in disjunctive logic programming. In
                 particular, we investigate three different semantics
                 for disjunctive Datalog: the minimal model semantics
                 the perfect model semantics, and the stable model
                 semantics. For each of these semantics, the expressive
                 power and complexity are studied. We show that the
                 possibility variants of these semantics express the
                 same set of queries. In fact, they precisely capture
                 the complexity class {\em P2}. Thus, unless the
                 Polynomial Hierarchy collapses, disjunctive Datalog is
                 more expressive that normal logic programming with
                 negation. These results are not only of theoretical
                 interest; we demonstrate that problems relevant in
                 practice such as computing the optimal tour value in
                 the Traveling Salesman Problem and eigenvector
                 computations can be handled in disjunctive Datalog, but
                 not Datalog with negation (unless the Polynomial
                 Hierarchy collapses). In addition, we study modularity
                 properties of disjunctive Datalog and investigate
                 syntactic restrictions of the formalisms.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "languages; theory",
  subject =      "{\bf H.2.3} Information Systems, DATABASE MANAGEMENT,
                 Languages, Datalog. {\bf H.2.3} Information Systems,
                 DATABASE MANAGEMENT, Languages, Query languages. {\bf
                 D.1.6} Software, PROGRAMMING TECHNIQUES, Logic
                 Programming. {\bf I.2.3} Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Deduction and Theorem Proving.
                 {\bf I.2.4} Computing Methodologies, ARTIFICIAL
                 INTELLIGENCE, Knowledge Representation Formalisms and
                 Methods. {\bf F.2.2} Theory of Computation, ANALYSIS OF
                 ALGORITHMS AND PROBLEM COMPLEXITY, Nonnumerical
                 Algorithms and Problems. {\bf F.4.1} Theory of
                 Computation, MATHEMATICAL LOGIC AND FORMAL LANGUAGES,
                 Mathematical Logic, Logic and constraint programming.",
}

@Article{Lakshmanan:1997:PFP,
  author =       "Laks V. S. Lakshmanan and Nicola Leone and Robert Ross
                 and V. S. Subrahmanian",
  title =        "{ProbView}: a flexible probabilistic database system",
  journal =      j-TODS,
  volume =       "22",
  number =       "3",
  pages =        "419--469",
  month =        sep,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-3/p419-lakshmanan/p419-lakshmanan.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-3/p419-lakshmanan/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1997-22-3/p419-lakshmanan/",
  abstract =     "Probability theory is mathematically the best
                 understood paradigm for modeling and manipulating
                 uncertain information. Probabilities of complex events
                 can be computed from those of basic events on which
                 they depend, using any of a number of strategies. Which
                 strategy is appropriate depends very much on the known
                 interdependencies among the events involved. Previous
                 work on probabilistic databases has assumed a {\em
                 fixed\/} and {\em restrictive\/} combination strategy
                 (e.g., assuming all events are pairwise independent).
                 In this article, we characterize, using postulates,
                 whole classes of strategies for conjunction,
                 disjunction, and negation, meaningful from the
                 viewpoint of probability theory. (1) We propose a
                 probabilistic relational data model and a {\em
                 generic\/}probabilistic relational algebra that neatly
                 captures {\em various strategies\/} satisfying the
                 postulates, within a {\em single unified framework.}
                 (2) We show that as long as the chosen strategies can
                 be computed in polynomial time, queries in the positive
                 fragment of the probabilistic relational algebra have
                 essentially the same data complexity as classical
                 relational algebra. (3) We establish various
                 containments and equivalences between algebraic
                 expressions, similar in spirit to those in classical
                 algebra. (4) We develop algorithms for maintaining
                 materialized probabilistic views. (5) Based on these
                 ideas, we have developed a prototype probabilistic
                 database system called ProbView on top of Dbase V.0. We
                 validate our complexity results with experiments and
                 show that rewriting certain types of queries to other
                 equivalent forms often yields substantial savings.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; languages; performance; probabilistic
                 databases; theory; view maintenance",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems. {\bf H.2.3} Information Systems, DATABASE
                 MANAGEMENT, Languages. {\bf I.2.3} Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Deduction and
                 Theorem Proving.",
}

@Article{Storey:1997:DDC,
  author =       "Veda C. Storey and Roger H. L. Chiang and Debabrata
                 Dey and Robert C. Goldstein and Shankar Sundaresan",
  title =        "Database design with common sense business reasoning
                 and learning",
  journal =      j-TODS,
  volume =       "22",
  number =       "4",
  pages =        "471--512",
  month =        dec,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-4/p471-storey/p471-storey.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-4/p471-storey/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1997-22-4/p471-storey/",
  abstract =     "Automated database design systems embody knowledge
                 about the database design process. However, their lack
                 of knowledge about the domains for which databases are
                 being developed significantly limits their usefulness.
                 A methodology for acquiring and using general world
                 knowledge about business for database design has been
                 developed and implemented in a system called the Common
                 Sense Business Reasoner, which acquires facts about
                 application domains and organizes them into a
                 hierarchical, context-dependent knowledge base. This
                 knowledge is used to make intelligent suggestions to a
                 user about the entities, attributes, and relationships
                 to include in a database design. A distance function
                 approach is employed for integrating specific facts,
                 obtained from individual design sessions, into the
                 knowledge base (learning) and for applying the
                 knowledge to subsequent design problems (reasoning).",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "common sense business reasoner; common sense learning;
                 common sense reasoning; database design; design;
                 entity-relationship model; theory",
  subject =      "{\bf D.2.1} Software, SOFTWARE ENGINEERING,
                 Requirements/Specifications. {\bf H.2.1} Information
                 Systems, DATABASE MANAGEMENT, Logical Design. {\bf
                 H.2.8} Information Systems, DATABASE MANAGEMENT,
                 Database Applications. {\bf I.2.4} Computing
                 Methodologies, ARTIFICIAL INTELLIGENCE, Knowledge
                 Representation Formalisms and Methods. {\bf I.2.6}
                 Computing Methodologies, ARTIFICIAL INTELLIGENCE,
                 Learning. {\bf I.2.1} Computing Methodologies,
                 ARTIFICIAL INTELLIGENCE, Applications and Expert
                 Systems.",
}

@Article{Tari:1997:ONF,
  author =       "Zahir Tari and John Stokes and Stefano Spaccapietra",
  title =        "Object normal forms and dependency constraints for
                 object-oriented schemata",
  journal =      j-TODS,
  volume =       "22",
  number =       "4",
  pages =        "513--569",
  month =        dec,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-4/p513-tari/p513-tari.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-4/p513-tari/",
  abstract =     "We address the development of a normalization theory
                 for object-oriented data models that have common
                 features to support objects. We first provide an
                 extension of functional dependencies to cope with the
                 richer semantics of relationships between objects,
                 called {\em path dependency}, {\em local dependency},
                 and {\em global dependency\/} constraints. Using these
                 dependency constraints, we provide normal forms for
                 object-oriented data models based on the notions of
                 {\em user interpretation\/} (user-specified dependency
                 constraints) and {\em object model}. In contrast to
                 conventional data models in which a normalized object
                 has a unique interpretation, in object-oriented data
                 models, an object may have many multiple
                 interpretations that form the model for that object. An
                 object will then be in a normal form if and only if the
                 user's interpretation is derivable from the model of
                 the object. Our normalization process is by nature
                 iterative, in which objects are restructured until
                 their models reflect the user's interpretation.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data model; functional and multivalued dependencies;
                 normal forms; object-oriented paradigm",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Normal forms}; Information
                 Systems --- Database Management --- Systems (H.2.4)",
}

@Article{Zaharioudakis:1997:AFG,
  author =       "Markos Zaharioudakis and Michael J. Carey and Michael
                 J. Franklin",
  title =        "Adaptive, fine-grained sharing in a client-server
                 {OODBMS}: a callback-based approach",
  journal =      j-TODS,
  volume =       "22",
  number =       "4",
  pages =        "570--627",
  month =        dec,
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1997-22-4/p570-zaharioudakis/p570-zaharioudakis.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1997-22-4/p570-zaharioudakis/",
  abstract =     "For reasons of simplicity and communication
                 efficiency, a number of existing object-oriented
                 database management systems are based on page server
                 architectures; data pages are their minimum unit of
                 transfer and client caching. Despite their efficiency,
                 page servers are often criticized as being too
                 restrictive when it comes to concurrency, as existing
                 systems use pages as the minimum locking unit as well.
                 In this paper we show how to support object-level
                 locking in a page-server context. Several approaches
                 are described, including an adaptive granularity
                 approach that uses page-level locking for most pages
                 but switches to object-level locking when finer-grained
                 sharing is demanded. Each of the approaches is based on
                 extending the idea of callback locking. We study the
                 performance of these approaches, comparing them to both
                 a pure page server and a pure object server. For the
                 range of workload that we have examined, our results
                 indicate that the adaptive page server provides very
                 good performance, usually outperforming the pure page
                 server and the other page-server variants as well. In
                 addition, the adaptive page server is often preferable
                 to the pure object server; our results provides insight
                 into when each approach is likely to perform better.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cache coherency; cache consistency; client-server
                 database; design; fine-grained sharing; object-oriented
                 databases; performance; performance analysis",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency; {\bf H.3.4} Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Systems and
                 Software, Distributed systems; {\bf H.2.4} Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing",
}

@Article{Anonymous:1997:AI,
  author =       "Anonymous",
  title =        "1997 Author Index",
  journal =      j-TODS,
  volume =       "22",
  number =       "4",
  pages =        "628--??",
  month =        "????",
  year =         "1997",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 7 10:36:24 MST 1998",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dyreson:1998:SVT,
  author =       "Curtis E. Dyreson and Richard T. Snodgrass",
  title =        "Supporting valid-time indeterminacy",
  journal =      j-TODS,
  volume =       "23",
  number =       "1",
  pages =        "1--57",
  month =        mar,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-1/p1-dyreson/p1-dyreson.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-1/p1-dyreson/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-1/p1-dyreson/",
  abstract =     "In {\em valid-time indeterminacy\/} it is known that
                 an event stored in a database did in fact occur, but it
                 is not known exactly {\em when}. In this paper we
                 extend the SQL data model and query language to support
                 valid-time indeterminacy. We represent the occurrence
                 time of an event with a set of possible instants,
                 delimiting when the event might have occurred, and a
                 probability distribution over that set. We also
                 describe query language constructs to retrieve
                 information in the presence of indeterminacy. These
                 constructs enable users to specify their {\em
                 credibility\/} in the underlying data and their {\em
                 plausibility\/} in the relationships among that data. A
                 denotational semantics for SQL's select statement with
                 optional credibility and plausibility constructs is
                 given. We show that this semantics is {\em reliable},
                 in that it never produces incorrect information, is
                 {\em maximal}, in that if it were extended to be more
                 informative, the results may not be reliable, and {\em
                 reduces\/} to the previous semantics when there is no
                 indeterminacy. Although the extended data model and
                 query language provide needed modeling capabilities,
                 these extensions appear initially to carry a
                 significant execution cost. A contribution of this
                 paper is to demonstrate that our approach is useful and
                 practical. An efficient representation of valid-time
                 indeterminacy and efficient query processing algorithms
                 are provided. The cost of support for indeterminacy is
                 empirically measured, and is shown to be modest.
                 Finally, we show that the approach is general, by
                 applying it to the temporal query language constructs
                 being proposed for SQL3.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; incomplete information; indeterminacy;
                 languages; probabilistic information; SQL; temporal
                 database; TSQL2; valid-time database",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design, Data models. {\bf H.2.3} Information
                 Systems, DATABASE MANAGEMENT, Languages, Query
                 languages. {\bf H.2.4} Information Systems, DATABASE
                 MANAGEMENT, Systems, Query processing.",
}

@Article{Revesz:1998:SQL,
  author =       "Peter Z. Revesz",
  title =        "Safe query languages for constraint databases",
  journal =      j-TODS,
  volume =       "23",
  number =       "1",
  pages =        "58--99",
  month =        mar,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-1/p58-revesz/p58-revesz.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-1/p58-revesz/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-1/p58-revesz/",
  abstract =     "In the database framework of Kanellakis et al. [1990]
                 it was argued that constraint query languages should
                 take constraint databases as input and give other
                 constraint databases that use the same type of atomic
                 constraints as output. This closed-form requirement has
                 been difficult to realize in constraint query languages
                 that contain the negation symbol. This paper describes
                 a general approach to restricting constraint query
                 languages with negation to safe subsets that contain
                 only programs that are evaluable in closed-form on any
                 valid constraint database input.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; languages; theory; verification",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.2.3} Information Systems,
                 DATABASE MANAGEMENT, Languages, Datalog.",
}

@Article{Stolboushkin:1998:SSD,
  author =       "Alexei P. Stolboushkin and Michael A. Taitslin",
  title =        "Safe stratified datalog with integer order does not
                 have syntax",
  journal =      j-TODS,
  volume =       "23",
  number =       "1",
  pages =        "100--109",
  month =        mar,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-1/p100-stolboushkin/p100-stolboushkin.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-1/p100-stolboushkin/",
  abstract =     "Stratified Datalog with integer (gap)-order (or {\em
                 Datalog***\/}) is considered. A {\em
                 Datalog***\/}-program is said to be safe if its
                 bottom-up processing terminates on all valid inputs. We
                 prove that safe {\em Datalog***\/}-programs do not have
                 effective syntax in the sense that there is no
                 recursively enumerable set $S$ of safe {\em
                 Datalog***\/}-programs such that every safe {\em
                 Datalog***\/}-program is equivalent to a program in
                 $S$.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "languages; theory; verification",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design; {\bf H.2.3} Information Systems,
                 DATABASE MANAGEMENT, Languages.",
}

@Article{Anonymous:1998:TR,
  author =       "Anonymous",
  title =        "{TODS} Referees",
  journal =      j-TODS,
  volume =       "23",
  number =       "1",
  pages =        "110--111",
  month =        mar,
  year =         "1998",
  bibdate =      "Mon Jan 18 18:22:17 1999",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Anonymous:1998:C,
  author =       "Anonymous",
  title =        "Corrigenda",
  journal =      j-TODS,
  volume =       "23",
  number =       "1",
  pages =        "112--112",
  month =        mar,
  year =         "1998",
  bibdate =      "Mon Jan 18 18:22:17 1999",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hellerstein:1998:OTQ,
  author =       "Joseph M. Hellerstein",
  title =        "Optimization techniques for queries with expensive
                 methods",
  journal =      j-TODS,
  volume =       "23",
  number =       "2",
  pages =        "113--157",
  month =        jun,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-2/p113-hellerstein/p113-hellerstein.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-2/p113-hellerstein/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-2/p113-hellerstein/",
  abstract =     "Object-relational database management systems allow
                 knowledgeable users to define new data types as well as
                 new methods (operators) for the types. This flexibility
                 produces an attendant complexity, which must be handled
                 in new ways for an object-relational database
                 management system to be efficient. In this article we
                 study techniques for optimizing queries that contain
                 time-consuming methods. The focus of traditional query
                 optimizers has been on the choice of join methods and
                 orders; selections have been handled by ``pushdown''
                 rules. These rules apply selections in an arbitrary
                 order before as many joins as possible, using th e
                 assumption that selection takes no time. However, users
                 of object-relational systems can embed complex methods
                 in selections. Thus selections may take significant
                 amounts of time, and the query optimization model must
                 be enhanced. In this article we carefully define a
                 query cost framework that incorporates both selectivity
                 and cost estimates for selections. We develop an
                 algorithm called Predicate Migration, and prove that it
                 produces optimal plans for queries with expensive
                 methods. We then describe our implementation of
                 Predicate Migration in the commercial object-relational
                 database management system Illustra, and discuss
                 practical issues that affect our earlier assumptions.
                 We compare Predicate Migration to a variety of simpler
                 optimization techniques, and demonstrate that Predicate
                 Migration is the best general solution to date. The
                 alternative techniques we present may be useful for
                 constrained workloads.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; expensive methods; extensibility;
                 object-relational databases; performance; predicate
                 migration; predicate placement; query optimization;
                 theory",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Query processing.",
}

@Article{Liu:1998:MAP,
  author =       "Xiangning Liu and Abdelsalam Helal and Weimin Du",
  title =        "Multiview access protocols for large-scale
                 replication",
  journal =      j-TODS,
  volume =       "23",
  number =       "2",
  pages =        "158--198",
  month =        jun,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-2/p158-liu/p158-liu.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-2/p158-liu/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-2/p158-liu/",
  abstract =     "The article proposes a scalable protocol for
                 replication management in large-scale replicated
                 systems. The protocol organizes sites and data replicas
                 into a tree-structured, hierarchical cluster
                 architecture. The basic idea of the protocol is to
                 accomplish the complex task of updating replicated data
                 with a very large number of replicas by a set of
                 related but independently committed transactions. Each
                 transaction is responsible for updating replicas in
                 exactly one cluster and invoking additional
                 transactions for member clusters. Primary copies (one
                 from each cluster) are updated by a cross-cluster
                 transaction. Then each cluster is independently updated
                 by a separate transaction. This decoupled update
                 propagation process results in possible multiple views
                 of replicated data in a cluster. Compared to other
                 replicated data management protocols, the proposed
                 protocol has several unique advantages. First, thanks
                 to a smaller number of replicas each transaction needs
                 to atomically update in a cluster, the protocol
                 significantly reduces the transaction abort rate, which
                 tends to soar in large transactional systems. Second,
                 the protocol improves user-level transaction response
                 time as top-level update transactions are allowed to
                 commit before all replicas have been updated. Third,
                 read-only queries have the flexibility to see database
                 views of different degrees of consistency and data
                 currency. This ranges from global, most up to date, and
                 consistent views, to local, consistent, but potentially
                 old views, to local, nearest to users but potentially
                 inconsistent views. Fourth, the protocol maintains its
                 scalability by allowing dynamic system reconfiguration
                 as it grows by splitting a cluster into two or more
                 smaller ones. Fifth, autonomy of the clusters is
                 preserved as no specific protocol is required to update
                 replicas within the same cluster. Clusters are,
                 therefore, free to use any valid replication or
                 concurrency control protocols.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Management;
                 Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; data replication; design; experimentation;
                 large-scale systems; management; measurement; multiview
                 access; performance",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Distributed databases. {\bf H.2.4} Information
                 Systems, DATABASE MANAGEMENT, Systems, Transaction
                 processing. {\bf C.2.4} Computer Systems Organization,
                 COMPUTER-COMMUNICATION NETWORKS, Distributed Systems,
                 Distributed databases.",
}

@Article{Mehrotra:1998:ECM,
  author =       "Sharad Mehrotra and Rajeev Rastogi and Henry F. Korth
                 and Abraham Silberschatz",
  title =        "Ensuring consistency in multidatabases by preserving
                 two-level serializability",
  journal =      j-TODS,
  volume =       "23",
  number =       "2",
  pages =        "199--230",
  month =        jun,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-2/p199-mehrotra/p199-mehrotra.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-2/p199-mehrotra/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-2/p199-mehrotra/",
  abstract =     "The concept of serializability has been the
                 traditionally accepted correctness criterion in
                 database systems. However in multidatabase systems
                 (MDBSs), ensuring global serializability is a difficult
                 task. The difficulty arises due to the {\em
                 heterogeneity\/} of the concurrency control protocols
                 used by the participating local database management
                 systems (DBMSs), and the desire to preserve the {\em
                 autonomy\/} of the local DBMSs. In general, solutions
                 to the global serializability problem result in
                 executions with a low degree of concurrency. The
                 alternative, relaxed serializability, may result in
                 data inconsistency.\par

                 In this article, we introduce a systematic approach to
                 relaxing the serializability requirement in MDBS
                 environments. Our approach exploits the structure of
                 the integrity constraints and the nature of transaction
                 programs to ensure consistency without requiring
                 executions to be serializable. We develop a simple yet
                 powerful classification of MDBSs based on the nature of
                 integrity constraints and transaction programs. For
                 each of the identified models we show how consistency
                 can be preserved by ensuring that executions are {\em
                 two-level serializable\/} (2LSR). 2LSR is a correctness
                 criterion for MDBS environments weaker than
                 serializability. What makes our approach interesting is
                 that unlike global serializability, ensuring 2LSR in
                 MDBS environments is relatively simple and protocols to
                 ensure 2LSR permit a high degree of concurrency.
                 Furthermore, we believe the range of models we consider
                 cover many practical MDBS environments to which the
                 results of this article can be applied to preserve
                 database consistency.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Management; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "beyond serializability; concurrency control; database
                 consistency; management; multidatabases; theory",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Concurrency. {\bf H.2.4} Information Systems,
                 DATABASE MANAGEMENT, Systems, Distributed databases.
                 {\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Transaction processing.",
}

@Article{Bertino:1998:ACM,
  author =       "Elisa Bertino and Claudio Bettini and Elena Ferrari
                 and Pierangela Samarati",
  title =        "An access control model supporting periodicity
                 constraints and temporal reasoning",
  journal =      j-TODS,
  volume =       "23",
  number =       "3",
  pages =        "231--285",
  month =        sep,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1998-23-3/p231-bertino/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-3/p231-bertino/",
  abstract =     "Access control models, such as the ones supported by
                 commercial DBMSs, are not yet able to fully meet many
                 application needs. An important requirement derives
                 from the temporal dimension that permissions have in
                 many real-world situations. Permissions are often
                 limited in time or may hold only for specific periods
                 of time. In this article, we present an access control
                 model in which periodic temporal intervals are
                 associated with authorizations. An authorization is
                 automatically granted in the specified intervals and
                 revoked when such intervals expire. Deductive temporal
                 rules with periodicity and order constraints are
                 provided to derive new authorizations based on the
                 presence or absence of other authorizations in specific
                 periods of time. We provide a solution to the problem
                 of ensuring the uniqueness of the global set of valid
                 authorizations derivable at each instant, and we
                 propose an algorithm to compute this set. Moreover, we
                 address issues related to the efficiency of access
                 control by adopting a materialization approach. The
                 resulting model provides a high degree of flexibility
                 and supports the specification of several protection
                 requirements that cannot be expressed in traditional
                 access control models.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "access control; periodic authorization; security;
                 temporal constraints; time management",
  subject =      "{\bf H.2.7} Information Systems, DATABASE MANAGEMENT,
                 Database Administration, Security, integrity, and
                 protection.",
}

@Article{Castano:1998:CSA,
  author =       "S. Castano and V. {De Antonellis} and M. G. Fugini and
                 B. Pernici",
  title =        "Conceptual schema analysis: techniques and
                 applications",
  journal =      j-TODS,
  volume =       "23",
  number =       "3",
  pages =        "286--333",
  month =        sep,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-3/p286-castano/p286-castano.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-3/p286-castano/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-3/p286-castano/",
  abstract =     "The problem of analyzing and classifying conceptual
                 schemas is becoming increasingly important due to the
                 availability of a large number of schemas related to
                 existing applications. The purposes of schema analysis
                 and classification activities can be different: to
                 extract information on intensional properties of legacy
                 systems in order to restructure or migrate to new
                 architectures; to build libraries of reference
                 conceptual components to be used in building new
                 applications in a given domain; and to identify
                 information flows and possible replication of data in
                 an organization. This article proposes a set of
                 techniques for schema analysis and classification to be
                 used separately or in combination. The techniques allow
                 the analyst to derive significant properties from
                 schemas, with human intervention limited as far as
                 possible. In particular, techniques for associating
                 descriptors with schemas, for abstracting reference
                 conceptual schemas based on schema clustering, and for
                 determining schema similarity are presented. A
                 methodology for systematic schema analysis is
                 illustrated, with the purpose of identifying and
                 abstracting into reference components the similar and
                 potentially reusable parts of a set of schemas.
                 Experiences deriving from the application of the
                 proposed techniques and methodology on a large set of
                 Entity-Relationship conceptual schemas of information
                 systems in the Italian Public Administration domain are
                 described",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Documentation; Management",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual modeling; design; documentation;
                 management; reference components; schema
                 classification; schema similarity",
  subject =      "{\bf H.2.1} Information Systems, DATABASE MANAGEMENT,
                 Logical Design. {\bf H.3.1} Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Content Analysis and
                 Indexing.",
}

@Article{Formica:1998:EMC,
  author =       "A. Formica and H. D. Groger and M. Missikoff",
  title =        "An efficient method for checking object-oriented
                 database schema correctness",
  journal =      j-TODS,
  volume =       "23",
  number =       "3",
  pages =        "334--369",
  month =        sep,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-3/p334-formica/p334-formica.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-3/p334-formica/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-3/p334-formica/",
  abstract =     "Inheritance is introduced in object-oriented systems
                 to enhance code reuse and create more compact and
                 readable software. Powerful object models adopt
                 multiple inheritance, allowing a type (or class)
                 definition to inherit from more than one supertype.
                 Unfortunately, in applying this powerful modeling
                 mechanism, inheritance conflicts may be generated,
                 which arise when the same property or operation is
                 defined in more than one supertype. Inheritance
                 conflicts identification and resolution is the key
                 issue of this article. In strongly typed
                 object-oriented systems the resolution of inheritance
                 conflicts depends on the compatibility of the types of
                 the conflicting definitions. In case of incompatible
                 types, a contradiction arises. This article focuses on
                 object-oriented databases (ODBs), providing a method
                 aimed at supporting the designer in the construction of
                 correct ODB schemas. The first necessary condition for
                 schema correctness is the absence of contradictions. A
                 second cause of schema incorrectness is due to the
                 presence of structurally recursive types that, when
                 defined within certain hierarchical patterns, cause the
                 nontermination of the inheritance process. In the
                 article, after the formal definition of a correct
                 schema, two graph-theoretic methods aimed at verifying
                 ODB schema correctness are analyzed. Although the first
                 method is intuitive but inefficient, the second allows
                 schema correctness to be checked in polynomial time, in
                 the size of the schema. The results of this study are
                 included in the implementation of Mosaico, an
                 environment for ODB application design.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Theory; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "databases; design; graph theory; inheritance
                 conflicts; inheritance process; languages;
                 object-oriented database schemas; recursive types;
                 theory; verification",
  subject =      "{\bf H.2.4} Information Systems, DATABASE MANAGEMENT,
                 Systems, Object-oriented databases. {\bf D.3.3}
                 Software, PROGRAMMING LANGUAGES, Language Constructs
                 and Features, Data types and structures. {\bf F.3.1}
                 Theory of Computation, LOGICS AND MEANINGS OF PROGRAMS,
                 Specifying and Verifying and Reasoning about Programs,
                 Mechanical verification. {\bf H.2.3} Information
                 Systems, DATABASE MANAGEMENT, Languages, Data
                 description languages (DDL). {\bf H.2.1} Information
                 Systems, DATABASE MANAGEMENT, Logical Design, Data
                 models. {\bf H.2.1} Information Systems, DATABASE
                 MANAGEMENT, Logical Design, Schema and subschema.",
}

@Article{Konopnicki:1998:IGW,
  author =       "David Konopnicki and Oded Shmueli",
  title =        "Information gathering in the {World-Wide Web}: the
                 {W3QL} query language and the {W3QS} system",
  journal =      j-TODS,
  volume =       "23",
  number =       "4",
  pages =        "369--410",
  month =        dec,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-4/p369-konopnicki/p369-konopnicki.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-4/p369-konopnicki/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-4/p369-konopnicki/",
  abstract =     "The World Wide Web (WWW) is a fast growing global
                 information resource. It contains an enormous amount of
                 information and provides access to a variety of
                 services. Since there is no central control and very
                 few standards of information organization or service
                 offering, searching for information and services is a
                 widely recognized problem. To some degree this problem
                 is solved by ``search services,'' also known as
                 ``indexers,'' such as Lycos, AltaVista, Yahoo, and
                 others. These sites employ search engines known as
                 ``robots'' or ``knowbots'' that scan the network
                 periodically and form text-based indices. These
                 services are limited in certain important aspects.
                 First, the structural information, namely, the
                 organization of the document into parts pointing to
                 each other, is usually lost. Second, one is limited by
                 the kind of textual analysis provided by the ``search
                 service.'' Third, search services are incapable of
                 navigating ``through'' forms. Finally, one cannot
                 prescribe a complex database-like search. We view the
                 WWW as a huge database. We have designed a high-level
                 SQL-like language called W3QL to support effective and
                 flexible query processing, which addresses the
                 structure and content of WWW nodes and their varied
                 sorts of data. We have implemented a system called W3QS
                 to execute W3QL queries. In W3QS, query results are
                 declaratively specified and continuously maintained as
                 views when desired. The current architecture of W3QS
                 provides a server that enables users to pose queries as
                 well as integrate their own data analysis tools. The
                 system and its query language set a framework for the
                 development of database-like tools over the WWW. A
                 significant contribution of this article is in
                 formalizing the WWW and query processing over it.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CGI; design; FORMS; HTML; HTTP; languages; PERL; query
                 language; query system; World-Wide Web",
  subject =      "{\bf H.2.3} Information Systems, DATABASE MANAGEMENT,
                 Languages, Query languages. {\bf H.1.0} Information
                 Systems, MODELS AND PRINCIPLES, General. {\bf H.3.3}
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Information Search and Retrieval.",
}

@Article{Sistla:1998:TTC,
  author =       "A. Prasad Sistla and Ouri Wolfson and Yelena Yesha and
                 Robert Sloan",
  title =        "Towards a theory of cost management for digital
                 libraries and electronic commerce",
  journal =      j-TODS,
  volume =       "23",
  number =       "4",
  pages =        "411--452",
  month =        dec,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-4/p411-sistla/p411-sistla.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-4/p411-sistla/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-4/p411-sistla/",
  abstract =     "One of the features that distinguishes digital
                 libraries from traditional databases is new cost models
                 for client access to intellectual property. Clients
                 will pay for accessing data items in digital libraries,
                 and we believe that optimizing these costs will be as
                 important as optimizing performance in traditional
                 databases. In this article we discuss cost models and
                 protocols for accessing digital libraries, with the
                 objective of determining the minimum cost protocol for
                 each model. We expect that in the future information
                 appliances will come equipped with a cost optimizer, in
                 the same way that computers today come with a built-in
                 operating system. This article makes the initial steps
                 towards a theory and practice of intellectual property
                 cost management.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Economics; Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; average case analysis; caching; cost
                 models; demand; economics; on-line services;
                 performance; protocols; subscription; theory; worst
                 case analysis",
  subject =      "{\bf H.2.m} Information Systems, DATABASE MANAGEMENT,
                 Miscellaneous. {\bf H.3.5} Information Systems,
                 INFORMATION STORAGE AND RETRIEVAL, Online Information
                 Services, Commercial services. {\bf H.3.5} Information
                 Systems, INFORMATION STORAGE AND RETRIEVAL, Online
                 Information Services, Web-based services. {\bf H.3.7}
                 Information Systems, INFORMATION STORAGE AND RETRIEVAL,
                 Digital Libraries, Dissemination.",
}

@Article{Zobel:1998:IFV,
  author =       "Justin Zobel and Alistair Moffat and Kotagiri
                 Ramamohanarao",
  title =        "Inverted files versus signature files for text
                 indexing",
  journal =      j-TODS,
  volume =       "23",
  number =       "4",
  pages =        "453--490",
  month =        dec,
  year =         "1998",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1998-23-4/p453-zobel/p453-zobel.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1998-23-4/p453-zobel/;
                 http://www.acm.org:80/pubs/citations/journals/tods/1998-23-4/p453-zobel/",
  abstract =     "Two well-known indexing methods are inverted files and
                 signature files. We have undertaken a detailed
                 comparison of these two approaches in the context of
                 text indexing, paying particular attention to query
                 evaluation speed and space requirements. We have
                 examined their relative performance using both
                 experimentation and a refined approach to modeling of
                 signature files, and demonstrate that inverted files
                 are distinctly superior to signature files. Not only
                 can inverted files be used to evaluate typical queries
                 in less time than can signature files, but inverted
                 files require less space and provide greater
                 functionality. Our results also show that a synthetic
                 text database can provide a realistic indication of the
                 behavior of an actual text database. The tools used to
                 generate the synthetic database have been made publicly
                 available",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "algorithms; indexing; inverted files; performance;
                 signature files; text databases; text indexing",
  subject =      "{\bf E.5} Data, FILES. {\bf H.2.2} Information
                 Systems, DATABASE MANAGEMENT, Physical Design. {\bf
                 H.3.3} Information Systems, INFORMATION STORAGE AND
                 RETRIEVAL, Information Search and Retrieval. {\bf
                 I.7.3} Computing Methodologies, DOCUMENT AND TEXT
                 PROCESSING, Index Generation**.",
}

@Article{Datta:1999:BPS,
  author =       "Anindya Datta and Debra E. Vandermeer and Aslihan
                 Celik and Vijay Kumar",
  title =        "Broadcast protocols to support efficient retrieval
                 from databases by mobile users",
  journal =      j-TODS,
  volume =       "24",
  number =       "1",
  pages =        "1--79",
  month =        mar,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-1/p1-datta/",
  abstract =     "Mobile computing has the potential for managing
                 information globally. Data management issues in mobile
                 computing have received some attention in recent times,
                 and the design of {\em adaptive broadcast protocols\/}
                 has been posed as an important problem. Such protocols
                 are employed by database servers to decide on the
                 content of broadcasts dynamically, in response to
                 client mobility and demand patterns. In this paper we
                 design such protocols and also propose efficient
                 retrieval strategies that may be employed by clients to
                 download information from broadcasts. The goal is to
                 design {\em cooperative\/} strategies between server
                 and client to provide access to information in such a
                 way as to minimize energy expenditure by clients. We
                 evaluate the performance of our protocols both
                 analytically and through simulation.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "adaptive broadcast protocols; client-server computing;
                 energy conservation; mobile databases",
  subject =      "Computer Systems Organization ---
                 Computer-Communication Networks --- Network
                 Architecture and Design (C.2.1); Information Systems
                 --- Database Management (H.2); Computer Systems
                 Organization --- Computer-Communication Networks ---
                 Network Protocols (C.2.2); Information Systems ---
                 Database Management --- Systems (H.2.4); Computer
                 Systems Organization --- Computer-Communication
                 Networks --- Network Architecture and Design (C.2.1):
                 {\bf Wireless communication}; Information Systems ---
                 Information Systems Applications --- Communications
                 Applications (H.4.3); Information Systems ---
                 Information Systems Applications --- Communications
                 Applications (H.4.3): {\bf Internet}",
}

@Article{Levene:1999:DDI,
  author =       "Mark Levene and George Loizou",
  title =        "Database design for incomplete relations",
  journal =      j-TODS,
  volume =       "24",
  number =       "1",
  pages =        "80--126",
  month =        mar,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-1/p80-levene/",
  abstract =     "Although there has been a vast amount of research in
                 the area of relational database design, to our
                 knowledge, there has been very little work that
                 considers whether this theory is still valid when
                 relations in the database may be incomplete. When
                 relations are incomplete and thus contain null values
                 the problem of whether satisfaction is additive arises.
                 Additivity is the property of the equivalence of the
                 satisfaction of a set of functional dependencies (FDs)
                 F with the individual satisfaction of each member of F
                 in an incomplete relation. It is well known that in
                 general, satisfaction of FDs is not additive.
                 Previously we have shown that satisfaction is additive
                 if and only if the set of FDs is monodependent. We
                 conclude that monodependence is a fundamental desirable
                 property of a set of FDs when considering incomplete
                 information in relational database design. We show
                 that, when the set of FDs F either satisfies the
                 intersection property or the split-freeness property,
                 then the problem of finding an optimum cover of F can
                 be solved in polynomial time in the size of F; in
                 general, this problem is known to be NP-complete. We
                 also show that when F satisfies the split-freeness
                 property then deciding whether there is a superkey of
                 cardinality k or less can be solved in polynomial time
                 in the size of F, since all the keys have the same
                 cardinality. If F only satisfies the intersection
                 property then this problem is NP-complete, as in the
                 general case. Moreover, we show that when F either
                 satisfies the intersection property or the
                 split-freeness property then deciding whether an
                 attribute is prime can be solved in polynomial time in
                 the size of F; in general, this problem is known to be
                 NP-complete. Assume that a relation schema R is an
                 appropriate normal form with respect to a set of FDs F.
                 We show that when F satisfies the intersection property
                 then the notions of second normal form and third normal
                 form are equivalent. We also show that when R is in
                 Boyce-Codd Normal Form (BCNF), then F is monodependent
                 if and only if either there is a unique key for R, or
                 for all keys X for R, the cardinality of X is one less
                 than the number of attributes associated with R.
                 Finally, we tackle a long-standing problem in
                 relational database theory by showing that when a set
                 of FDs F over R satisfies the intersection property, it
                 also satisfies the split-freeness property (i.e., is
                 monodependent), if and only if every lossless join
                 decomposition of R with respect to F is also dependency
                 preserving. As a corollary of this result we are able
                 to show that when F satisfies the intersection
                 property, it also satisfies the intersection property,
                 it also satisfies the split-freeness property(i.e., is
                 monodependent), if and only if every lossless join
                 decomposition of R, which is in BCNF, is also
                 dependency preserving. Our final result is that when F
                 is monodependent, then there exists a unique optimum
                 lossless join decomposition of R, which is in BCNF, and
                 is also dependency preserving. Furthermore, this
                 ultimate decomposition can be attained in polynomial
                 time in the size of F.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "additivity problem; complexity; dependency preserving
                 decomposition; incomplete information; intersection
                 property; lossless join decomposition; monodependence;
                 normal forms; null functional dependencies; optimum
                 cover; prime attribute problem; split-freeness
                 property; superkey of cardinality k problem",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Normal forms}",
}

@Article{Wijsen:1999:TFC,
  author =       "Jef Wijsen",
  title =        "Temporal {FDs} on complex objects",
  journal =      j-TODS,
  volume =       "24",
  number =       "1",
  pages =        "127--176",
  month =        mar,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-1/p127-wijsen/",
  abstract =     "{\em Temporal functional dependencies\/} (TFD) are
                 defined for temporal databases that include object
                 identity. It is argued that object identity can
                 overcome certain semantic difficulties with existing
                 temporal relational data models. Practical applications
                 of TFDs in object bases are discussed. Reasoning about
                 TFDs is at the center of this paper. It turns out that
                 the distinction between acyclic and cyclic schemas is
                 significant. For acyclic schemas, a complete
                 axiomatization for finite implication is given and an
                 algorithm for deciding finite implication provided. The
                 same axiomatization is proven complete for unrestricted
                 implication in unrestricted schemas, which can be
                 cyclic. An interesting result is that there are cyclic
                 schemas for which unrestricted and finite implication
                 do not coincide. TFDs relate and extend some earlier
                 work on dependency theory in temporal databases.
                 Throughout this paper, the construct of TFD is compared
                 with the notion of temporal FD introduced by Wang et
                 al. (1997). A comparison with other related work is
                 provided at the end of the article.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database constraints; functional dependency;
                 object-identity; temporal databases; time granularity",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1)",
}

@Article{Chaudhuri:1999:OQU,
  author =       "Surajit Chaudhuri and Kyuseok Shim",
  title =        "Optimization of queries with user-defined predicates",
  journal =      j-TODS,
  volume =       "24",
  number =       "2",
  pages =        "177--228",
  month =        jun,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-2/p177-chaudhuri/",
  abstract =     "Relational databases provide the ability to store
                 user-defined functions and predicates which can be
                 invoked in SQL queries. When evaluation of a
                 user-defined predicate is relatively expensive, the
                 traditional method of evaluating predicates as early as
                 possible is no longer a sound heuristic. There are two
                 previous approaches for optimizing such queries.
                 However, neither is able to guarantee the optimal plan
                 over the desired execution space. We present efficient
                 techniques that are able to guarantee the choice of an
                 optimal plan over the desired execution space. The {\em
                 optimization algorithm with complete rank-ordering\/}
                 improves upon the naive optimization algorithm by
                 exploiting the nature of the cost formulas for join
                 methods and is polynomial in the number of user-defined
                 predicates (for a given number of relations.) We also
                 propose {\em pruning rules\/} that significantly reduce
                 the cost of searching the execution space for both the
                 naive algorithm as well as for the optimization
                 algorithm with complete rank-ordering, without
                 compromising optimality. We also propose a {\em
                 conservative local heuristic\/} that is simpler and has
                 low optimization overhead. Although it is not always
                 guaranteed to find the optimal plans, it produces close
                 to optimal plans in most cases. We discuss how,
                 depending on application requirements, to determine the
                 algorithm of choice. It should be emphasized that our
                 optimization algorithms handle user-defined selections
                 as well as user-defined join predicates uniformly. We
                 present complexity analysis and experimental comparison
                 of the algorithms.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Management; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "dynamic programming; query optimization; user-defined
                 predicates",
  subject =      "Information Systems --- Database Management (H.2);
                 Information Systems --- Database Management --- General
                 (H.2.0); Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Relational databases}",
}

@Article{Gravano:1999:GTS,
  author =       "Luis Gravano and H{\'e}ctor Garc{\'\i}a-Molina and
                 Anthony Tomasic",
  title =        "{GlOSS}: text-source discovery over the {Internet}",
  journal =      j-TODS,
  volume =       "24",
  number =       "2",
  pages =        "229--264",
  month =        jun,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-2/p229-gravano/",
  abstract =     "The dramatic growth of the Internet has created a new
                 problem for users: location of the relevant sources of
                 documents. This article presents a framework for (and
                 experimentally analyzes a solution to) this problem,
                 which we call the {\em text-source discovery problem}.
                 Our approach consists of two phases. First, each text
                 source exports its contents to a centralized service.
                 Second, users present queries to the service, which
                 returns an ordered list of promising text sources. This
                 article describes {\em GlOSS}, Glossary of Servers
                 Server, with two versions: {\em bGlOSS}, which provides
                 a Boolean query retrieval model, and {\em vGlOSS},
                 which provides a vector-space retrieval model. We also
                 present {\em hGlOSS}, which provides a decentralized
                 version of the system. We extensively describe the
                 methodology for measuring the retrieval effectiveness
                 of these systems and provide experimental evidence,
                 based on actual data, that all three systems are highly
                 effective in determining promising text sources for a
                 given query.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Measurement; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "digital libraries; distributed information retrieval;
                 Internet search and retrieval; text databases",
  subject =      "Information Systems --- Information Storage and
                 Retrieval --- General (H.3.0); Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3); Information Systems ---
                 Information Storage and Retrieval --- Digital Libraries
                 (H.3.7); Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Textual databases};
                 Information Systems --- Information Storage and
                 Retrieval (H.3)",
}

@Article{Hjaltason:1999:DBS,
  author =       "G{\'\i}sli R. Hjaltason and Hanan Samet",
  title =        "Distance browsing in spatial databases",
  journal =      j-TODS,
  volume =       "24",
  number =       "2",
  pages =        "265--318",
  month =        jun,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 21 16:01:19 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-2/p265-hjaltason/",
  abstract =     "We compare two different techniques for browsing
                 through a collection of spatial objects stored in an
                 R-tree spatial data structure on the basis of their
                 distances from an arbitrary spatial query object. The
                 conventional approach is one that makes use of a
                 $k$-nearest neighbor algorithm where $k$ is known prior
                 to the invocation of the algorithm. Thus if $ m k $
                 neighbors are needed, the $k$-nearest neighbor
                 algorithm has to be reinvoked for $m$ neighbors,
                 thereby possibly performing some redundant
                 computations. The second approach is incremental in the
                 sense that having obtained the $k$ nearest neighbors,
                 the $ k + 1 {\em st \/ } $ neighbor can be obtained
                 without having to calculate the $ k + 1 $ nearest
                 neighbors from scratch. The incremental approach is
                 useful when processing complex queries where one of the
                 conditions involves spatial proximity (e.g., the
                 nearest city to Chicago with population greater than a
                 million), in which case a query engine can make use of
                 a pipelined strategy. We present a general incremental
                 nearest neighbor algorithm that is applicable to a
                 large class of hierarchical spatial data structures.
                 This algorithm is adapted to the R-tree and its
                 performance is compared to an existing $k$-nearest
                 neighbor algorithm for R-trees [Rousseopoulos et al.
                 1995]. Experiments show that the incremental nearest
                 neighbor algorithm significantly outperforms the
                 $k$-nearest neighbor algorithm for distance browsing
                 queries in a spatial database that uses the R-tree as a
                 spatial index. Moreover, the incremental nearest
                 neighbor algorithm usually outperforms the $k$-nearest
                 neighbor algorithm when applied to the $k$-nearest
                 neighbor problem for the R-tree, although the
                 improvement is not nearly as large as for distance
                 browsing queries. In fact, we prove informally that at
                 any step in its execution the incremental nearest
                 neighbor algorithm is optimal with respect to the
                 spatial data structure that is employed. Furthermore,
                 based on some simplifying assumptions, we prove that in
                 two dimensions the number of distance computations and
                 leaf nodes accesses made by the algorithm for finding
                 $k$ neighbors is $ O(k + k) $.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "distance browsing; hierarchical spatial data
                 structures; nearest neighbors; R-trees; ranking",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Spatial databases
                 and GIS}; Data --- Data Structures (E.1): {\bf Trees}",
}

@Article{Alagic:1999:TCO,
  author =       "Suad Alag{\'\i}c",
  title =        "Type-checking {OQL} queries in the {ODMG} type
                 systems",
  journal =      j-TODS,
  volume =       "24",
  number =       "3",
  pages =        "319--360",
  month =        sep,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-3/p319-alagic/",
  abstract =     "Several negative results are proved about the ability
                 to type-check queries in the only existing proposed
                 standard for object-oriented databases. The first of
                 these negative results is that it is not possible to
                 type-check OQL queries in the type system underlying
                 the ODMG object model and its definition language ODL.
                 The second negative result is that OQL queries cannot
                 be type-checked in the type system of the Java binding
                 of the ODMG standard either. A solution proposed in
                 this paper is to extend the ODMG object model with
                 explicit support for parametric polymorphism (universal
                 type quantification). These results show that Java
                 cannot be a viable database programming language unless
                 extended with parametric polymorphism. This is why
                 type-checking OQL queries presents no problem for the
                 type system of the C++ binding of the ODMG standard.
                 However, a type system that is strictly more powerful
                 than any of the type systems of the ODMG standard is
                 required in order to properly type ordered collections
                 and indices. The required form of polymorphism is
                 bounded type quantification (constrained genericity)
                 and even F-bounded polymorphism. A further result is
                 that neither static nor the standard dynamic
                 object-oriented type-checking is possible for Java OQL,
                 in spite of the fact that Java OQL combines features of
                 two strongly and mostly statically-typed languages.
                 Contrary to one of the promises of object-oriented
                 database technology, this result shows that the
                 impedance mismatch does not disappear in the ODMG
                 standard. A type-safe reflective technique is proposed
                 for overcoming this mismatch.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Standardization; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "C++; Java; ODMG standard; OQL; parametric
                 polymorphism; type systems",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Data description languages (DDL)}; Information
                 Systems --- Database Management --- Languages (H.2.3):
                 {\bf Database (persistent) programming languages};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Object-oriented databases}; Software ---
                 Programming Languages --- Language Classifications
                 (D.3.2): {\bf Object-oriented languages}; Software ---
                 Programming Languages --- Language Constructs and
                 Features (D.3.3): {\bf Classes and objects}; Software
                 --- Programming Languages --- Language Constructs and
                 Features (D.3.3): {\bf Inheritance}; Software ---
                 Programming Languages --- Language Constructs and
                 Features (D.3.3): {\bf Polymorphism}",
}

@Article{Bozkaya:1999:ILM,
  author =       "Tolga Bozkaya and Meral Ozsoyoglu",
  title =        "Indexing large metric spaces for similarity search
                 queries",
  journal =      j-TODS,
  volume =       "24",
  number =       "3",
  pages =        "361--404",
  month =        sep,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-3/p361-bozkaya/",
  abstract =     "One of the common queries in many database
                 applications is finding approximate matches to a given
                 query item from a collection of data items. For
                 example, given an image database, one may want to
                 retrieve all images that are similar to a given query
                 image. Distance-based index structures are proposed for
                 applications where the distance computations between
                 objects of the data domain are expensive (such as
                 high-dimensional data) and the distance function is
                 metric. In this paper we consider using distance-based
                 index structures for similarity queries on large metric
                 spaces. We elaborate on the approach that uses
                 reference points (vantage points) to partition the data
                 space into spherical shell-like regions in a
                 hierarchical manner. We introduce the multivantage
                 point tree structure (mvp-tree) that uses more than one
                 vantage point to partition the space into spherical
                 cuts at each level. In answering similarity-based
                 queries, the mvp-tree also utilizes the precomputed (at
                 construction time) distances between the data points
                 and the vantage points.\par We summarize the
                 experiments comparing mvp-trees to vp-trees that have a
                 similar partitioning strategy, but use only one vantage
                 point at each level and do not make use of the
                 precomputed distances. Empirical studies show that the
                 mvp-tree outperforms the vp-tree by 20\% to 80\% for
                 varying query ranges and different distance
                 distributions. Next, we generalize the idea of using
                 multiple vantage points and discuss the results of
                 experiments we have made to see how varying the number
                 of vantage points in a node affects performance and how
                 much is gained in performance by making use of
                 precomputed distances. The results show that, after
                 all, it may be best to use a large number of vantage
                 points in an internal node in order to end up with a
                 single directory node and keep as many of the
                 precomputed distances as possible to provide more
                 efficient filtering during search operations. Finally,
                 we provide some experimental results that compare
                 mvp-trees with M-trees, which is a dynamic
                 distance-based index structure for metric domains.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Measurement; Performance;
                 Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  subject =      "Data --- Data Structures (E.1): {\bf Trees};
                 Information Systems --- Information Storage and
                 Retrieval --- Content Analysis and Indexing (H.3.1):
                 {\bf Indexing methods}; Information Systems ---
                 Information Storage and Retrieval --- Information
                 Search and Retrieval (H.3.3): {\bf Search process}",
}

@Article{Casati:1999:SIE,
  author =       "Fabio Casati and Stefano Ceri and Stefano Paraboschi
                 and Guiseppe Pozzi",
  title =        "Specification and implementation of exceptions in
                 workflow management systems",
  journal =      j-TODS,
  volume =       "24",
  number =       "3",
  pages =        "405--451",
  month =        sep,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-3/p405-casati/",
  abstract =     "Although workflow management systems are most
                 applicable when an organization follows standard
                 business processes and routines, any of these processes
                 faces the need for handling exceptions, i.e.,
                 asynchronous and anomalous situations that fall outside
                 the normal control flow.\par

                 In this paper we concentrate upon anomalous situations
                 that, although unusual, are part of the semantics of
                 workflow applications, and should be specified and
                 monitored coherently; in most real-life applications,
                 such exceptions affect a significant fraction of
                 workflow cases. However, very few workflow management
                 systems are integrated with a highly expressive
                 language for specifying this kind of exception and with
                 a system component capable of handling it.\par

                 We present Chimera-Exc, a language for the
                 specification of exceptions for workflows based on
                 detached active rules, and then describe the
                 architecture of a system, called FAR, that implements
                 Chimera-Exc and integrates it with a commercial
                 workflow management system and database server. We
                 discuss the main issues that were solved by our
                 implementation, and report on the performance of FAR.
                 We also discuss design criteria for exceptions in light
                 of the formal properties of their execution. Finally,
                 we focus on the portability of FAR on its unbundling to
                 a generic architecture with detached active rules.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Languages; Management; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active rules; asynchronous events; exceptions;
                 workflow management systems",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Rule-based databases};
                 Information Systems --- Information Systems
                 Applications --- General (H.4.0)",
}

@Article{Dey:1999:IDD,
  author =       "Debabrata Dey and Veda C. Storey and Terence M.
                 Barron",
  title =        "Improving database design through the analysis of
                 relationships",
  journal =      j-TODS,
  volume =       "24",
  number =       "4",
  pages =        "453--486",
  month =        dec,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/1999-24-4/p453-dey/p453-dey.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/1999-24-4/p453-dey/",
  abstract =     "Much of the work on conceptual modeling involves the
                 use of an entity-relationship model in which binary
                 relationships appear as associations between two
                 entities. Relationships involving more than two
                 entities are considered rare and, therefore, have not
                 received adequate attention. This research provides a
                 general framework for the analysis of relationships in
                 which binary relationships simply become a special
                 case. The framework helps a designer to identify
                 ternary and other higher-degree relationships that are
                 commonly represented, often inappropriately, as either
                 entities or binary relationships. Generalized rules are
                 also provided for representing higher-degree
                 relationships in the relational model. This uniform
                 treatment of relationships should significantly ease
                 the burden on a designer by enabling him or her to
                 extract more information from a real-world situation
                 and represent it properly in a conceptual design.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual model; ER model; integrity constraint;
                 min-max cardinality; relationship degree; weak
                 relationship",
  subject =      "Information Systems --- Models and Principles ---
                 General (H.1.0); Information Systems --- Database
                 Management --- Logical Design (H.2.1): {\bf Data
                 models}; Information Systems --- Database Management
                 --- Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}",
}

@Article{Muralidhar:1999:SRD,
  author =       "Krishnamurty Muralidhar and Rathindra Sarathy",
  title =        "Security of random data perturbation methods",
  journal =      j-TODS,
  volume =       "24",
  number =       "4",
  pages =        "487--493",
  month =        dec,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-4/p487-muralidhar/",
  abstract =     "Statistical databases often use random data
                 perturbation (RDP) methods to protect against
                 disclosure of confidential numerical attributes. One of
                 the key requirements of RDP methods is that they
                 provide the appropriate level of security against
                 snoopers who attempt to obtain information on
                 confidential attributes through statistical inference.
                 In this study, we evaluate the security provided by
                 three methods of perturbation. The results of this
                 study allow the database administrator to select the
                 most effective RDP method that assures adequate
                 protection against disclosure of confidential
                 information.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Measurement; Security",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bias; covariance; noise addition; random data
                 perturbation",
  subject =      "Information Systems --- Database Management ---
                 Database Administration (H.2.7): {\bf Security,
                 integrity, and protection}; Information Systems ---
                 Database Management --- Database Applications (H.2.8):
                 {\bf Statistical databases}",
}

@Article{Wand:1999:OAR,
  author =       "Yair Wand and Veda C. Storey and Ron Weber",
  title =        "An ontological analysis of the relationship construct
                 in conceptual modeling",
  journal =      j-TODS,
  volume =       "24",
  number =       "4",
  pages =        "494--528",
  month =        dec,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-4/p494-wand/",
  abstract =     "Conceptual models or semantic data models were
                 developed to capture the meaning of an application
                 domain as perceived by someone. Moreover, concepts
                 employed in semantic data models have recently been
                 adopted in object-oriented approaches to systems
                 analysis and design. To employ conceptual modeling
                 constructs effectively, their meanings have to be
                 defined rigorously. Often, however, rigorous
                 definitions of these constructs are missing. This
                 situation occurs especially in the case of the
                 relationship construct. Empirical evidence shows that
                 use of relationships is often problematical as a way of
                 communicating the meaning of an application domain. For
                 example, users of conceptual modeling methodologies are
                 frequently confused about whether to show an
                 association between things via a relationship, an
                 entity, or an attribute. Because conceptual models are
                 intended to capture knowledge about a real-world
                 domain, we take the view that the meaning of modeling
                 constructs should be sought in models of reality.
                 Accordingly, we use ontology, which is the branch of
                 philosophy dealing with models of reality, to analyze
                 the meaning of common conceptual modeling constructs.
                 Our analysis provides a precise definition of several
                 conceptual modeling constructs. Based on our analysis,
                 we derive rules for the use of relationships in
                 entity-relationship conceptual modeling. Moreover, we
                 show how the rules resolve ambiguities that exist in
                 current practice and how they can enrich the capacity
                 of an entity-relationship conceptual model to capture
                 knowledge about an application domain.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conceptual modeling; database design;
                 entity-relationship model; object-oriented modeling;
                 ontology; semantic data modeling",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Computing
                 Milieux --- Management of Computing and Information
                 Systems --- Project and People Management (K.6.1): {\bf
                 Systems analysis and design}",
}

@Article{Yan:1999:SID,
  author =       "Tak W. Yan and Hector Garcia-Molina",
  title =        "The {SIFT} information dissemination system",
  journal =      j-TODS,
  volume =       "24",
  number =       "4",
  pages =        "529--565",
  month =        dec,
  year =         "1999",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Sep 26 08:44:02 MDT 2000",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/1999-24-4/p529-yan/",
  abstract =     "Information dissemination is a powerful mechanism for
                 finding information in wide-area environments. An
                 information dissemination server accepts long-term user
                 queries, collects new documents from information
                 sources, matches the documents against the queries, and
                 continuously updates the users with relevant
                 information. This paper is a retrospective of the
                 Stanford Information Filtering Service (SIFT), a system
                 that as of April 1996 was processing over 40,000
                 worldwide subscriptions and over 80,000 daily
                 documents. The paper describes some of the indexing
                 mechanisms that were developed for SIFT, as well as the
                 evaluations that were conducted to select a scheme to
                 implement. It also describes the implementation of
                 SIFT, and experimental results for the actual system.
                 Finally, it also discusses and experimentally evaluates
                 techniques for distributing a service such as SIFT for
                 added performance and availability.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Boolean queries; dissemination; filtering; indexing;
                 vector space queries",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4); Information Systems --- Information
                 Storage and Retrieval --- Information Storage (H.3.2);
                 Information Systems --- Information Storage and
                 Retrieval --- Information Search and Retrieval (H.3.3);
                 Information Systems --- Information Storage and
                 Retrieval --- Systems and Software (H.3.4)",
}

@Article{Morris:19xx:DON,
  author =       "K. Morris and J. D. Ullman and A. VanGelder",
  title =        "Design Overview of the {NAIL!} System",
  journal =      j-TODS,
  volume =       "??",
  number =       "??",
  pages =        "??--??",
  month =        "????",
  year =         "19xx",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Also published in/as: Proc. of International
                 Conference on Logic Programming, BCS 3, 1986. Also
                 published in/as: Stanford Un., CSD, TR-CS-86-1108.",
  annote =       "The {NAIL!} System seems to be a much more powerful
                 query language than the ones commercially available
                 today. It adds the power and dexterity of Prolog-like
                 logic to standard query techniques. The {NAIL!} System
                 exhibits a tendency to swing the database community
                 from object-oriented query languages back to
                 value-oriented query languages. There seemed to be
                 ambiguity as to how to handle recursive rules. The
                 paper gave some techniques but didn't prefer one over
                 the other. Overall, the {NAIL!} System appears to be a
                 superior attempt at strengthening conventional database
                 query operations.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Spiegler:19xx:DRA,
  author =       "I. Spiegler and Y. Noff",
  title =        "Dynamic Recovery as an Alternative to Data Base
                 Restoration",
  journal =      j-TODS,
  volume =       "??",
  number =       "??",
  pages =        "??--??",
  month =        "????",
  year =         "19xx",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibsource =    "Database/Wiederhold.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Submitted, March 1987.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Guting:2000:FRQ,
  author =       "Ralf Hartmut G{\"u}ting and Michael H. B{\"o}hlen and
                 Martin Erwig and Christian S. Jensen and Nikos A.
                 Lorentzos and Markus Schneider and Michalis
                 Vazirgiannis",
  title =        "A foundation for representing and querying moving
                 objects",
  journal =      j-TODS,
  volume =       "25",
  number =       "1",
  pages =        "1--42",
  month =        mar,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-1/p1-guting/p1-guting.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-1/p1-guting/",
  abstract =     "Spatio-temporal databases deal with geometries
                 changing over time. The goal of our work is to provide
                 a DBMS data model and query language capable of
                 handling such time-dependent geometries, including
                 those changing continuously that describe {\em moving
                 objects}. Two fundamental abstractions are {\em moving
                 point\/} and {\em moving region}, describing objects
                 for which only the time-dependent position, or position
                 and extent, respectively, are of interest. We propose
                 to present such time-dependent geometries as attribute
                 data types with suitable operations, that is, to
                 provide an abstract data type extension to a DBMS data
                 model and query language. This paper presents a design
                 of such a system of abstract data types. It turns out
                 that besides the main types of interest, moving point
                 and moving region, a relatively large number of
                 auxiliary data types are needed. For example, one needs
                 a line type to represent the projection of a moving
                 point into the plane, or a ``moving real'' to represent
                 the time-dependent distance of two points. It then
                 becomes crucial to achieve (i) orthogonality in the
                 design of the system, i.e., type constructors can be
                 applied uniformly; (ii) genericity and consistency of
                 operations, i.e., operations range over as many types
                 as possible and behave consistently; and (iii) closure
                 and consistency between structure and operations of
                 nontemporal and related temporal types. Satisfying
                 these goal leads to a simple and expressive system of
                 abstract data types that may be integrated into a query
                 language to yield a powerful language for querying
                 spatio-temporal data, including moving objects. The
                 paper formally defines the types and operations, offers
                 detailed insight into the considerations that went into
                 the design, and exemplifies the use of the abstract
                 data types using SQL. The paper offers a precise and
                 conceptually clean foundation for implementing a
                 spatio-temporal DBMS extension.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "abstract data types; algebra; moving objects; moving
                 point; moving region; spatio-temporal data types;
                 spatio-temporal databases",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Query languages}; Information
                 Systems --- Database Management --- Database
                 Applications (H.2.8): {\bf Spatial databases and GIS}",
}

@Article{Kossmann:2000:IDP,
  author =       "Donald Kossmann and Konrad Stocker",
  title =        "Iterative dynamic programming: a new class of query
                 optimization algorithms",
  journal =      j-TODS,
  volume =       "25",
  number =       "1",
  pages =        "43--82",
  month =        mar,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-1/p43-kossmann/p43-kossmann.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-1/p43-kossmann/",
  abstract =     "The query optimizer is one of the most important
                 components of a database system. Most commercial query
                 optimizers today are based on a dynamic-programming
                 algorithm, as proposed in Selinger et al. [1979]. While
                 this algorithm produces good optimization results (i.e,
                 good plans), its high complexity can be prohibitive if
                 complex queries need to be processed, new query
                 execution techniques need to be integrated, or in
                 certain programming environments (e.g., distributed
                 database systems). In this paper, we present and
                 thoroughly evaluate a new class of query optimization
                 algorithms that are based on a principle that we call
                 {\em iterative dynamic programming}, or IDP for short.
                 IDP has several important advantages: First,
                 IDP-algorithms produce the best plans of all known
                 algorithms in situations in which dynamic programming
                 is not viable because of its high complexity. Second,
                 some IDP variants are adaptive and produce as good
                 plans as dynamic programming if dynamic programming is
                 viable and as good-as possible plans if dynamic
                 programming turns out to be not viable. Three, all
                 IDP-algorithms can very easily be integrated into an
                 existing optimizer which is based on dynamic
                 programming.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "dynamic programming; greedy algorithm; iterative
                 dynamic programming; plan evaluation function; query
                 optimization; query optimiztion; randomized
                 optimization",
  subject =      "Theory of Computation --- Analysis of Algorithms and
                 Problem Complexity --- Nonnumerical Algorithms and
                 Problems (F.2.2); Information Systems --- Database
                 Management --- Systems (H.2.4): {\bf Distributed
                 databases}; Information Systems --- Database Management
                 --- Systems (H.2.4): {\bf Relational databases}",
}

@Article{Lerner:2000:MCT,
  author =       "Barbara Staudt Lerner",
  title =        "A model for compound type changes encountered in
                 schema evolution",
  journal =      j-TODS,
  volume =       "25",
  number =       "1",
  pages =        "83--127",
  month =        mar,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-1/p83-lerner/p83-lerner.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-1/p83-lerner/",
  abstract =     "Schema evolution is a problem that is faced by
                 long-lived data. When a schema changes, existing
                 persistent data can become inaccessible unless the
                 database system provides mechanisms to access data
                 created with previous versions of the schema. Most
                 existing systems that support schema evolution focus on
                 changes local to individual types within the schema,
                 thereby limiting the changes that the database
                 maintainer can perform. We have developed a model of
                 type changes involving multiple types. The model
                 describes both type changes and their impact on data by
                 defining derivation rules to initialize new data based
                 on the existing data. The derivation rules can describe
                 local and nonlocal changes to types to capture the
                 intent of a large class of type change operations. We
                 have built a system called Tess (Type Evolution
                 Software System) that uses this model to recognize type
                 changes by comparing schemas and then produces a
                 transformer that can update data in a database to
                 correspond to a newer version of the schema.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Languages",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "persistent programming languages; schema evolution",
  subject =      "Information Systems --- Database Management ---
                 Miscellaneous (H.2.m); Information Systems --- Database
                 Management --- Languages (H.2.3): {\bf Database
                 (persistent) programming languages}; Software ---
                 Software Engineering --- Distribution, Maintenance, and
                 Enhancement (D.2.7): {\bf Restructuring, reverse
                 engineering, and reengineering}",
}

@Article{Bohm:2000:CMQ,
  author =       "Christian B{\"o}hm",
  title =        "A cost model for query processing in high dimensional
                 data spaces",
  journal =      j-TODS,
  volume =       "25",
  number =       "2",
  pages =        "129--178",
  month =        jun,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-2/p129-bohm/p129-bohm.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-2/p129-bohm/",
  abstract =     "During the last decade, multimedia databases have
                 become increasingly important in many application areas
                 such as medicine, CAD, geography, and molecular
                 biology. An important research topic in multimedia
                 databases is similarity search in large data sets. Most
                 current approaches that address similarity search use
                 the feature approach, which transforms important
                 properties of the stored objects into points of a
                 high-dimensional space (feature vectors). Thus,
                 similarity search is transformed into a neighborhood
                 search in feature space. Multidimensional index
                 structures are usually applied when managing feature
                 vectors. Query processing can be improved substantially
                 with optimization techniques such as blocksize
                 optimization, data space quantization, and dimension
                 reduction. To determine optimal parameters, an accurate
                 estimate of index-based query processing performance is
                 crucial. In this paper we develop a cost model for
                 index structures for point databases such as the
                 R*-tree and the X-tree. It provides accurate estimates
                 of the number of data page accesses for range queries
                 and nearest-neighbor queries under a Euclidean metric
                 and a maximum metric and a maximum metric. The problems
                 specific to high-dimensional data spaces, called
                 boundary effects, are considered. The concept of the
                 fractal dimension is used to take the effects of
                 correlated data into account.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Performance; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cost model; multidimensional index",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8); Information Systems ---
                 Information Storage and Retrieval --- Content Analysis
                 and Indexing (H.3.1)",
}

@Article{Cui:2000:TLV,
  author =       "Yingwei Cui and Jennifer Widom and Janet L. Wiener",
  title =        "Tracing the lineage of view data in a warehousing
                 environment",
  journal =      j-TODS,
  volume =       "25",
  number =       "2",
  pages =        "179--227",
  month =        jun,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-2/p179-cui/p179-cui.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-2/p179-cui/",
  abstract =     "We consider the {\em view data lineage\/} problem in a
                 warehousing environment: For a given data item in a
                 materialized warehouse view, we want to identify the
                 set of source data items that produced the view item.
                 We formally define the lineage problem, develop lineage
                 tracing algorithms for relational views with
                 aggregation, and propose mechanisms for performing
                 consistent lineage tracing in a multisource data
                 warehousing environment. Our result can form the basis
                 of a tool that allows analysts to browse warehouse
                 data, select view tuples of interest, and then
                 ``drill-through'' to examine the exact source tuples
                 that produced the view tuples of interest.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Design",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data warehouse; derivation; lineage; materialized
                 views",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Database Manager}",
}

@Article{Parsons:2000:EIT,
  author =       "Jeffrey Parsons and Yair Wand",
  title =        "Emancipating instances from the tyranny of classes in
                 information modeling",
  journal =      j-TODS,
  volume =       "25",
  number =       "2",
  pages =        "228--268",
  month =        jun,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-2/p228-parsons/p228-parsons.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-2/p228-parsons/",
  abstract =     "Database design commonly assumes, explicitly or
                 implicitly, that instances must belong to classes. This
                 can be termed the {\em assumption of inherent
                 classification}. We argue that the extent and
                 complexity of problems in schema integration, schema
                 evolution, and interoperability are, to a large degree,
                 consequences of inherent classification. Furthermore,
                 we make the case that the assumption of inherent
                 classification violates philosophical and cognitive
                 guidelines on classification and is, therefore,
                 inappropriate in view of the role of data modeling in
                 representing knowledge about application domains.
                 \par

                 As an alternative, we propose a layered approach to
                 modeling in which information about instances is
                 separated from any particular classification. Two data
                 modeling layers are proposed: (1) an {\em instance
                 model\/} consisting of an instance base (i.e.,
                 information about instances and properties) and
                 operations to populate, use, and maintain it; and (2) a
                 {\em class model\/} consisting of a class base (i.e.,
                 information about classes defined in terms of
                 properties) and operations to populate, use, and
                 maintain it. The two-layered model provides {\em class
                 independence}. This is analogous to the arguments of
                 data independence offered by the relational model in
                 comparison to hierarchical and network models. We show
                 that a two-layered approach yields several advantages.
                 In particular, schema integration is shown to be
                 partially an artifact of inherent classification that
                 can be greatly simplified in designing a database based
                 on a layered model; schema evolution is supported
                 without the complexity of operations currently required
                 by class-based models; and the difficulties associated
                 with interoperability among heterogeneous databases are
                 reduced because there is no need to agree on the
                 semantics of classes among independent databases. We
                 conclude by considering the adequacy of a two-layered
                 approach, outlining possible implementation strategies,
                 and drawing attention to some practical
                 considerations.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Management; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "classification; conceptual modeling; database design;
                 interoperability; ontology; schema evolution; schema
                 integration",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1): {\bf Data models}; Information
                 Systems --- Database Management --- Logical Design
                 (H.2.1): {\bf Schema and subschema}; Information
                 Systems --- Database Management --- Heterogeneous
                 Databases (H.2.5): {\bf Data translation**};
                 Information Systems --- Database Management ---
                 Heterogeneous Databases (H.2.5)",
}

@Article{Baralis:2000:AAS,
  author =       "Elena Baralis and Jennifer Widom",
  title =        "An algebraic approach to static analysis of active
                 database rules",
  journal =      j-TODS,
  volume =       "25",
  number =       "3",
  pages =        "269--332",
  month =        sep,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/2000-25-3/p269-baralis/",
  abstract =     "Rules in active database systems can be very difficult
                 to program due to the unstructured and unpredictable
                 nature of rule processing. We provide static analysis
                 techniques for predicting whether a given rule set is
                 guaranteed to terminate and whether rule execution is
                 confluent (guaranteed to have a unique final state).
                 Our methods are based on previous techniques for
                 analyzing rules in active database systems. We improve
                 considerably on the previous techniques by providing
                 analysis criteria that are much less conservative: our
                 methods often determine that a rule set will terminate
                 or is confluent when previous methods could not make
                 this determination. Our improved analysis is based on a
                 ``propagation'' algorithm, which uses an extended
                 relational algebra to accurately determine when the
                 action of one rule can affect the condition of another,
                 and determine when rule actions commute. We consider
                 both condition-action rules and
                 event-condition-action-rules, making our approach
                 widely applicable to relational active database rule
                 languages and to the trigger language in the SQL:1999
                 standard.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Verification",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "active database systems; confluence; database rule
                 processing; database trigger processing; termination",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Rule-based databases};
                 Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf SQL}",
}

@Article{Kemme:2000:NAD,
  author =       "Bettina Kemme and Gustavo Alonso",
  title =        "A new approach to developing and implementing eager
                 database replication protocols",
  journal =      j-TODS,
  volume =       "25",
  number =       "3",
  pages =        "333--379",
  month =        sep,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/2000-25-3/p333-kemme/",
  abstract =     "Database replication is traditionally seen as a way to
                 increase the availability and performance of
                 distributed databases. Although a large number of
                 protocols providing data consistency and
                 fault-tolerance have been proposed, few of these ideas
                 have ever been used in commercial products due to their
                 complexity and performance implications. Instead,
                 current products allow inconsistencies and often resort
                 to centralized approaches which eliminates some of the
                 advantages of replication. As an alternative, we
                 propose a suite of replication protocols that addresses
                 the main problems related to database replication. On
                 the one hand, our protocols maintain data consistency
                 and the same transactional semantics found in
                 centralized systems. On the other hand, they provide
                 flexibility and reasonable performance. To do so, our
                 protocols take advantage of the rich semantics of group
                 communication primitives and the relaxed isolation
                 guarantees provided by most databases. This allows us
                 to eliminate the possibility of deadlocks, reduce the
                 message overhead and increase performance. A detailed
                 simulation study shows the feasibility of the approach
                 and the flexibility with which different types of
                 bottlenecks can be circumvented.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Management; Performance; Reliability",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database replication; fault-tolerance; group
                 communication; isolation levels;
                 one-copy-serializability; replica control; total error
                 multicast",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Concurrency}; Information Systems
                 --- Database Management --- Systems (H.2.4): {\bf
                 Distributed databases}; Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Transaction processing}; Computer Systems Organization
                 --- Computer-Communication Networks --- Distributed
                 Systems (C.2.4); Computer Systems Organization ---
                 Performance of Systems (C.4)",
}

@Article{Meo:2000:TDV,
  author =       "Rosa Meo",
  title =        "Theory of dependence values",
  journal =      j-TODS,
  volume =       "25",
  number =       "3",
  pages =        "380--406",
  month =        sep,
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Apr 14 10:34:48 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/citations/journals/tods/2000-25-3/p380-meo/",
  abstract =     "A new model to evaluate dependencies in data mining
                 problems is presented and discussed. The well-known
                 concept of the association rule is replaced by the new
                 definition of dependence value, which is a single real
                 number uniquely associated with a given itemset.
                 Knowledge of dependence values is sufficient to
                 describe all the dependencies characterizing a given
                 data mining problem. The dependence value of an itemset
                 is the difference between the occurrence probability of
                 the itemset and a corresponding ``maximum independence
                 estimate.'' This can be determined as a function of
                 joint probabilities of the subsets of the itemset being
                 considered by maximizing a suitable entropy function.
                 So it is possible to separate in an itemset of
                 cardinality $k$ the dependence inherited from its
                 subsets of cardinality ($ k 1 $ ) and the specific
                 inherent dependence of that itemset. The absolute value
                 of the difference between the probability p($i$ ) of
                 the event $i$ that indicates the presence of the
                 itemset $ \{ a, b, \ldots {} \} $ and its maximum
                 independence estimate is constant for any combination
                 of values of $ Q a, b, \ldots {} Q $. In addition, the
                 Boolean function specifying the combination of values
                 for which the dependence is positive is a parity
                 function. So the determination of such combinations is
                 immediate. The model appears to be simple and
                 powerful.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Experimentation; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "association rules; dependence rules; entropy;
                 variables independence",
  subject =      "Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Data mining};
                 Information Systems --- Database Management ---
                 Database Applications (H.2.8): {\bf Statistical
                 databases}; Information Systems --- Models and
                 Principles --- Systems and Information Theory (H.1.1):
                 {\bf Information theory}; Computing Methodologies ---
                 Artificial Intelligence --- Knowledge Representation
                 Formalisms and Methods (I.2.4)",
}

@Article{Bohlen:2000:TSM,
  author =       "Michael H. B{\"o}hlen and Christian S. Jensen and
                 Richard T. Snodgrass",
  title =        "Temporal statement modifiers",
  journal =      j-TODS,
  volume =       "25",
  number =       "4",
  pages =        "407--456",
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 26 08:20:52 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-4/p407-bohlen/p407-bohlen.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-4/p407-bohlen/",
  abstract =     "A wide range of database applications manage
                 time-varying data. Many temporal query languages have
                 been proposed, each one the result of many carefully
                 made yet subtly interacting design decisions. In this
                 article we advocate a different approach to
                 articulating a set of requirements, or desiderata, that
                 directly imply the syntactic structure and core
                 semantics of a temporal extension of an (arbitrary)
                 nontemporal query language. These desiderata facilitate
                 transitioning applications from a nontemporal query
                 language and data model, which has received only scant
                 attention thus far. \par

                 The paper then introduces the notion of {\em statement
                 modifiers\/} that provide a means of systematically
                 adding temporal support to an existing query language.
                 Statement modifiers apply to all query language
                 statements, for example, queries, cursor definitions,
                 integrity constraints, assertions, views, and data
                 manipulation statements. We also provide a way to
                 systematically add temporal support to an existing
                 implementation. The result is a temporal query language
                 syntax, semantics, and implementation that derives from
                 first principles. \par

                 We exemplify this approach by extending SQL-92 with
                 statement modifiers. This extended language, termed
                 ATSQL, is formally defined via a
                 denotational-semantics-style mapping of temporal
                 statements to expressions using a combination of
                 temporal and conventional relational algebraic
                 operators.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Languages; Theory",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "ATSQL; statement modifiers; temporal databases",
  subject =      "Information Systems --- Database Management ---
                 Languages (H.2.3): {\bf Data manipulation languages
                 (DML)}; Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Query processing}; Information
                 Systems --- Database Management --- Systems (H.2.4):
                 {\bf Relational databases}; Information Systems ---
                 Database Management --- Languages (H.2.3): {\bf Data
                 description languages (DDL)}",
}

@Article{Fegaras:2000:OOQ,
  author =       "Leonidas Fegaras and David Maier",
  title =        "Optimizing object queries using an effective
                 calculus",
  journal =      j-TODS,
  volume =       "25",
  number =       "4",
  pages =        "457--516",
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 26 08:20:52 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-4/p457-fegaras/p457-fegaras.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-4/p457-fegaras/",
  abstract =     "Object-oriented databases (OODBs) provide powerful
                 data abstractions and modeling facilities, but they
                 generally lack a suitable framework for query
                 processing and optimization. The development of an
                 effective query optimizer is one of the key factors for
                 OODB systems to successfully compete with relational
                 systems, as well as to meet the performance
                 requirements of many nontraditional applications. We
                 propose an effective framework with a solid theoretical
                 basis for optimizing OODB query languages. Our
                 calculus, called the monoid comprehension calculus,
                 captures most features of ODMG OQL, and is a good basis
                 for expressing various optimization algorithms
                 concisely. This article concentrates on query unnesting
                 (also known as query decorrelation), an optimization
                 that, even though it improves performance considerably,
                 is not treated properly (if at all) by most OODB
                 systems. Our framework generalizes many unnesting
                 techniques proposed recently in the literature, and is
                 capable of removing any form of query nesting using a
                 very simple and efficient algorithm. The simplicity of
                 our method is due to the use of the monoid
                 comprehension calculus as an intermediate form for OODB
                 queries. The monoid comprehension calculus treats
                 operations over multiple collection types, aggregates,
                 and quantifiers in a similar way, resulting in a
                 uniform method of unnesting queries, regardless of
                 their type of nesting.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Design; Experimentation; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "nested relations; object-oriented databases; query
                 decorrelation; query optimization",
  subject =      "Information Systems --- Database Management ---
                 Logical Design (H.2.1); Information Systems ---
                 Database Management --- Systems (H.2.4): {\bf
                 Object-oriented databases}",
}

@Article{Kossmann:2000:CII,
  author =       "Donald Kossmann and Michael J. Franklin and Gerhard
                 Drasch and Wig Ag",
  title =        "Cache investment: integrating query optimization and
                 distributed data placement",
  journal =      j-TODS,
  volume =       "25",
  number =       "4",
  pages =        "517--558",
  year =         "2000",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 26 08:20:52 MDT 2001",
  bibsource =    "http://www.acm.org/pubs/toc/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.acm.org/pubs/articles/journals/tods/2000-25-4/p517-kossmann/p517-kossmann.pdf;
                 http://www.acm.org/pubs/citations/journals/tods/2000-25-4/p517-kossmann/",
  abstract =     "Emerging distributed query-processing systems support
                 flexible execution strategies in which each query can
                 be run using a combination of data shipping and query
                 shipping. As in any distributed environment, these
                 systems can obtain tremendous performance and
                 availability benefits by employing dynamic data
                 caching. When flexible execution and dynamic caching
                 are combined, however, a circular dependency arises:
                 Caching occurs as a by-product of query operator
                 placement, but query operator placement decisions are
                 based on (cached) data location. The practical impact
                 of this dependency is that query optimization decisions
                 that appear valid on a per-query basis can actually
                 cause suboptimal performance for all queries in the
                 long run. \par

                 To address this problem, we developed {\em Cache
                 Investment\/} - a novel approach for integrating query
                 optimization and data placement that looks beyond the
                 performance of a single query. Cache Investment
                 sometimes intentionally generates a ``suboptimal'' plan
                 for a particular query in the interest of effecting a
                 better data placement for subsequent queries. Cache
                 Investment can be integrated into a distributed
                 database system without changing the internals of the
                 query optimizer. In this paper, we propose Cache
                 Investment mechanisms and policies and analyze their
                 performance. The analysis uses results from both an
                 implementation on the SHORE storage manager and a
                 detailed simulation model. Our results show that Cache
                 Investment can significantly improve the overall
                 performance of a system and demonstrate the trade-offs
                 among various alternative policies.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  generalterms = "Algorithms; Performance",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cache investment; caching; client-server database
                 systems; data shipping; dynamic data placement; query
                 optimization; query shipping",
  subject =      "Information Systems --- Database Management ---
                 Systems (H.2.4): {\bf Distributed databases};
                 Information Systems --- Database Management --- Systems
                 (H.2.4): {\bf Relational databases}",
}

@Article{Andries:2001:AUM,
  author =       "Marc Andries and Luca Cabibbo and Jan Paredaens and
                 Jan van den Bussche",
  title =        "Applying an update method to a set of receivers",
  journal =      j-TODS,
  volume =       "26",
  number =       "1",
  pages =        "1--40",
  month =        mar,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:14:50 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dekhtyar:2001:PTD,
  author =       "Alex Dekhtyar and Robert Ross and V. S. Subrahmanian",
  title =        "Probabilistic temporal databases, {I}: algebra",
  journal =      j-TODS,
  volume =       "26",
  number =       "1",
  pages =        "41--95",
  month =        mar,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:14:50 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hsu:2001:RBP,
  author =       "Windsor W. Hsu and Alan Jay Smith and Honesty C.
                 Young",
  title =        "{I/O} reference behavior of production database
                 workloads and the {TPC} benchmarks --- an analysis at
                 the logical level",
  journal =      j-TODS,
  volume =       "26",
  number =       "1",
  pages =        "96--143",
  month =        mar,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:14:50 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chomicki:2001:QAD,
  author =       "Jan Chomicki and David Toman and Michael H.
                 B{\"o}hlen",
  title =        "Querying {ATSQL} databases with temporal logic",
  journal =      j-TODS,
  volume =       "26",
  number =       "2",
  pages =        "145--178",
  month =        jun,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fukuda:2001:DMO,
  author =       "Takeshi Fukuda and Yasuhiko Morimoto and Shimichi
                 Morishita and Takeshi Tokuyama",
  title =        "Data mining with optimized two-dimensional association
                 rules",
  journal =      j-TODS,
  volume =       "26",
  number =       "2",
  pages =        "179--213",
  month =        jun,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jajodia:2001:FSM,
  author =       "Sushil Jajodia and Pierangela Samarati and Maria Luisa
                 Sapino and V. S. Subrahmanian",
  title =        "Flexible support for multiple access control
                 policies",
  journal =      j-TODS,
  volume =       "26",
  number =       "2",
  pages =        "214--260",
  month =        jun,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Eiter:2001:POB,
  author =       "Thomas Eiter and James J. Lu and Thomas Lukasiewicz
                 and V. S. Subrahmanian",
  title =        "Probabilistic object bases",
  journal =      j-TODS,
  volume =       "26",
  number =       "3",
  pages =        "264--312",
  month =        sep,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2001:AQU,
  author =       "Chen Li and Edward Chang",
  title =        "Answering queries with useful bindings",
  journal =      j-TODS,
  volume =       "26",
  number =       "3",
  pages =        "313--343",
  month =        sep,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ng:2001:ERD,
  author =       "Wilfred Ng",
  title =        "An extension of the relational data model to
                 incorporate ordered domains",
  journal =      j-TODS,
  volume =       "26",
  number =       "3",
  pages =        "344--383",
  month =        sep,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:15 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kotidis:2001:CDV,
  author =       "Yannis Kotidis and Nick Roussopoulos",
  title =        "A case for dynamic view management",
  journal =      j-TODS,
  volume =       "26",
  number =       "4",
  pages =        "388--423",
  month =        dec,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:16 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Mamoulis:2001:MSJ,
  author =       "Nikos Mamoulis and Dimitris Papadias",
  title =        "Multiway spatial joins",
  journal =      j-TODS,
  volume =       "26",
  number =       "4",
  pages =        "424--475",
  month =        dec,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:16 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lakshmanan:2001:SES,
  author =       "Laks V. S. Lakshmanan and Fereidoon Sadri and Subbu N.
                 Subramanian",
  title =        "{SchemaSQL}: {An} extension to {SQL} for multidatabase
                 interoperability",
  journal =      j-TODS,
  volume =       "26",
  number =       "4",
  pages =        "476--519",
  month =        dec,
  year =         "2001",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Feb 19 16:12:16 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Suciu:2002:DQE,
  author =       "Dan Suciu",
  title =        "Distributed query evaluation on semistructured data",
  journal =      j-TODS,
  volume =       "27",
  number =       "1",
  pages =        "1--62",
  month =        mar,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schuldt:2002:AIT,
  author =       "Heiko Schuldt and Gustavo Alonso and Catriel Beeri and
                 Hans-J{\"o}rg Schek",
  title =        "Atomicity and isolation for transactional processes",
  journal =      j-TODS,
  volume =       "27",
  number =       "1",
  pages =        "63--116",
  month =        mar,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Liu:2002:LFD,
  author =       "Mengchi Liu and Gillian Dobbie and Tok Wang Ling",
  title =        "A logical foundation for deductive object-oriented
                 databases",
  journal =      j-TODS,
  volume =       "27",
  number =       "1",
  pages =        "117--151",
  month =        mar,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bruno:2002:TSQ,
  author =       "Nicolas Bruno and Surajit Chaudhuri and Luis Gravano",
  title =        "Top-$k$ selection queries over relational databases:
                 {Mapping} strategies and performance evaluation",
  journal =      j-TODS,
  volume =       "27",
  number =       "2",
  pages =        "153--187",
  month =        jun,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chakrabarti:2002:LAD,
  author =       "Kaushik Chakrabarti and Eamonn Keogh and Sharad
                 Mehrotra and Michael Pazzani",
  title =        "Locally adaptive dimensionality reduction for indexing
                 large time series databases",
  journal =      j-TODS,
  volume =       "27",
  number =       "2",
  pages =        "188--228",
  month =        jun,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Calders:2002:SDM,
  author =       "Toon Calders and Raymond T. Ng and Jef Wijsen",
  title =        "Searching for dependencies at multiple abstraction
                 levels",
  journal =      j-TODS,
  volume =       "27",
  number =       "3",
  pages =        "229--260",
  month =        sep,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gibbons:2002:FIM,
  author =       "Phillip B. Gibbons and Yossi Matias and Viswanath
                 Poosala",
  title =        "Fast incremental maintenance of approximate
                 histograms",
  journal =      j-TODS,
  volume =       "27",
  number =       "3",
  pages =        "261--298",
  month =        sep,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tao:2002:CMO,
  author =       "Yufei Tao and Dimitris Papadias and Jun Zhang",
  title =        "Cost models for overlapping and multiversion
                 structures",
  journal =      j-TODS,
  volume =       "27",
  number =       "3",
  pages =        "299--342",
  month =        sep,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:13 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{May:2002:UGS,
  author =       "Wolfgang May and Bertram Lud{\"a}scher",
  title =        "Understanding the global semantics of referential
                 actions using logic rules",
  journal =      j-TODS,
  volume =       "27",
  number =       "4",
  pages =        "343--397",
  month =        dec,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:14 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ciaccia:2002:SMS,
  author =       "Paolo Ciaccia and Marco Patella",
  title =        "Searching in metric spaces with user-defined and
                 approximate distances",
  journal =      j-TODS,
  volume =       "27",
  number =       "4",
  pages =        "398--437",
  month =        dec,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:14 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fernandez:2002:SFP,
  author =       "Mary Fern{\'a}ndez and Yana Kadiyska and Dan Suciu and
                 Atsuyuki Morishima and Wang-Chiew Tan",
  title =        "{SilkRoute}: a framework for publishing relational
                 data in {XML}",
  journal =      j-TODS,
  volume =       "27",
  number =       "4",
  pages =        "438--493",
  month =        dec,
  year =         "2002",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Nov 5 11:23:14 MST 2002",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Beneventano:2003:DLS,
  author =       "Domenico Beneventano and Sonia Bergamaschi and Claudio
                 Sartori",
  title =        "Description logics for semantic query optimization in
                 object-oriented database systems",
  journal =      j-TODS,
  volume =       "28",
  number =       "1",
  pages =        "1--50",
  month =        mar,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Karp:2003:SAF,
  author =       "Richard M. Karp and Scott Shenker and Christos H.
                 Papadimitriou",
  title =        "A simple algorithm for finding frequent elements in
                 streams and bags",
  journal =      j-TODS,
  volume =       "28",
  number =       "1",
  pages =        "51--55",
  month =        mar,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Melnik:2003:AAS,
  author =       "Sergey Melnik and Hector Garcia-Molina",
  title =        "Adaptive algorithms for set containment joins",
  journal =      j-TODS,
  volume =       "28",
  number =       "1",
  pages =        "56--99",
  month =        mar,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tao:2003:SQD,
  author =       "Yufei Tao and Dimitris Papadias",
  title =        "Spatial queries in dynamic environments",
  journal =      j-TODS,
  volume =       "28",
  number =       "2",
  pages =        "101--139",
  month =        jun,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gunopulos:2003:DAM,
  author =       "Dimitrios Gunopulos and Roni Khardon and Heikki
                 Mannila and Sanjeev Saluja and Hannu Toivonen and Ram
                 Sewak Sharma",
  title =        "Discovering all most specific sentences",
  journal =      j-TODS,
  volume =       "28",
  number =       "2",
  pages =        "140--174",
  month =        jun,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lechtenborger:2003:CRV,
  author =       "Jens Lechtenb{\"o}rger and Gottfried Vossen",
  title =        "On the computation of relational view complements",
  journal =      j-TODS,
  volume =       "28",
  number =       "2",
  pages =        "175--208",
  month =        jun,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 7 13:51:37 MDT 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wijesekera:2003:RPF,
  author =       "Duminda Wijesekera and Sushil Jajodia and Francesco
                 Parisi-Presicce and {\AA}sa Hagstr{\"o}m",
  title =        "Removing permissions in the flexible authorization
                 framework",
  journal =      j-TODS,
  volume =       "28",
  number =       "3",
  pages =        "209--229",
  month =        sep,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Oct 31 05:55:40 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jacox:2003:ISJ,
  author =       "Edwin H. Jacox and Hanan Samet",
  title =        "Iterative spatial join",
  journal =      j-TODS,
  volume =       "28",
  number =       "3",
  pages =        "230--256",
  month =        sep,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Oct 31 05:55:40 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jimenez-Peris:2003:QAD,
  author =       "Ricardo Jim{\'e}nez-Peris and M.
                 Pati{\~n}o-Mart{\'\i}nez and Gustavo Alonso and Bettina
                 Kemme",
  title =        "Are quorums an alternative for data replication?",
  journal =      j-TODS,
  volume =       "28",
  number =       "3",
  pages =        "257--294",
  month =        sep,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Oct 31 05:55:40 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tao:2003:APS,
  author =       "Yufei Tao and Jimeng Sun and Dimitris Papadias",
  title =        "Analysis of predictive spatio-temporal queries",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "295--336",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lakshmanan:2003:EDM,
  author =       "Laks V. S. Lakshmanan and Carson Kai-Sang Leung and
                 Raymond T. Ng",
  title =        "Efficient dynamic mining of constrained frequent
                 sets",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "337--389",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cho:2003:EPR,
  author =       "Junghoo Cho and Hector Garcia-Molina",
  title =        "Effective page refresh policies for {Web} crawlers",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "390--426",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chomicki:2003:PFR,
  author =       "Jan Chomicki",
  title =        "Preference formulas in relational queries",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "427--466",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Diao:2003:PSP,
  author =       "Yanlei Diao and Mehmet Altinel and Michael J. Franklin
                 and Hao Zhang and Peter Fischer",
  title =        "Path sharing and predicate evaluation for
                 high-performance {XML} filtering",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "467--516",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hjaltason:2003:IDS,
  author =       "Gisli R. Hjaltason and Hanan Samet",
  title =        "Index-driven similarity search in metric spaces",
  journal =      j-TODS,
  volume =       "28",
  number =       "4",
  pages =        "517--580",
  month =        dec,
  year =         "2003",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Dec 13 18:01:35 MST 2003",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kolaitis:2004:F,
  author =       "Phokion Kolaitis and Michael J. Franklin",
  title =        "Foreword",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "1--1",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Buneman:2004:ASD,
  author =       "Peter Buneman and Sanjeev Khanna and Keishi Tajima and
                 Wang-Chiew Tan",
  title =        "Archiving scientific data",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "2--42",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Garofalakis:2004:PWS,
  author =       "Minos Garofalakis and Phillip B. Gibbons",
  title =        "Probabilistic wavelet synopses",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "43--90",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Grust:2004:AXE,
  author =       "Torsten Grust and Maurice {Van Keulen} and Jens
                 Teubner",
  title =        "Accelerating {XPath} evaluation in any {RDBMS}",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "91--131",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ross:2004:SCM,
  author =       "Kenneth A. Ross",
  title =        "Selection conditions in main memory",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "132--161",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arasu:2004:CMR,
  author =       "Arvind Arasu and Brian Babcock and Shivnath Babu and
                 Jon McAlister and Jennifer Widom",
  title =        "Characterizing memory requirements for queries over
                 continuous data streams",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "162--194",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2004:NFX,
  author =       "Marcelo Arenas and Leonid Libkin",
  title =        "A normal form for {XML} documents",
  journal =      j-TODS,
  volume =       "29",
  number =       "1",
  pages =        "195--232",
  month =        mar,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Amer-Yahia:2004:DAO,
  author =       "Sihem Amer-Yahia and Sophie Cluet",
  title =        "A declarative approach to optimize bulk loading into
                 databases",
  journal =      j-TODS,
  volume =       "29",
  number =       "2",
  pages =        "233--281",
  month =        jun,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sadri:2004:EOS,
  author =       "Reza Sadri and Carlo Zaniolo and Amir Zarkesh and
                 Jafar Adibi",
  title =        "Expressing and optimizing sequence queries in database
                 systems",
  journal =      j-TODS,
  volume =       "29",
  number =       "2",
  pages =        "282--318",
  month =        jun,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Marian:2004:ETQ,
  author =       "Am{\'e}lie Marian and Nicolas Bruno and Luis Gravano",
  title =        "Evaluating top-$k$ queries over {Web}-accessible
                 databases",
  journal =      j-TODS,
  volume =       "29",
  number =       "2",
  pages =        "319--362",
  month =        jun,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yu:2004:CAM,
  author =       "Ting Yu and Divesh Srivastava and Laks V. S.
                 Lakshmanan and H. V. Jagadish",
  title =        "A compressed accessibility map for {XML}",
  journal =      j-TODS,
  volume =       "29",
  number =       "2",
  pages =        "363--402",
  month =        jun,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Datta:2004:PBA,
  author =       "Anindya Datta and Kaushik Dutta and Helen Thomas and
                 Debra Vandermeer and Krithi Ramamritham",
  title =        "Proxy-based acceleration of dynamically generated
                 content on the {World Wide Web}: {An} approach and
                 implementation",
  journal =      j-TODS,
  volume =       "29",
  number =       "2",
  pages =        "403--443",
  month =        jun,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Jul 10 10:03:25 MDT 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Vincent:2004:SFD,
  author =       "Millist W. Vincent and Jixue Liu and Chengfei Liu",
  title =        "Strong functional dependencies and their application
                 to normal forms in {XML}",
  journal =      j-TODS,
  volume =       "29",
  number =       "3",
  pages =        "445--462",
  month =        sep,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Nov 4 08:30:22 MST 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Trajcevski:2004:MUM,
  author =       "Goce Trajcevski and Ouri Wolfson and Klaus Hinrichs
                 and Sam Chamberlain",
  title =        "Managing uncertainty in moving objects databases",
  journal =      j-TODS,
  volume =       "29",
  number =       "3",
  pages =        "463--507",
  month =        sep,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Nov 4 08:30:22 MST 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chaudhuri:2004:EPM,
  author =       "Surajit Chaudhuri and Vivek Narasayya and Sunita
                 Sarawagi",
  title =        "Extracting predicates from mining models for efficient
                 query evaluation",
  journal =      j-TODS,
  volume =       "29",
  number =       "3",
  pages =        "508--544",
  month =        sep,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Nov 4 08:30:22 MST 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Babu:2004:ECR,
  author =       "Shivnath Babu and Utkarsh Srivastava and Jennifer
                 Widom",
  title =        "Exploiting $k$-constraints to reduce memory overhead
                 in continuous queries over data streams",
  journal =      j-TODS,
  volume =       "29",
  number =       "3",
  pages =        "545--580",
  month =        sep,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Nov 4 08:30:22 MST 2004",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ozsoyoglu:2004:QWM,
  author =       "G{\"u}ltekin {\"O}zsoyo{\u{g}}lu and Ismail Seng{\"o}r
                 Alting{\"o}vde and Abdullah Al-Hamdani and Selma
                 Ay{\c{s}}e {\"O}zel and {\"O}zg{\"u}r Ulusoy and Zehra
                 Meral {\"o}zsoyo{\u{g}}lu",
  title =        "Querying {Web} metadata: {Native} score management and
                 text support in databases",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "581--634",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Baralis:2004:ECR,
  author =       "Elena Baralis and Silvia Chiusano",
  title =        "Essential classification rule sets",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "635--674",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chen:2004:MBV,
  author =       "Songting Chen and Bin Liu and Elke A. Rundensteiner",
  title =        "Multiversion-based view maintenance over distributed
                 data sources",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "675--709",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Balmin:2004:IVX,
  author =       "Andrey Balmin and Yannis Papakonstantinou and Victor
                 Vianu",
  title =        "Incremental validation of {XML} documents",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "710--751",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Green:2004:PXS,
  author =       "Todd J. Green and Ashish Gupta and Gerome Miklau and
                 Makoto Onizuka and Dan Suciu",
  title =        "Processing {XML} streams with deterministic automata
                 and stream indexes",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "752--788",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Samet:2004:DPG,
  author =       "Hanan Samet",
  title =        "Decoupling partitioning and grouping: {Overcoming}
                 shortcomings of spatial indexing with bucketing",
  journal =      j-TODS,
  volume =       "29",
  number =       "4",
  pages =        "789--830",
  month =        dec,
  year =         "2004",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 14 10:39:39 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Milo:2005:EIX,
  author =       "Tova Milo and Serge Abiteboul and Bernd Amann and Omar
                 Benjelloun and Fred Dang Ngoc",
  title =        "Exchanging intensional {XML} data",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "1--40",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papadias:2005:PSC,
  author =       "Dimitris Papadias and Yufei Tao and Greg Fu and
                 Bernhard Seeger",
  title =        "Progressive skyline computation in database systems",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "41--82",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Witkowski:2005:ASM,
  author =       "Andrew Witkowski and Srikanth Bellamkonda and Tolga
                 Bozkaya and Nathan Folkert and Abhinav Gupta and John
                 Haydu and Lei Sheng and Sankar Subramanian",
  title =        "Advanced {SQL} modeling in {RDBMS}",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "83--121",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Madden:2005:TAQ,
  author =       "Samuel R. Madden and Michael J. Franklin and Joseph M.
                 Hellerstein and Wei Hong",
  title =        "{TinyDB}: an acquisitional query processing system for
                 sensor networks",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "122--173",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fagin:2005:DEG,
  author =       "Ronald Fagin and Phokion G. Kolaitis and Lucian Popa",
  title =        "Data exchange: getting to the core",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "174--210",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pu:2005:CDS,
  author =       "Ken Q. Pu and Alberto O. Mendelzon",
  title =        "Concise descriptions of subsets of structured sets",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "211--248",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cormode:2005:WHW,
  author =       "Graham Cormode and S. Muthukrishnan",
  title =        "What's hot and what's not: tracking most frequent
                 items dynamically",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "249--278",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Garofalakis:2005:XSP,
  author =       "Minos Garofalakis and Amit Kumar",
  title =        "{XML} stream processing using tree-edit distance
                 embeddings",
  journal =      j-TODS,
  volume =       "30",
  number =       "1",
  pages =        "279--332",
  month =        mar,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat May 7 08:01:30 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Geerts:2005:TUB,
  author =       "Floris Geerts and Bart Goethals and Jan {Van Den
                 Bussche}",
  title =        "Tight upper bounds on the number of candidate
                 patterns",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "333--363",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jagadish:2005:IAB,
  author =       "H. V. Jagadish and Beng Chin Ooi and Kian-Lee Tan and
                 Cui Yu and Rui Zhang",
  title =        "{iDistance}: {An} adaptive {B$^+$}-tree based indexing
                 method for nearest neighbor search",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "364--397",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Braga:2005:XXU,
  author =       "Daniele Braga and Alessandro Campi and Stefano Ceri",
  title =        "{{\em XQBE} ({\em XQ\/}uery {\em B\/}y {\em
                 E\/}xample)}: a visual interface to the standard {XML}
                 query language",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "398--443",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gottlob:2005:EAP,
  author =       "Georg Gottlob and Christoph Koch and Reinhard
                 Pichler",
  title =        "Efficient algorithms for processing {XPath} queries",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "444--491",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fekete:2005:MSI,
  author =       "Alan Fekete and Dimitrios Liarokapis and Elizabeth
                 O'Neil and Patrick O'Neil and Dennis Shasha",
  title =        "Making snapshot isolation serializable",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "492--528",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papadias:2005:ANN,
  author =       "Dimitris Papadias and Yufei Tao and Kyriakos
                 Mouratidis and Chun Kit Hui",
  title =        "Aggregate nearest neighbor queries in spatial
                 databases",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "529--576",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Peng:2005:XSX,
  author =       "Feng Peng and Sudarshan S. Chawathe",
  title =        "{XSQ}: a streaming {XPath} engine",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "577--623",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wyss:2005:RLM,
  author =       "Catharine M. Wyss and Edward L. Robertson",
  title =        "Relational languages for metadata integration",
  journal =      j-TODS,
  volume =       "30",
  number =       "2",
  pages =        "624--660",
  month =        jun,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 7 14:14:12 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhang:2005:GMD,
  author =       "Rui Zhang and Panos Kalnis and Beng Chin Ooi and
                 Kian-Lee Tan",
  title =        "Generalized multidimensional data mapping and query
                 processing",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "661--697",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pang:2005:IMS,
  author =       "Chaoyi Pang and Guozhu Dong and Kotagiri
                 Ramamohanarao",
  title =        "Incremental maintenance of shortest distance and
                 transitive closure in first-order logic and {SQL}",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "698--721",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wijsen:2005:DRU,
  author =       "Jef Wijsen",
  title =        "Database repairing using updates",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "722--768",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Litwin:2005:LHA,
  author =       "Witold Litwin and Rim Moussa and Thomas Schwarz",
  title =        "{LH*RS}---a highly-available scalable distributed data
                 structure",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "769--811",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{He:2005:STC,
  author =       "Zhen He and Byung Suk Lee and Robert Snapp",
  title =        "Self-tuning cost modeling of user-defined functions in
                 an object-relational {DBMS}",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "812--853",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hurtado:2005:CSI,
  author =       "Carlos A. Hurtado and Claudio Gutierrez and Alberto O.
                 Mendelzon",
  title =        "Capturing summarizability with integrity constraints
                 in {OLAP}",
  journal =      j-TODS,
  volume =       "30",
  number =       "3",
  pages =        "854--886",
  month =        sep,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Oct 12 07:55:28 MDT 2005",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Suciu:2005:F,
  author =       "Dan Suciu and Gerhard Weikum",
  title =        "Foreword",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "887--887",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Garofalakis:2005:WSG,
  author =       "Minos Garofalakis and Amit Kumar",
  title =        "Wavelet synopses for general error metrics",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "888--928",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Marx:2005:CX,
  author =       "Maarten Marx",
  title =        "Conditional {XPath}",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "929--959",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yan:2005:GIB,
  author =       "Xifeng Yan and Philip S. Yu and Jiawei Han",
  title =        "Graph indexing based on discriminative frequent
                 structure analysis",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "960--993",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fagin:2005:CSM,
  author =       "Ronald Fagin and Phokion G. Kolaitis and Lucian Popa
                 and Wang-Chiew Tan",
  title =        "Composing schema mappings: {Second-order} dependencies
                 to the rescue",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "994--1055",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bowman:2005:OQS,
  author =       "Ivan T. Bowman and Kenneth Salem",
  title =        "Optimization of query streams using semantic
                 prefetching",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "1056--1101",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kaushik:2005:SQO,
  author =       "Raghav Kaushik and Jeffrey F. Naughton and Raghu
                 Ramakrishnan and Venkatesan T. Chakravarthy",
  title =        "Synopses for query optimization: a space-complexity
                 perspective",
  journal =      j-TODS,
  volume =       "30",
  number =       "4",
  pages =        "1102--1127",
  month =        dec,
  year =         "2005",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Feb 16 11:31:47 MST 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wu:2006:OBI,
  author =       "Kesheng Wu and Ekow J. Otoo and Arie Shoshani",
  title =        "Optimizing bitmap indices with efficient compression",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "1--38",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schneider:2006:TRB,
  author =       "Markus Schneider and Thomas Behr",
  title =        "Topological relationships between complex spatial
                 objects",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "39--81",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jaluta:2006:BTC,
  author =       "Ibrahim Jaluta and Seppo Sippu and Eljas
                 Soisalon-Soininen",
  title =        "{B}-tree concurrency control and recovery in
                 page-server database systems",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "82--132",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gray:2006:CTC,
  author =       "Jim Gray and Leslie Lamport",
  title =        "Consensus on transaction commit",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "133--160",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Guha:2006:IXD,
  author =       "Sudipto Guha and H. V. Jagadish and Nick Koudas and
                 Divesh Srivastava and Ting Yu",
  title =        "Integrating {XML} data sources using approximate
                 joins",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "161--207",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Abiteboul:2006:RQX,
  author =       "Serge Abiteboul and Luc Segoufin and Victor Vianu",
  title =        "Representing and querying {XML} with incomplete
                 information",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "208--254",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pelanis:2006:IPP,
  author =       "Mindaugas Pelanis and Simonas {\v{S}}altenis and
                 Christian S. Jensen",
  title =        "Indexing the past, present, and anticipated future
                 positions of moving objects",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "255--298",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Rao:2006:SXD,
  author =       "Praveen Rao and Bongki Moon",
  title =        "Sequencing {XML} data and query twigs for fast pattern
                 matching",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "299--345",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{He:2006:ACS,
  author =       "Bin He and Kevin Chen-Chuan Chang",
  title =        "Automatic complex schema matching across {Web} query
                 interfaces: a correlation mining approach",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "346--395",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Guha:2006:ASA,
  author =       "Sudipto Guha and Nick Koudas and Kyuseok Shim",
  title =        "Approximation and streaming algorithms for histogram
                 construction problems",
  journal =      j-TODS,
  volume =       "31",
  number =       "1",
  pages =        "396--438",
  month =        mar,
  year =         "2006",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri May 26 08:20:49 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Qian:2006:DIM,
  author =       "Gang Qian and Qiang Zhu and Qiang Xue and Sakti
                 Pramanik",
  title =        "Dynamic indexing for multidimensional non-ordered
                 discrete data spaces using a data-partitioning
                 approach",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "439--484",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138395",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Iwerks:2006:MNS,
  author =       "Glenn S. Iwerks and Hanan Samet and Kenneth P. Smith",
  title =        "Maintenance of {$K$}-nn and spatial join queries on
                 continuously moving points",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "485--536",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138396",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pentaris:2006:QOD,
  author =       "Fragkiskos Pentaris and Yannis Ioannidis",
  title =        "Query optimization in distributed networks of
                 autonomous database systems",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "537--583",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138397",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lin:2006:SLT,
  author =       "Xuemin Lin and Qing Liu and Yidong Yuan and Xiaofang
                 Zhou and Hongjun Lu",
  title =        "Summarizing level-two topological relations in large
                 spatial datasets",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "584--630",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138398",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bright:2006:APB,
  author =       "Laura Bright and Avigdor Gal and Louiqa Raschid",
  title =        "Adaptive pull-based policies for wide area data
                 delivery",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "631--671",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138399",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cohen:2006:RQA,
  author =       "Sara Cohen and Werner Nutt and Yehoshua Sagiv",
  title =        "Rewriting queries with arbitrary aggregation functions
                 using views",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "672--715",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138400",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kalashnikov:2006:DID,
  author =       "Dmitri V. Kalashnikov and Sharad Mehrotra",
  title =        "Domain-independent data cleaning via analysis of
                 entity-relationship graph",
  journal =      j-TODS,
  volume =       "31",
  number =       "2",
  pages =        "716--767",
  month =        jun,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1138394.1138401",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 14 10:11:33 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Eiter:2006:ISI,
  author =       "Thomas Eiter and Leonid Libkin",
  title =        "Introduction to special {ICDT} section",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "769--769",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166075",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Martens:2006:ECX,
  author =       "Wim Martens and Frank Neven and Thomas Schwentick and
                 Geert Jan Bex",
  title =        "Expressiveness and complexity of {XML Schema}",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "770--813",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166076",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Shaft:2006:TNN,
  author =       "Uri Shaft and Raghu Ramakrishnan",
  title =        "Theory of nearest neighbors indexability",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "814--838",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166077",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Braganholo:2006:PFA,
  author =       "Vanessa P. Braganholo and Susan B. Davidson and Carlos
                 A. Heuser",
  title =        "{PATAX{\'O}}: a framework to allow updates through
                 {XML} views",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "839--886",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166078",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Harizopoulos:2006:IIC,
  author =       "Stavros Harizopoulos and Anastassia Ailamaki",
  title =        "Improving instruction cache performance in {OLTP}",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "887--920",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166079",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Shao:2006:TNV,
  author =       "Feng Shao and Antal Novak and Jayavel
                 Shanmugasundaram",
  title =        "Triggers over nested views of relational data",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "921--967",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166080",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{May:2006:SQU,
  author =       "Norman May and Sven Helmer and Guido Moerkotte",
  title =        "Strategies for query unnesting in {XML} databases",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "968--1013",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166081",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Polyzotis:2006:XSX,
  author =       "Neoklis Polyzotis and Minos Garofalakis",
  title =        "{XSKETCH} synopses for {XML} data graphs",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "1014--1063",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166082",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sugumaran:2006:RDO,
  author =       "Vijayan Sugumaran and Veda C. Storey",
  title =        "The role of domain ontologies in database design: {An}
                 ontology management and conceptual modeling
                 environment",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "1064--1094",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166083",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Metwally:2006:IES,
  author =       "Ahmed Metwally and Divyakant Agrawal and Amr {El
                 Abbadi}",
  title =        "An integrated efficient solution for computing
                 frequent and top-$k$ elements in data streams",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "1095--1133",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166084",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See comments in \cite{Liu:2010:CIE}.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chaudhuri:2006:PIR,
  author =       "Surajit Chaudhuri and Gautam Das and Vagelis Hristidis
                 and Gerhard Weikum",
  title =        "Probabilistic information retrieval approach for
                 ranking of database query results",
  journal =      j-TODS,
  volume =       "31",
  number =       "3",
  pages =        "1134--1168",
  month =        sep,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1166074.1166085",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 17 05:41:01 MDT 2006",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Calders:2006:EPA,
  author =       "Toon Calders and Laks V. S. Lakshmanan and Raymond T.
                 Ng and Jan Paredaens",
  title =        "Expressive power of an algebra for data mining",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1169--1214",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189770",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The relational data model has simple and clear
                 foundations on which significant theoretical and
                 systems research has flourished. By contrast, most
                 research on data mining has focused on algorithmic
                 issues. A major open question is: what's an appropriate
                 foundation for data mining, which can accommodate
                 disparate mining tasks? We address this problem by
                 presenting a database model and an algebra for data
                 mining. The database model is based on the 3W-model
                 introduced by Johnson et al. [2000]. This model relied
                 on black box mining operators. A main contribution of
                 this article is to open up these black boxes, by using
                 generic operators in a data mining algebra. Two key
                 operators in this algebra are regionize, which creates
                 regions (or models) from data tuples, and a restricted
                 form of looping called mining loop. Then the resulting
                 data mining algebra MA is studied and properties
                 concerning expressive power and complexity are
                 established. We present results in three directions:
                 (1) expressiveness of the mining algebra; (2) relations
                 with alternative frameworks, and (3) interactions
                 between regionize and mining loop.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Algebra; data mining; expressive power",
}

@Article{Koch:2006:CNX,
  author =       "Christoph Koch",
  title =        "On the complexity of nonrecursive {XQuery} and
                 functional query languages on complex values",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1215--1256",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189771",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article studies the complexity of evaluating
                 functional query languages for complex values such as
                 monad algebra and the recursion-free fragment of
                 XQuery. We show that monad algebra, with equality
                 restricted to atomic values, is complete for the class
                 TA[2 O ( n ), O ( n )] of problems solvable in linear
                 exponential time with a linear number of alternations
                 if the query is assumed to be part of the input. The
                 monotone fragment of monad algebra with atomic value
                 equality but without negation is NEXPTIME-complete. For
                 monad algebra with deep value equality, that is,
                 equality of complex values, we establish TA[2 O ( n ),
                 O ( n )] lower and exponential-space upper bounds. We
                 also study a fragment of XQuery, Core XQuery, that
                 seems to incorporate all the features of a query
                 language on complex values that are traditionally
                 deemed essential. A close connection between monad
                 algebra on lists and Core XQuery (with ``child'' as the
                 only axis) is exhibited. The two languages are shown
                 expressively equivalent up to representation issues. We
                 show that Core XQuery is just as hard as monad algebra
                 with respect to query and combined complexity. As Core
                 XQuery is NEXPTIME-hard, the best-known techniques for
                 processing such problems require exponential amounts of
                 working memory and doubly exponential time in the worst
                 case. We present a property of queries---the lack of a
                 certain form of composition---that virtually all
                 real-world XQueries have and that allows for query
                 evaluation in PSPACE and thus singly exponential time.
                 Still, we are able to show for an important special
                 case---Core XQuery with equality testing restricted to
                 atomic values---that the composition-free language is
                 just as expressive as the language with composition.
                 Thus, under widely-held complexity-theoretic
                 assumptions, the language with composition is an
                 exponentially more succinct version of the
                 composition-free language.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Complex values; complexity; conservativity;
                 expressiveness; monad algebra; nested-relational
                 algebra; XML; XQuery",
}

@Article{Ilyas:2006:ARA,
  author =       "Ihab F. Ilyas and Walid G. Aref and Ahmed K.
                 Elmagarmid and Hicham G. Elmongui and Rahul Shah and
                 Jeffrey Scott Vitter",
  title =        "Adaptive rank-aware query optimization in relational
                 databases",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1257--1304",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189772",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Rank-aware query processing has emerged as a key
                 requirement in modern applications. In these
                 applications, efficient and adaptive evaluation of top-
                 k queries is an integral part of the application
                 semantics. In this article, we introduce a rank-aware
                 query optimization framework that fully integrates
                 rank-join operators into relational query engines. The
                 framework is based on extending the System R dynamic
                 programming algorithm in both enumeration and pruning.
                 We define ranking as an interesting physical property
                 that triggers the generation of rank-aware query plans.
                 Unlike traditional join operators, optimizing for
                 rank-join operators depends on estimating the input
                 cardinality of these operators. We introduce a
                 probabilistic model for estimating the input
                 cardinality, and hence the cost of a rank-join
                 operator. To our knowledge, this is the first effort in
                 estimating the needed input size for optimal rank
                 aggregation algorithms. Costing ranking plans is key to
                 the full integration of rank-join operators in
                 real-world query processing engines. Since optimal
                 execution strategies picked by static query optimizers
                 lose their optimality due to estimation errors and
                 unexpected changes in the computing environment, we
                 introduce several adaptive execution strategies for
                 top- k queries that respond to these unexpected changes
                 and costing errors. Our reactive reoptimization
                 techniques change the execution plan at runtime to
                 significantly enhance the performance of running
                 queries. Since top- k query plans are usually pipelined
                 and maintain a complex ranking state, altering the
                 execution strategy of a running ranking query is an
                 important and challenging task. We conduct an extensive
                 experimental study to evaluate the performance of the
                 proposed framework. The experimental results are
                 twofold: (1) we show the effectiveness of our
                 cost-based approach of integrating ranking plans in
                 dynamic programming cost-based optimizers; and (2) we
                 show a significant speedup (up to 300\%) when using our
                 adaptive execution of ranking plans over the
                 state-of-the-art mid-query reoptimization strategies.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "adaptive processing; Advanced query processing;
                 rank-aware optimization; ranking; top-k",
}

@Article{Jiao:2006:MSS,
  author =       "Yishan Jiao",
  title =        "Maintaining stream statistics over multiscale sliding
                 windows",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1305--1334",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189773",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we propose a new multiscale sliding
                 window model which differentiates data items in
                 different time periods of the data stream, based on a
                 reasonable monotonicity of resolution assumption. Our
                 model, as a well-motivated extension of the sliding
                 window model, stands halfway between the traditional
                 all-history and time-decaying models. We also present
                 algorithms for estimating two significant data stream
                 statistics--- $ F_0 $ and Jacard's similarity
                 coefficient---with reasonable accuracies under the new
                 model. Our algorithms use space logarithmic in the data
                 stream size and linear in the number of windows; they
                 support update time logarithmic in the number of
                 windows and independent of the accuracy required. Our
                 algorithms are easy to implement. Experimental results
                 demonstrate the efficiencies of our algorithms. Our
                 techniques apply to scenarios in which universe
                 sampling is used.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data stream; F 0; Jacard's similarity coefficient;
                 multiscale sliding window model",
}

@Article{Pei:2006:TMS,
  author =       "Jian Pei and Yidong Yuan and Xuemin Lin and Wen Jin
                 and Martin Ester and Qing Liu and Wei Wang and Yufei
                 Tao and Jeffrey Xu Yu and Qing Zhang",
  title =        "Towards multidimensional subspace skyline analysis",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1335--1381",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189774",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The skyline operator is important for multicriteria
                 decision-making applications. Although many recent
                 studies developed efficient methods to compute skyline
                 objects in a given space, none of them considers
                 skylines in multiple subspaces simultaneously. More
                 importantly, the fundamental problem on the semantics
                 of skylines remains open: Why and in which subspaces is
                 (or is not) an object in the skyline? Practically,
                 users may also be interested in the skylines in any
                 subspaces. Then, what is the relationship between the
                 skylines in the subspaces and those in the
                 super-spaces? How can we effectively analyze the
                 subspace skylines? Can we efficiently compute skylines
                 in various subspaces and answer various analytical
                 queries?In this article, we tackle the problem of
                 multidimensional subspace skyline computation and
                 analysis. We explore skylines in subspaces. First, we
                 propose the concept of Skycube, which consists of
                 skylines of all possible nonempty subspaces of a given
                 full space. Once a Skycube is materialized, any
                 subspace skyline queries can be answered online.
                 However, Skycube cannot fully address the semantic
                 concerns and may contain redundant information. To
                 tackle the problem, we introduce a novel notion of
                 skyline group which essentially is a group of objects
                 that coincide in the skylines of some subspaces. We
                 identify the decisive subspaces that qualify skyline
                 groups in the subspace skylines. The new notions
                 concisely capture the semantics and the structures of
                 skylines in various subspaces. Multidimensional roll-up
                 and drill-down analysis is introduced. We also develop
                 efficient algorithms to compute Skycube, skyline groups
                 and their decisive subspaces. A systematic performance
                 study using both real data sets and synthetic data sets
                 is reported to evaluate our approach.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data cubing; multidimensional data analysis; Skyline
                 query",
}

@Article{Jermaine:2006:SMS,
  author =       "Christopher Jermaine and Alin Dobra and Subramanian
                 Arumugam and Shantanu Joshi and Abhijit Pol",
  title =        "The {Sort-Merge-Shrink} join",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1382--1416",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189775",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "One of the most common operations in analytic query
                 processing is the application of an aggregate function
                 to the result of a relational join. We describe an
                 algorithm called the Sort-Merge-Shrink (SMS) Join for
                 computing the answer to such a query over large,
                 disk-based input tables. The key innovation of the SMS
                 join is that if the input data are clustered in a
                 statistically random fashion on disk, then at all
                 times, the join provides an online, statistical
                 estimator for the eventual answer to the query as well
                 as probabilistic confidence bounds. Thus, a user can
                 monitor the progress of the join throughout its
                 execution and stop the join when satisfied with the
                 estimate's accuracy or run the algorithm to completion
                 with a total time requirement that is not much longer
                 than that of other common join algorithms. This
                 contrasts with other online join algorithms, which
                 either do not offer such statistical guarantees or can
                 only offer guarantees so long as the input data can fit
                 into main memory.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "nonparametric statistics; OLAP; Online algorithms",
}

@Article{Afrati:2006:FSS,
  author =       "Foto Afrati and Jennifer Widom",
  title =        "Foreword to special section on {SIGMOD\slash PODS}
                 2005",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1417--1417",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189776",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yan:2006:FBS,
  author =       "Xifeng Yan and Feida Zhu and Philip S. Yu and Jiawei
                 Han",
  title =        "Feature-based similarity search in graph structures",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1418--1453",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189777",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Similarity search of complex structures is an
                 important operation in graph-related applications since
                 exact matching is often too restrictive. In this
                 article, we investigate the issues of substructure
                 similarity search using indexed features in graph
                 databases. By transforming the edge relaxation ratio of
                 a query graph into the maximum allowed feature misses,
                 our structural filtering algorithm can filter graphs
                 without performing pairwise similarity computation. It
                 is further shown that using either too few or too many
                 features can result in poor filtering performance. Thus
                 the challenge is to design an effective feature set
                 selection strategy that could maximize the filtering
                 capability. We prove that the complexity of optimal
                 feature set selection is $ \Omega (2^m) $ in the worst
                 case, where $m$ is the number of features for
                 selection. In practice, we identify several criteria to
                 build effective feature sets for filtering, and
                 demonstrate that combining features with similar size
                 and selectivity can improve the filtering and search
                 performance significantly within a multifilter
                 composition framework. The proposed feature-based
                 filtering concept can be generalized and applied to
                 searching approximate nonconsecutive sequences, trees,
                 and other structured data as well.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "complexity; Graph database; index; similarity search",
}

@Article{Fuxman:2006:PDE,
  author =       "Ariel Fuxman and Phokion G. Kolaitis and Ren{\'e}e J.
                 Miller and Wang-Chiew Tan",
  title =        "Peer data exchange",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1454--1498",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189778",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we introduce and study a framework,
                 called peer data exchange, for sharing and exchanging
                 data between peers. This framework is a special case of
                 a full-fledged peer data management system and a
                 generalization of data exchange between a source schema
                 and a target schema. The motivation behind peer data
                 exchange is to model authority relationships between
                 peers, where a source peer may contribute data to a
                 target peer, specified using source-to-target
                 constraints, and a target peer may use target-to-source
                 constraints to restrict the data it is willing to
                 receive, but cannot modify the data of the source
                 peer.A fundamental algorithmic problem in this
                 framework is that of deciding the existence of a
                 solution: given a source instance and a target instance
                 for a fixed peer data exchange setting, can the target
                 instance be augmented in such a way that the source
                 instance and the augmented target instance satisfy all
                 constraints of the setting? We investigate the
                 computational complexity of the problem for peer data
                 exchange settings in which the constraints are given by
                 tuple generating dependencies. We show that this
                 problem is always in NP, and that it can be NP-complete
                 even for ``acyclic'' peer data exchange settings. We
                 also show that the data complexity of the certain
                 answers of target conjunctive queries is in coNP, and
                 that it can be coNP-complete even for ``acyclic'' peer
                 data exchange settings. After this, we explore the
                 boundary between tractability and intractability for
                 deciding the existence of a solution and for computing
                 the certain answers of target conjunctive queries. To
                 this effect, we identify broad syntactic conditions on
                 the constraints between the peers under which the
                 existence-of-solutions problem is solvable in
                 polynomial time. We also identify syntactic conditions
                 between peer data exchange settings and target
                 conjunctive queries that yield polynomial-time
                 algorithms for computing the certain answers. For both
                 problems, these syntactic conditions turn out to be
                 tight, in the sense that minimal relaxations of them
                 lead to intractability. Finally, we introduce the
                 concept of a universal basis of solutions in peer data
                 exchange and explore its properties.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "certain answers; conjunctive queries; Data exchange;
                 data integration; metadata model management; schema
                 mapping",
}

@Article{Cheng:2006:DMM,
  author =       "David Cheng and Ravi Kannan and Santosh Vempala and
                 Grant Wang",
  title =        "A divide-and-merge methodology for clustering",
  journal =      j-TODS,
  volume =       "31",
  number =       "4",
  pages =        "1499--1525",
  month =        dec,
  year =         "2006",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1189769.1189779",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:31 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We present a divide-and-merge methodology for
                 clustering a set of objects that combines a top-down
                 ``divide'' phase with a bottom-up ``merge'' phase. In
                 contrast, previous algorithms use either top-down or
                 bottom-up methods to construct a hierarchical
                 clustering or produce a flat clustering using local
                 search (e.g., k -means). For the divide phase, which
                 produces a tree whose leaves are the elements of the
                 set, we suggest an efficient spectral algorithm. When
                 the data is in the form of a sparse document-term
                 matrix, we show how to modify the algorithm so that it
                 maintains sparsity and runs in linear space. The merge
                 phase quickly finds the optimal partition that respects
                 the tree for many natural objective functions, for
                 example, k -means, min-diameter, min-sum, correlation
                 clustering, etc. We present a thorough experimental
                 evaluation of the methodology. We describe the
                 implementation of a meta-search engine that uses this
                 methodology to cluster results from web searches. We
                 also give comparative empirical results on several real
                 datasets.",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Clustering; data mining; information retrieval",
}

@Article{Snodgrass:2007:ESV,
  author =       "Richard T. Snodgrass",
  title =        "Editorial: {Single}- versus double-blind reviewing",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206050",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This editorial analyzes from a variety of perspectives
                 the controversial issue of single-blind versus
                 double-blind reviewing. In single-blind reviewing, the
                 reviewer is unknown to the author, but the identity of
                 the author is known to the reviewer. Double-blind
                 reviewing is more symmetric: The identity of the author
                 and the reviewer are not revealed to each other. We
                 first examine the significant scholarly literature
                 regarding blind reviewing. We then list six benefits
                 claimed for double-blind reviewing and 21 possible
                 costs. To compare these benefits and costs, we propose
                 a double-blind policy for TODS that attempts to
                 minimize the costs while retaining the core benefit of
                 fairness that double-blind reviewing provides, and
                 evaluate that policy against each of the listed
                 benefits and costs. Following that is a general
                 discussion considering several questions: What does
                 this have to do with TODS, does bias exist in computer
                 science, and what is the appropriate decision
                 procedure? We explore the ``knobs'' a policy design can
                 manipulate to fine-tune a double-blind review policy.
                 This editorial ends with a specific decision.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Anonymous citation; blinding efficacy; double-blind
                 review; gender bias; single-blind review; status bias",
}

@Article{Pourabbas:2007:EEJ,
  author =       "Elaheh Pourabbas and Arie Shoshani",
  title =        "Efficient estimation of joint queries from multiple
                 {OLAP} databases",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206051",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Given an OLAP query expressed over multiple source
                 OLAP databases, we study the problem of estimating the
                 resulting OLAP target database. The problem arises when
                 it is not possible to derive the result from a single
                 database. The method we use is linear indirect
                 estimation, commonly used for statistical estimation.
                 We examine two obvious computational methods for
                 computing such a target database, called the full
                 cross-product (F) and preaggregation (P) methods. We
                 study the accuracy and computational cost of these
                 methods. While the F method provides a more accurate
                 estimate, it is more expensive computationally than P.
                 Our contribution is in proposing a third, new method,
                 called the partial preaggregation method (PP), which is
                 significantly less expensive than F, but just as
                 accurate. We prove formally that the PP method yields
                 the same results as the F method, and provide
                 analytical and experimental results on the accuracy and
                 computational benefits of the PP method.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "multiple summary databases; OLAP; query estimation",
}

@Article{Olteanu:2007:FNS,
  author =       "Dan Olteanu",
  title =        "Forward node-selecting queries over trees",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206052",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Node-selecting queries over trees lie at the core of
                 several important XML languages for the web, such as
                 the node-selection language XPath, the query language
                 XQuery, and the transformation language XSLT. The main
                 syntactic constructs of such queries are the backward
                 predicates, for example, ancestor and preceding, and
                 the forward predicates, for example, descendant and
                 following. Forward predicates are included in the
                 depth-first, left-to-right preorder relation associated
                 with the input tree, whereas backward predicates are
                 included in the inverse of this preorder
                 relation.\par

                 This work is devoted to an expressiveness study of
                 node-selecting queries with proven theoretical and
                 practical applicability, especially in the field of
                 query evaluation against XML streams. The main question
                 it answers positively is whether, for each input query
                 with forward and backward predicates, there exists an
                 equivalent forward-only output query. This question is
                 then positively answered for input and output queries
                 of varying structural complexity, using LOGLIN and
                 PSPACE reductions.\par

                 Various existing applications based on the results of
                 this work are reported, including query optimization
                 and streamed evaluation.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Expressiveness; rewriting; streams; XML; XPath",
}

@Article{Nash:2007:CMG,
  author =       "Alan Nash and Philip A. Bernstein and Sergey Melnik",
  title =        "Composition of mappings given by embedded
                 dependencies",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "4:1--4:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206053",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Composition of mappings between schemas is essential
                 to support schema evolution, data exchange, data
                 integration, and other data management tasks. In many
                 applications, mappings are given by embedded
                 dependencies. In this article, we study the issues
                 involved in composing such mappings.\par

                 Our algorithms and results extend those of Fagin et al.
                 [2004], who studied the composition of mappings given
                 by several kinds of constraints. In particular, they
                 proved that full source-to-target tuple-generating
                 dependencies (tgds) are closed under composition, but
                 embedded source-to-target tgds are not. They introduced
                 a class of second-order constraints, SO tgds, that is
                 closed under composition and has desirable properties
                 for data exchange.\par

                 We study constraints that need not be source-to-target
                 and we concentrate on obtaining (first-order) embedded
                 dependencies. As part of this study, we also consider
                 full dependencies and second-order constraints that
                 arise from Skolemizing embedded dependencies. For each
                 of the three classes of mappings that we study, we
                 provide: (a) an algorithm that attempts to compute the
                 composition; and (b) sufficient conditions on the input
                 mappings which guarantee that the algorithm will
                 succeed.\par

                 In addition, we give several negative results. In
                 particular, we show that full and second-order
                 dependencies that are not limited to be
                 source-to-target are not closed under composition (for
                 the latter, under the additional restriction that no
                 new function symbols are introduced). Furthermore, we
                 show that determining whether the composition can be
                 given by these kinds of dependencies is undecidable.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "database theory; Metadata management",
}

@Article{Hwang:2007:OTK,
  author =       "Seung-won Hwang and Kevin Chen-chuan Chang",
  title =        "Optimizing top-k queries for middleware access: a
                 unified cost-based approach",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "5:1--5:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206054",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article studies optimizing top- k queries in
                 middlewares. While many assorted algorithms have been
                 proposed, none is generally applicable to a wide range
                 of possible scenarios. Existing algorithms lack both
                 the ``generality'' to support a wide range of access
                 scenarios and the systematic ``adaptivity'' to account
                 for runtime specifics. To fulfill this critical
                 lacking, we aim at taking a cost-based optimization
                 approach: By runtime search over a space of algorithms,
                 cost-based optimization is general across a wide range
                 of access scenarios, yet adaptive to the specific
                 access costs at runtime. While such optimization has
                 been taken for granted for relational queries from
                 early on, it has been clearly lacking for ranked
                 queries. In this article, we thus identify and address
                 the barriers of realizing such a unified framework. As
                 the first barrier, we need to define a
                 ``comprehensive'' space encompassing all possibly
                 optimal algorithms to search over. As the second
                 barrier and a conflicting goal, such a space should
                 also be ``focused'' enough to enable efficient search.
                 For SQL queries that are explicitly composed of
                 relational operators, such a space, by definition,
                 consists of schedules of relational operators (or
                 ``query plans''). In contrast, top- k queries do not
                 have logical tasks, such as relational operators. We
                 thus define the logical tasks of top- k queries as
                 building blocks to identify a comprehensive and focused
                 space for top- k queries. We then develop efficient
                 search schemes over such space for identifying the
                 optimal algorithm. Our study indicates that our
                 framework not only unifies, but also outperforms
                 existing algorithms specifically designed for their
                 scenarios.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "middlewares; Top-k query processing",
}

@Article{Ceri:2007:MCV,
  author =       "Stefano Ceri and Francesco {Di Giunta} and Pier Luca
                 Lanzi",
  title =        "Mining constraint violations",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "6:1--6:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206055",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we introduce pesudoconstraints, a
                 novel data mining pattern aimed at identifying rare
                 events in databases. At first, we formally define
                 pesudoconstraints using a probabilistic model and
                 provide a statistical test to identify
                 pesudoconstraints in a database. Then, we focus on a
                 specific class of pesudoconstraints, named cycle
                 pesudoconstraints, which often occur in databases. We
                 define cycle pesudoconstraints in the context of the ER
                 model and present an automatic method for detecting
                 cycle pesudoconstraints from a relational database.
                 Finally, we present an experiment to show cycle
                 pesudoconstraints ``at work'' on real data.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Deviation detection; probabilistic models; relational
                 data mining",
}

@Article{Jacox:2007:SJT,
  author =       "Edwin H. Jacox and Hanan Samet",
  title =        "Spatial join techniques",
  journal =      j-TODS,
  volume =       "32",
  number =       "1",
  pages =        "7:1--7:??",
  month =        mar,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1206049.1206056",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:36:55 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A variety of techniques for performing a spatial join
                 are reviewed. Instead of just summarizing the
                 literature and presenting each technique in its
                 entirety, distinct components of the different
                 techniques are described and each is decomposed into an
                 overall framework for performing a spatial join. A
                 typical spatial join technique consists of the
                 following components: partitioning the data, performing
                 internal-memory spatial joins on subsets of the data,
                 and checking if the full polygons intersect. Each
                 technique is decomposed into these components and each
                 component addressed in a separate section so as to
                 compare and contrast similar aspects of each technique.
                 The goal of this survey is to describe the algorithms
                 within each component in detail, comparing and
                 contrasting competing methods, thereby enabling further
                 analysis and experimentation with each component and
                 allowing the best algorithms for a particular situation
                 to be built piecemeal, or, even better, enabling an
                 optimizer to choose which algorithms to use.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "External memory algorithms; plane-sweep; spatial
                 join",
}

@Article{Athitsos:2007:QSE,
  author =       "Vassilis Athitsos and Marios Hadjieleftheriou and
                 George Kollios and Stan Sclaroff",
  title =        "Query-sensitive embeddings",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242525",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A common problem in many types of databases is
                 retrieving the most similar matches to a query object.
                 Finding these matches in a large database can be too
                 slow to be practical, especially in domains where
                 objects are compared using computationally expensive
                 similarity (or distance) measures. Embedding methods
                 can significantly speed-up retrieval by mapping objects
                 into a vector space, where distances can be measured
                 rapidly using a Minkowski metric. In this article we
                 present a novel way to improve embedding quality. In
                 particular, we propose to construct embeddings that use
                 a query-sensitive distance measure for the target space
                 of the embedding. This distance measure is used to
                 compare those vectors that the query and database
                 objects are mapped to. The term ``query-sensitive''
                 means that the distance measure changes, depending on
                 the current query object. We demonstrate theoretically
                 that using a query-sensitive distance measure increases
                 the modeling power of embeddings and allows them to
                 capture more of the structure of the original space. We
                 also demonstrate experimentally that query-sensitive
                 embeddings can significantly improve retrieval
                 performance. In experiments with an image database of
                 handwritten digits and a time-series database, the
                 proposed method outperforms existing state-of-the-art
                 non-Euclidean indexing methods, meaning that it
                 provides significantly better tradeoffs between
                 efficiency and retrieval accuracy.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Embedding methods; nearest-neighbor retrieval;
                 non-Euclidean spaces; nonmetric spaces; similarity
                 matching",
}

@Article{Chaudhuri:2007:OSS,
  author =       "Surajit Chaudhuri and Gautam Das and Vivek Narasayya",
  title =        "Optimized stratified sampling for approximate query
                 processing",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242526",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The ability to approximately answer aggregation
                 queries accurately and efficiently is of great benefit
                 for decision support and data mining tools. In contrast
                 to previous sampling-based studies, we treat the
                 problem as an optimization problem where, given a
                 workload of queries, we select a stratified random
                 sample of the original data such that the error in
                 answering the workload queries using the sample is
                 minimized. A key novelty of our approach is that we can
                 tailor the choice of samples to be robust, even for
                 workloads that are ``similar'' but not necessarily
                 identical to the given workload. Finally, our
                 techniques recognize the importance of taking into
                 account the variance in the data distribution in a
                 principled manner. We show how our solution can be
                 implemented on a database system, and present results
                 of extensive experiments on Microsoft SQL Server that
                 demonstrate the superior quality of our method compared
                 to previous work.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximation; query processing; Random sampling",
}

@Article{Deligiannakis:2007:EWM,
  author =       "Antonios Deligiannakis and Minos Garofalakis and Nick
                 Roussopoulos",
  title =        "Extended wavelets for multiple measures",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242527",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Several studies have demonstrated the effectiveness of
                 the Haar wavelet decomposition as a tool for reducing
                 large amounts of data down to compact wavelet synopses
                 that can be used to obtain fast, accurate approximate
                 answers to user queries. Although originally designed
                 for minimizing the overall mean-squared (i.e., $ L^2
                 $-norm) error in the data approximation, recently
                 proposed methods also enable the use of Haar wavelets
                 in minimizing other error metrics, such as the relative
                 error in data value reconstruction, which is arguably
                 the most important for approximate query answers.
                 Relatively little attention, however, has been paid to
                 the problem of using wavelet synopses as an approximate
                 query answering tool over complex tabular datasets
                 containing multiple measures, such as those typically
                 found in real-life OLAP applications. Existing
                 decomposition approaches will either operate on each
                 measure individually, or treat all measures as a vector
                 of values and process them simultaneously. As we
                 demonstrate in this article, these existing individual
                 or combined storage approaches for the wavelet
                 coefficients of different measures can easily lead to
                 suboptimal storage utilization, resulting in
                 drastically reduced accuracy for approximate query
                 answers. To address this problem, in this work, we
                 introduce the notion of an extended wavelet coefficient
                 as a flexible, efficient storage method for wavelet
                 coefficients over multimeasure data. We also propose
                 novel algorithms for constructing effective (optimal or
                 near-optimal) extended wavelet-coefficient synopses
                 under a given storage constraint, for both sum-squared
                 error and relative-error norms. Experimental results
                 with both real-life and synthetic datasets validate our
                 approach, demonstrating that our techniques
                 consistently obtain significant gains in approximation
                 accuracy compared to existing solutions.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate query processing; data synopses;
                 Wavelets",
}

@Article{Rusu:2007:PRN,
  author =       "Florin Rusu and Alin Dobra",
  title =        "Pseudo-random number generation for sketch-based
                 estimations",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "11:1--11:48",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242528",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The exact computation of aggregate queries, like the
                 size of join of two relations, usually requires large
                 amounts of memory (constrained in data-streaming) or
                 communication (constrained in distributed computation)
                 and large processing times. In this situation,
                 approximation techniques with provable guarantees, like
                 sketches, are one possible solution. The performance of
                 sketches depends crucially on the ability to generate
                 particular pseudo-random numbers. In this article we
                 investigate both theoretically and empirically the
                 problem of generating k -wise independent pseudo-random
                 numbers and, in particular, that of generating 3- and
                 4-wise independent pseudo-random numbers that are fast
                 range-summable (i.e., they can be summed in sublinear
                 time). Our specific contributions are: (a) we provide a
                 thorough comparison of the various pseudo-random number
                 generating schemes; (b) we study both theoretically and
                 empirically the fast range-summation property of 3- and
                 4-wise independent generating schemes; (c) we provide
                 algorithms for the fast range-summation of two 3-wise
                 independent schemes, BCH and extended Hamming; and (d)
                 we show convincing theoretical and empirical evidence
                 that the extended Hamming scheme performs as well as
                 any 4-wise independent scheme for estimating the size
                 of join of two relations using AMS sketches, even
                 though it is only 3-wise independent. We use this
                 scheme to generate estimators that significantly
                 outperform state-of-the-art solutions for two problems,
                 namely, size of spatial joins and selectivity
                 estimation.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate query processing; data synopses; fast
                 range-summation; Sketches",
}

@Article{Mazeika:2007:ESA,
  author =       "Arturas Mazeika and Michael H. B{\"o}hlen and Nick
                 Koudas and Divesh Srivastava",
  title =        "Estimating the selectivity of approximate string
                 queries",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242529",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Approximate queries on string data are important due
                 to the prevalence of such data in databases and various
                 conventions and errors in string data. We present the
                 VSol estimator, a novel technique for estimating the
                 selectivity of approximate string queries. The VSol
                 estimator is based on inverse strings and makes the
                 performance of the selectivity estimator independent of
                 the number of strings. To get inverse strings we
                 decompose all database strings into overlapping
                 substrings of length q (q-grams) and then associate
                 each q-gram with its inverse string: the IDs of all
                 strings that contain the q-gram. We use signatures to
                 compress inverse strings, and clustering to group
                 similar signatures.\par

                 We study our technique analytically and experimentally.
                 The space complexity of our estimator only depends on
                 the number of neighborhoods in the database and the
                 desired estimation error. The time to estimate the
                 selectivity is independent of the number of database
                 strings and linear with respect to the length of query
                 string. We give a detailed empirical performance
                 evaluation of our solution for synthetic and real-world
                 datasets. We show that VSol is effective for large
                 skewed databases of short strings.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Inverse strings; min-wise hash signatures; q-grams",
}

@Article{Zeng:2007:CCC,
  author =       "Zhiping Zeng and Jianyong Wang and Lizhu Zhou and
                 George Karypis",
  title =        "Out-of-core coherent closed quasi-clique mining from
                 large dense graph databases",
  journal =      j-TODS,
  volume =       "32",
  number =       "2",
  pages =        "13:1--13:??",
  month =        jun,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1242524.1242530",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:09 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Due to the ability of graphs to represent more generic
                 and more complicated relationships among different
                 objects, graph mining has played a significant role in
                 data mining, attracting increasing attention in the
                 data mining community. In addition, frequent coherent
                 subgraphs can provide valuable knowledge about the
                 underlying internal structure of a graph database, and
                 mining frequently occurring coherent subgraphs from
                 large dense graph databases has witnessed several
                 applications and received considerable attention in the
                 graph mining community recently. In this article, we
                 study how to efficiently mine the complete set of
                 coherent closed quasi-cliques from large dense graph
                 databases, which is an especially challenging task due
                 to the fact that the downward-closure property no
                 longer holds. By fully exploring some properties of
                 quasi-cliques, we propose several novel optimization
                 techniques which can prune the unpromising and
                 redundant subsearch spaces effectively. Meanwhile, we
                 devise an efficient closure checking scheme to
                 facilitate the discovery of closed quasi-cliques only.
                 Since large databases cannot be held in main memory, we
                 also design an out-of-core solution with efficient
                 index structures for mining coherent closed
                 quasi-cliques from large dense graph databases. We call
                 this Cocain*. Thorough performance study shows that
                 Cocain* is very efficient and scalable for large dense
                 graph databases.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "coherent subgraph; frequent closed subgraph; Graph
                 mining; out-of-core algorithm; quasi-clique",
}

@Article{Ipeirotis:2007:MMC,
  author =       "Panagiotis G. Ipeirotis and Alexandros Ntoulas and
                 Junghoo Cho and Luis Gravano",
  title =        "Modeling and managing changes in text databases",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "14:1--14:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272744",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:57 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Large amounts of (often valuable) information are
                 stored in web-accessible text databases.
                 ``Metasearchers'' provide unified interfaces to query
                 multiple such databases at once. For efficiency,
                 metasearchers rely on succinct statistical summaries of
                 the database contents to select the best databases for
                 each query. So far, database selection research has
                 largely assumed that databases are static, so the
                 associated statistical summaries do not evolve over
                 time. However, databases are rarely static and the
                 statistical summaries that describe their contents need
                 to be updated periodically to reflect content changes.
                 In this article, we first report the results of a study
                 showing how the content summaries of 152 real web
                 databases evolved over a period of 52 weeks. Then, we
                 show how to use ``survival analysis'' techniques in
                 general, and Cox's proportional hazards regression in
                 particular, to model database changes over time and
                 predict when we should update each content summary.
                 Finally, we exploit our change model to devise update
                 schedules that keep the summaries up to date by
                 contacting databases only when needed, and then we
                 evaluate the quality of our schedules experimentally
                 over real web databases.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "distributed information retrieval; Metasearching; text
                 database selection",
}

@Article{Tao:2007:RSM,
  author =       "Yufei Tao and Xiaokui Xiao and Reynold Cheng",
  title =        "Range search on multidimensional uncertain data",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272745",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:22 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In an uncertain database, every object o is associated
                 with a probability density function, which describes
                 the likelihood that o appears at each position in a
                 multidimensional workspace. This article studies two
                 types of range retrieval fundamental to many analytical
                 tasks. Specifically, a nonfuzzy query returns all the
                 objects that appear in a search region $ r_q $ with at
                 least a certain probability $ t_q $. On the other hand,
                 given an uncertain object $q$, fuzzy search retrieves
                 the set of objects that are within distance $
                 \varepsilon_q $ from $q$ with no less than probability
                 $ t_q $. The core of our methodology is a novel concept
                 of ``probabilistically constrained rectangle'', which
                 permits effective pruning\slash validation of
                 nonqualifying\slash qualifying data. We develop a new
                 index structure called the U-tree for minimizing the
                 query overhead. Our algorithmic findings are
                 accompanied with a thorough theoretical analysis, which
                 reveals valuable insight into the problem
                 characteristics, and mathematically confirms the
                 efficiency of our solutions. We verify the
                 effectiveness of the proposed techniques with extensive
                 experiments.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "range search; Uncertain databases",
}

@Article{Sinha:2007:MRB,
  author =       "Rishi Rakesh Sinha and Marianne Winslett",
  title =        "Multi-resolution bitmap indexes for scientific data",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272746",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:57 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The unique characteristics of scientific data and
                 queries cause traditional indexing techniques to
                 perform poorly on scientific workloads, occupy
                 excessive space, or both. Refinements of bitmap indexes
                 have been proposed previously as a solution to this
                 problem. In this article, we describe the difficulties
                 we encountered in deploying bitmap indexes with
                 scientific data and queries from two real-world
                 domains. In particular, previously proposed methods of
                 binning, encoding, and compressing bitmap vectors
                 either were quite slow for processing the large-range
                 query conditions our scientists used, or required
                 excessive storage space. Nor could the indexes easily
                 be built or used on parallel platforms. In this
                 article, we show how to solve these problems through
                 the use of multi-resolution, parallelizable bitmap
                 indexes, which support a fine-grained trade-off between
                 storage requirements and query performance. Our
                 experiments with large data sets from two scientific
                 domains show that multi-resolution, parallelizable
                 bitmap indexes occupy an acceptable amount of storage
                 while improving range query performance by roughly a
                 factor of 10, compared to a single-resolution bitmap
                 index of reasonable size.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "bitmap index; parallel index; Query processing;
                 scientific data management",
}

@Article{Chen:2007:IHJ,
  author =       "Shimin Chen and Anastassia Ailamaki and Phillip B.
                 Gibbons and Todd C. Mowry",
  title =        "Improving hash join performance through prefetching",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272747",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:57 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Hash join algorithms suffer from extensive CPU cache
                 stalls. This article shows that the standard hash join
                 algorithm for disk-oriented databases (i.e. GRACE)
                 spends over 80\% of its user time stalled on CPU cache
                 misses, and explores the use of CPU cache prefetching
                 to improve its cache performance. Applying prefetching
                 to hash joins is complicated by the data dependencies,
                 multiple code paths, and inherent randomness of
                 hashing. We present two techniques, group prefetching
                 and software-pipelined prefetching, that overcome these
                 complications. These schemes achieve 1.29--4.04X
                 speedups for the join phase and 1.37--3.49X speedups
                 for the partition phase over GRACE and simple
                 prefetching approaches. Moreover, compared with
                 previous cache-aware approaches (i.e. cache
                 partitioning), the schemes are at least 36\% faster on
                 large relations and do not require exclusive use of the
                 CPU cache to be effective. Finally, comparing the
                 elapsed real times when disk I/Os are in the picture,
                 our cache prefetching schemes achieve 1.12--1.84X
                 speedups for the join phase and 1.06--1.60X speedups
                 for the partition phase over the GRACE hash join
                 algorithm.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "CPU cache performance; CPU cache prefetching; group
                 prefetching; Hash join; software-pipelined
                 prefetching",
}

@Article{Cao:2007:SQO,
  author =       "Bin Cao and Antonio Badia",
  title =        "{SQL} query optimization through nested relational
                 algebra",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272748",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:57 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Most research work on optimization of nested queries
                 focuses on aggregate subqueries. In this article, we
                 show that existing approaches are not adequate for
                 nonaggregate subqueries, especially for those having
                 multiple subqueries and certain comparison operators.
                 We then propose a new efficient approach, the nested
                 relational approach, based on the nested relational
                 algebra. The nested relational approach treats all
                 subqueries in a uniform manner, being able to deal with
                 nested queries of any type and any level. We report on
                 experimental work that confirms that existing
                 approaches have difficulties dealing with nonaggregate
                 subqueries, and that the nested relational approach
                 offers better performance. We also discuss algebraic
                 optimization rules for further optimizing the nested
                 relational approach and the issue of integrating it
                 into relational database systems.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Nested queries; nested relational algebra;
                 nonrelational query processing",
}

@Article{Mamoulis:2007:ETA,
  author =       "Nikos Mamoulis and Man Lung Yiu and Kit Hung Cheng and
                 David W. Cheung",
  title =        "Efficient top-$k$ aggregation of ranked inputs",
  journal =      j-TODS,
  volume =       "32",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1272743.1272749",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:22 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A top- k query combines different rankings of the same
                 set of objects and returns the k objects with the
                 highest combined score according to an aggregate
                 function. We bring to light some key observations,
                 which impose two phases that any top- k algorithm,
                 based on sorted accesses, should go through. Based on
                 them, we propose a new algorithm, which is designed to
                 minimize the number of object accesses, the
                 computational cost, and the memory requirements of top-
                 k search with monotone aggregate functions. We provide
                 an analysis for its cost and show that it is always no
                 worse than the baseline ``no random accesses''
                 algorithm in terms of computations, accesses, and
                 memory required. As a side contribution, we perform a
                 space analysis, which indicates the memory requirements
                 of top- k algorithms that only perform sorted accesses.
                 For the case, where the required space exceeds the
                 available memory, we propose disk-based variants of our
                 algorithm. We propose and optimize a multiway top- k
                 join operator, with certain advantages over evaluation
                 trees of binary top- k join operators. Finally, we
                 define and study the computation of top- k cubes and
                 the implementation of roll-up and drill-down operations
                 in such cubes. Extensive experiments with synthetic and
                 real data show that, compared to previous techniques,
                 our method accesses fewer objects, while being orders
                 of magnitude faster.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "rank aggregation; Top- k queries",
}

@Article{Chaudhuri:2007:IAS,
  author =       "Surajit Chaudhuri",
  title =        "Introduction to {ACM SIGMOD} 2006 conference papers",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "20:1--20:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292610",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ipeirotis:2007:TQO,
  author =       "Panagiotis G. Ipeirotis and Eugene Agichtein and
                 Pranay Jain and Luis Gravano",
  title =        "Towards a query optimizer for text-centric tasks",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "21:1--21:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292611",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Text is ubiquitous and, not surprisingly, many
                 important applications rely on textual data for a
                 variety of tasks. As a notable example, information
                 extraction applications derive structured relations
                 from unstructured text; as another example, focused
                 crawlers explore the Web to locate pages about specific
                 topics. Execution plans for text-centric tasks follow
                 two general paradigms for processing a text database:
                 either we can scan, or ``crawl,'' the text database or,
                 alternatively, we can exploit search engine indexes and
                 retrieve the documents of interest via carefully
                 crafted queries constructed in task-specific ways. The
                 choice between crawl- and query-based execution plans
                 can have a substantial impact on both execution time
                 and output ``completeness'' (e.g., in terms of recall).
                 Nevertheless, this choice is typically ad hoc and based
                 on heuristics or plain intuition. In this article, we
                 present fundamental building blocks to make the choice
                 of execution plans for text-centric tasks in an
                 informed, cost-based way. Towards this goal, we show
                 how to analyze query- and crawl-based plans in terms of
                 both execution time and output completeness. We adapt
                 results from random-graph theory and statistics to
                 develop a rigorous cost model for the execution plans.
                 Our cost model reflects the fact that the performance
                 of the plans depends on fundamental task-specific
                 properties of the underlying text databases. We
                 identify these properties and present efficient
                 techniques for estimating the associated parameters of
                 the cost model. We also present two optimization
                 approaches for text-centric tasks that rely on the
                 cost-model parameters and select efficient execution
                 plans. Overall, our optimization approaches help build
                 efficient execution plans for a task, resulting in
                 significant efficiency and output completeness
                 benefits. We complement our results with a large-scale
                 experimental evaluation for three important
                 text-centric tasks and over multiple real-life data
                 sets.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "distributed information retrieval; focused crawling;
                 information extraction; Metasearching; text database
                 selection",
}

@Article{Petropoulos:2007:EIQ,
  author =       "Michalis Petropoulos and Alin Deutsch and Yannis
                 Papakonstantinou and Yannis Katsis",
  title =        "Exporting and interactively querying {Web}
                 service-accessed sources: {The CLIDE System}",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "22:1--22:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292612",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The CLIDE System assists the owners of sources that
                 participate in Web service-based data publishing
                 systems to publish a restricted set of parameterized
                 queries over the schema of their sources and package
                 them as WSDL services. The sources may be relational
                 databases, which naturally have a schema, or ad hoc
                 information/application systems whereas the owner
                 publishes a virtual schema. CLIDE allows information
                 clients to pose queries over the published schema and
                 utilizes prior work on answering queries using views to
                 answer queries that can be processed by combining and
                 processing the results of one or more Web service
                 calls. These queries are called feasible. Contrary to
                 prior work, where infeasible queries are rejected
                 without an explanatory feedback, leading the user into
                 a frustrating trial-and-error cycle, CLIDE features a
                 query formulation interface, which extends the QBE-like
                 query builder of Microsoft's SQL Server with a color
                 scheme that guides the user toward formulating feasible
                 queries. CLIDE guarantees that the suggested query edit
                 actions are complete (i.e., each feasible query can be
                 built by following only suggestions), rapidly
                 convergent (the suggestions are tuned to lead to the
                 closest feasible completions of the query), and
                 suitably summarized (at each interaction step, only a
                 minimal number of actions needed to preserve
                 completeness are suggested). We present the algorithms,
                 implementation, and performance evaluation showing that
                 CLIDE is a viable on-line tool.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "limited access patterns; Middleware; query rewriting;
                 Web services",
}

@Article{Sharfman:2007:GAM,
  author =       "Izchak Sharfman and Assaf Schuster and Daniel Keren",
  title =        "A geometric approach to monitoring threshold functions
                 over distributed data streams",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "23:1--23:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292613",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Monitoring data streams in a distributed system is the
                 focus of much research in recent years. Most of the
                 proposed schemes, however, deal with monitoring simple
                 aggregated values, such as the frequency of appearance
                 of items in the streams. More involved challenges, such
                 as the important task of feature selection (e.g., by
                 monitoring the information gain of various features),
                 still require very high communication overhead using
                 naive, centralized algorithms.\par

                 We present a novel geometric approach which reduces
                 monitoring the value of a function (vis-{\`a}-vis a
                 threshold) to a set of constraints applied locally on
                 each of the streams. The constraints are used to
                 locally filter out data increments that do not affect
                 the monitoring outcome, thus avoiding unnecessary
                 communication. As a result, our approach enables
                 monitoring of arbitrary threshold functions over
                 distributed data streams in an efficient manner.\par

                 We present experimental results on real-world data
                 which demonstrate that our algorithms are highly
                 scalable, and considerably reduce communication load in
                 comparison to centralized algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Distributed monitoring",
}

@Article{VandenBussche:2007:IPS,
  author =       "Jan {Van den Bussche}",
  title =        "Introduction to the {PODS} 2006 special section",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "24:1--24:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292614",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fagin:2007:ISM,
  author =       "Ronald Fagin",
  title =        "Inverting schema mappings",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "25:1--25:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292615",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A schema mapping is a specification that describes how
                 data structured under one schema (the source schema) is
                 to be transformed into data structured under a
                 different schema (the target schema). Although the
                 notion of an inverse of a schema mapping is important,
                 the exact definition of an inverse mapping is somewhat
                 elusive. This is because a schema mapping may associate
                 many target instances with each source instance, and
                 many source instances with each target instance. Based
                 on the notion that the composition of a mapping and its
                 inverse is the identity, we give a formal definition
                 for what it means for a schema mapping $ M \prime $ to
                 be an inverse of a schema mapping $M$ for a class $S$
                 of source instances. We call such an inverse an
                 $S$-inverse. A particular case of interest arises when
                 $S$ is the class of all source instances, in which case
                 an $S$-inverse is a global inverse. We focus on the
                 important and practical case of schema mappings
                 specified by source-to-target tuple-generating
                 dependencies, and uncover a rich theory. When $S$ is
                 specified by a set of dependencies with a finite chase,
                 we show how to construct an $S$-inverse when one
                 exists. In particular, we show how to construct a
                 global inverse when one exists. Given $M$ and $ M
                 \prime $, we show how to define the largest class $S$
                 such that $ M \prime $ is an $S$-inverse of $M$.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "chase; computational complexity; Data exchange; data
                 integration; dependencies; inverse; metadata model
                 management; schema mapping; second-order logic",
}

@Article{Bender:2007:APM,
  author =       "Michael A. Bender and Haodong Hu",
  title =        "An adaptive packed-memory array",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "26:1--26:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292616",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The packed-memory array (PMA) is a data structure that
                 maintains a dynamic set of $N$ elements in sorted order
                 in a $ \Theta (N) $-sized array. The idea is to
                 intersperse $ \Theta (N) $ empty spaces or gaps among
                 the elements so that only a small number of elements
                 need to be shifted around on an insert or delete.
                 Because the elements are stored physically in sorted
                 order in memory or on disk, the PMA can be used to
                 support extremely efficient range queries.
                 Specifically, the cost to scan $L$ consecutive elements
                 is $ O(1 + L / B) $ memory transfers.\par

                 This article gives the first adaptive packed-memory
                 array (APMA), which automatically adjusts to the input
                 pattern. Like the traditional PMA, any pattern of
                 updates costs only $ O(\log_2 N) $ amortized element
                 moves and $ O(1 + (\log_2 N) / B) $ amortized memory
                 transfers per update. However, the APMA performs even
                 better on many common input distributions achieving
                 only $ O(\log N) $ amortized element moves and $ O(1 +
                 (\log N) / B) $ amortized memory transfers. The article
                 analyzes sequential inserts, where the insertions are
                 to the front of the APMA, hammer inserts, where the
                 insertions ``hammer'' on one part of the APMA, random
                 inserts, where the insertions are after random elements
                 in the APMA, and bulk inserts, where for constant $
                 \alpha $ $ \epsilon $ $ [0, 1] $, $N$ $ \alpha $
                 elements are inserted after random elements in the
                 APMA. The article then gives simulation results that
                 are consistent with the asymptotic bounds. For
                 sequential insertions of roughly 1.4 million elements,
                 the APMA has four times fewer element moves per
                 insertion than the traditional PMA and running times
                 that are more than seven times faster.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "adaptive packed-memory array; cache oblivious;
                 locality preserving; packed-memory array; range query;
                 rebalance; sequential file maintenance; sequential
                 scan; sparse array",
}

@Article{Ioannidis:2007:IES,
  author =       "Yannis Ioannidis",
  title =        "Introduction to the {EDBT} 2006 special section",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "27:1--27:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292617",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bruno:2007:PDR,
  author =       "Nicolas Bruno and Surajit Chaudhuri",
  title =        "Physical design refinement: {The} `merge-reduce'
                 approach",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "28:1--28:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292618",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Physical database design tools rely on a DBA-provided
                 workload to pick an ``optimal'' set of indexes and
                 materialized views. Such tools allow either creating a
                 new such configuration or adding new structures to
                 existing ones. However, these tools do not provide
                 adequate support for the incremental and flexible
                 refinement of existing physical structures. Although
                 such refinements are often very valuable for DBAs, a
                 completely manual approach to refinement can lead to
                 infeasible solutions (e.g., excessive use of space). In
                 this article, we focus on the important problem of
                 physical design refinement and propose a
                 transformational architecture that is based upon two
                 novel primitive operations, called merging and
                 reduction. These operators help refine a configuration,
                 treating indexes and materialized views in a unified
                 way, as well as succinctly explain the refinement
                 process to DBAs.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Physical database design; physical design refinment;
                 view merging and reduction",
}

@Article{Skopal:2007:UFF,
  author =       "Tom{\'a}{\v{s}} Skopal",
  title =        "Unified framework for fast exact and approximate
                 search in dissimilarity spaces",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "29:1--29:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292619",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In multimedia systems we usually need to retrieve
                 database (DB) objects based on their similarity to a
                 query object, while the similarity assessment is
                 provided by a measure which defines a (dis)similarity
                 score for every pair of DB objects. In most existing
                 applications, the similarity measure is required to be
                 a metric, where the triangle inequality is utilized to
                 speed up the search for relevant objects by use of
                 metric access methods (MAMs), for example, the M-tree.
                 A recent research has shown, however, that nonmetric
                 measures are more appropriate for similarity modeling
                 due to their robustness and ease to model a
                 made-to-measure similarity. Unfortunately, due to the
                 lack of triangle inequality, the nonmetric measures
                 cannot be directly utilized by MAMs. From another point
                 of view, some sophisticated similarity measures could
                 be available in a black-box nonanalytic form (e.g., as
                 an algorithm or even a hardware device), where no
                 information about their topological properties is
                 provided, so we have to consider them as nonmetric
                 measures as well. From yet another point of view, the
                 concept of similarity measuring itself is inherently
                 imprecise and we often prefer fast but approximate
                 retrieval over an exact but slower one.\par

                 To date, the mentioned aspects of similarity retrieval
                 have been solved separately, that is, exact versus
                 approximate search or metric versus nonmetric search.
                 In this article we introduce a similarity retrieval
                 framework which incorporates both of the aspects into a
                 single unified model. Based on the framework, we show
                 that for any dissimilarity measure (either a metric or
                 nonmetric) we are able to change the ``amount'' of
                 triangle inequality, and so obtain an approximate or
                 full metric which can be used for MAM-based retrieval.
                 Due to the varying ``amount'' of triangle inequality,
                 the measure is modified in a way suitable for either an
                 exact but slower or an approximate but faster
                 retrieval. Additionally, we introduce the TriGen
                 algorithm aimed at constructing the desired
                 modification of any black-box distance automatically,
                 using just a small fraction of the database.",
  acknowledgement = ack-nhfb,
  articleno =    "29",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate and exact search; Similarity retrieval",
}

@Article{Li:2007:NGN,
  author =       "Yunyao Li and Huahai Yang and H. V. Jagadish",
  title =        "{NaLIX}: a generic natural language search environment
                 for {XML} data",
  journal =      j-TODS,
  volume =       "32",
  number =       "4",
  pages =        "30:1--30:??",
  month =        nov,
  year =         "2007",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1292609.1292620",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:32 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We describe the construction of a generic natural
                 language query interface to an XML database. Our
                 interface can accept a large class of English sentences
                 as a query, which can be quite complex and include
                 aggregation, nesting, and value joins, among other
                 things. This query is translated, potentially after
                 reformulation, into an XQuery expression. The
                 translation is based on mapping grammatical proximity
                 of natural language parsed tokens in the parse tree of
                 the query sentence to proximity of corresponding
                 elements in the XML data to be retrieved. Iterative
                 search in the form of followup queries is also
                 supported. Our experimental assessment, through a user
                 study, demonstrates that this type of natural language
                 interface is good enough to be usable now, with no
                 restrictions on the application domain.",
  acknowledgement = ack-nhfb,
  articleno =    "30",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "dialog system; iterative search; Natural language
                 interface; XML; XQuery",
}

@Article{Hristidis:2008:ABK,
  author =       "Vagelis Hristidis and Heasoo Hwang and Yannis
                 Papakonstantinou",
  title =        "Authority-based keyword search in databases",
  journal =      j-TODS,
  volume =       "33",
  number =       "1",
  pages =        "1:1--1:40",
  month =        mar,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1331904.1331905",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:49 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Our system applies authority-based ranking to keyword
                 search in databases modeled as labeled graphs. Three
                 ranking factors are used: the relevance to the query,
                 the specificity and the importance of the result. All
                 factors are handled using authority-flow techniques
                 that exploit the link-structure of the data graph, in
                 contrast to traditional Information Retrieval. We
                 address the performance challenges in computing the
                 authority flows in databases by using precomputation
                 and exploiting the database schema if present. We
                 conducted user surveys and performance experiments on
                 multiple real and synthetic datasets, to assess the
                 semantic meaningfulness and performance of our
                 system.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Authority flow; PageRank; quality experiments;
                 ranking; specificity",
}

@Article{Aggarwal:2008:SDM,
  author =       "Charu C. Aggarwal and Philip S. Yu",
  title =        "On static and dynamic methods for condensation-based
                 privacy-preserving data mining",
  journal =      j-TODS,
  volume =       "33",
  number =       "1",
  pages =        "2:1--2:39",
  month =        mar,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1331904.1331906",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:49 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In recent years, privacy-preserving data mining has
                 become an important problem because of the large amount
                 of personal data which is tracked by many business
                 applications. In many cases, users are unwilling to
                 provide personal information unless the privacy of
                 sensitive information is guaranteed. In this paper, we
                 propose a new framework for privacy-preserving data
                 mining of multidimensional data. Previous work for
                 privacy-preserving data mining uses a perturbation
                 approach which reconstructs data distributions in order
                 to perform the mining. Such an approach treats each
                 dimension independently and therefore ignores the
                 correlations between the different dimensions. In
                 addition, it requires the development of a new
                 distribution-based algorithm for each data mining
                 problem, since it does not use the multidimensional
                 records, but uses aggregate distributions of the data
                 as input. This leads to a fundamental re-design of data
                 mining algorithms. In this paper, we will develop a new
                 and flexible approach for privacy-preserving data
                 mining that does not require new problem-specific
                 algorithms, since it maps the original data set into a
                 new anonymized data set. These anonymized data closely
                 match the characteristics of the original data
                 including the correlations among the different
                 dimensions. We will show how to extend the method to
                 the case of data streams. We present empirical results
                 illustrating the effectiveness of the method. We also
                 show the efficiency of the method for data streams.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "databases data mining; k -anonymity; Privacy",
}

@Article{Balazinska:2008:FTB,
  author =       "Magdalena Balazinska and Hari Balakrishnan and Samuel
                 R. Madden and Michael Stonebraker",
  title =        "Fault-tolerance in the {Borealis} distributed stream
                 processing system",
  journal =      j-TODS,
  volume =       "33",
  number =       "1",
  pages =        "3:1--3:44",
  month =        mar,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1331904.1331907",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:49 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Over the past few years, Stream Processing Engines
                 (SPEs) have emerged as a new class of software systems,
                 enabling low latency processing of streams of data
                 arriving at high rates. As SPEs mature and get used in
                 monitoring applications that must continuously run
                 (e.g., in network security monitoring), a significant
                 challenge arises: SPEs must be able to handle various
                 software and hardware faults that occur, masking them
                 to provide high availability (HA). In this article, we
                 develop, implement, and evaluate DPC (Delay, Process,
                 and Correct), a protocol to handle crash failures of
                 processing nodes and network failures in a distributed
                 SPE.\par

                 Like previous approaches to HA, DPC uses replication
                 and masks many types of node and network failures. In
                 the presence of network partitions, the designer of any
                 replication system faces a choice between providing
                 availability or data consistency across the replicas.
                 In DPC, this choice is made explicit: the user
                 specifies an availability bound (no result should be
                 delayed by more than a specified delay threshold even
                 under failure if the corresponding input is available),
                 and DPC attempts to minimize the resulting
                 inconsistency between replicas (not all of which might
                 have seen the input data) while meeting the given delay
                 threshold. Although conceptually simple, the DPC
                 protocol tolerates the occurrence of multiple
                 simultaneous failures as well as any further failures
                 that occur during recovery.\par

                 This article describes DPC and its implementation in
                 the Borealis SPE. We show that DPC enables a
                 distributed SPE to maintain low-latency processing at
                 all times, while also achieving eventual consistency,
                 where applications eventually receive the complete and
                 correct output streams. Furthermore, we show that,
                 independent of system size and failure location, it is
                 possible to handle failures almost up-to the
                 user-specified bound in a manner that meets the
                 required availability without introducing any
                 inconsistency.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "availability; consistency; Distributed stream
                 processing; fault-tolerance",
}

@Article{Fan:2008:IPX,
  author =       "Wenfei Fan and Philip Bohannon",
  title =        "Information preserving {XML} schema embedding",
  journal =      j-TODS,
  volume =       "33",
  number =       "1",
  pages =        "4:1--4:44",
  month =        mar,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1331904.1331908",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:49 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A fundamental concern of data integration in an XML
                 context is the ability to embed one or more source
                 documents in a target document so that (a) the target
                 document conforms to a target schema and (b) the
                 information in the source documents is preserved. In
                 this paper, information preservation for XML is
                 formally studied, and the results of this study guide
                 the definition of a novel notion of schema embedding
                 between two XML DTD schemas represented as graphs.
                 Schema embedding generalizes the conventional notion of
                 graph similarity by allowing an edge in a source DTD
                 schema to be mapped to a path in the target DTD.
                 Instance-level embeddings can be derived from the
                 schema embedding in a straightforward manner, such that
                 conformance to a target schema and information
                 preservation are guaranteed. We show that it is
                 NP-complete to find an embedding between two DTD
                 schemas. We also outline efficient heuristic algorithms
                 to find candidate embeddings, which have proved
                 effective by our experimental study. These yield the
                 first systematic and effective approach to finding
                 information preserving XML mappings.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data transformation; information integration;
                 information preservation; schema embedding; schema
                 mapping; XML; XSLT",
}

@Article{Sharaf:2008:AMP,
  author =       "Mohamed A. Sharaf and Panos K. Chrysanthis and
                 Alexandros Labrinidis and Kirk Pruhs",
  title =        "Algorithms and metrics for processing multiple
                 heterogeneous continuous queries",
  journal =      j-TODS,
  volume =       "33",
  number =       "1",
  pages =        "5:1--5:44",
  month =        mar,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1331904.1331909",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jun 12 16:37:49 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The emergence of monitoring applications has
                 precipitated the need for Data Stream Management
                 Systems (DSMSs), which constantly monitor incoming data
                 feeds (through registered continuous queries), in order
                 to detect events of interest. In this article, we
                 examine the problem of how to schedule multiple
                 Continuous Queries (CQs) in a DSMS to optimize
                 different Quality of Service (QoS) metrics. We show
                 that, unlike traditional online systems, scheduling
                 policies in DSMSs that optimize for average response
                 time will be different from policies that optimize for
                 average slowdown, which is a more appropriate metric to
                 use in the presence of a heterogeneous workload.
                 Towards this, we propose policies to optimize for the
                 average-case performance for both metrics.
                 Additionally, we propose a hybrid scheduling policy
                 that strikes a fine balance between performance and
                 fairness, by looking at both the average- and
                 worst-case performance, for both metrics. We also show
                 how our policies can be adaptive enough to handle the
                 inherent dynamic nature of monitoring applications.
                 Furthermore, we discuss how our policies can be
                 efficiently implemented and extended to exploit sharing
                 in optimized multi-query plans and multi-stream CQs.
                 Finally, we experimentally show using real data that
                 our policies consistently outperform currently used
                 ones.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "continuous queries; Data stream management system;
                 operator scheduling",
}

@Article{Fan:2008:CFD,
  author =       "Wenfei Fan and Floris Geerts and Xibei Jia and
                 Anastasios Kementsietsidis",
  title =        "Conditional functional dependencies for capturing data
                 inconsistencies",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "6:1--6:48",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366103",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We propose a class of integrity constraints for
                 relational databases, referred to as {\em conditional
                 functional dependencies\/} (CFDs), and study their
                 applications in data cleaning. In contrast to
                 traditional functional dependencies (FDs) that were
                 developed mainly for schema design, CFDs aim at
                 capturing the consistency of data by enforcing bindings
                 of semantically related values. For static analysis of
                 CFDs we investigate {\em the consistency problem},
                 which is to determine whether or not there exists a
                 nonempty database satisfying a given set of CFDs, and
                 {\em the implication problem}, which is to decide
                 whether or not a set of CFDs entails another CFD. We
                 show that while any set of transitional FDs is
                 trivially consistent, the consistency problem is
                 NP-complete for CFDs, but it is in PTIME when either
                 the database schema is predefined or no attributes
                 involved in the CFDs have a finite domain. For the
                 implication analysis of CFDs, we provide an inference
                 system analogous to Armstrong's axioms for FDs, and
                 show that the implication problem is coNP-complete for
                 CFDs in contrast to the linear-time complexity for
                 their traditional counterpart. We also present an
                 algorithm for computing a minimal cover of a set of
                 CFDs. Since CFDs allow data bindings, in some cases
                 CFDs may be physically large, complicating the
                 detection of constraint violations. We develop
                 techniques for detecting CFD violations in SQL as well
                 as novel techniques for checking multiple constraints
                 by a single query. We also provide incremental methods
                 for checking CFDs in response to changes to the
                 database. We experimentally verify the effectiveness of
                 our CFD-based methods for inconsistency detection. This
                 work not only yields a constraint theory for CFDs but
                 is also a step toward a practical constraint-based
                 method for improving data quality.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data cleaning; functional dependency; SQL",
}

@Article{Jacox:2008:MSS,
  author =       "Edwin H. Jacox and Hanan Samet",
  title =        "Metric space similarity joins",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "7:1--7:38",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366104",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Similarity join algorithms find pairs of objects that
                 lie within a certain distance $ \epsilon $ of each
                 other. Algorithms that are adapted from spatial join
                 techniques are designed primarily for data in a vector
                 space and often employ some form of a multidimensional
                 index. For these algorithms, when the data lies in a
                 metric space, the usual solution is to embed the data
                 in vector space and then make use of a multidimensional
                 index. Such an approach has a number of drawbacks when
                 the data is high dimensional as we must eventually find
                 the most discriminating dimensions, which is not
                 trivial. In addition, although the maximum distance
                 between objects increases with dimension, the ability
                 to discriminate between objects in each dimension does
                 not. These drawbacks are overcome via the introduction
                 of a new method called {\em Quickjoin\/} that does not
                 require a multidimensional index and instead adapts
                 techniques used in distance-based indexing for use in a
                 method that is conceptually similar to the Quicksort
                 algorithm. A formal analysis is provided of the
                 Quickjoin method. Experiments show that the Quickjoin
                 method significantly outperforms two existing
                 techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "distance-based indexing; external memory algorithms;
                 nearest neighbor queries; range queries; ranking;
                 similarity join",
}

@Article{He:2008:COD,
  author =       "Bingsheng He and Qiong Luo",
  title =        "Cache-oblivious databases: {Limitations} and
                 opportunities",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "8:1--8:42",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366105",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Cache-oblivious techniques, proposed in the theory
                 community, have optimal asymptotic bounds on the amount
                 of data transferred between any two adjacent levels of
                 an arbitrary memory hierarchy. Moreover, this optimal
                 performance is achieved without any hardware platform
                 specific tuning. These properties are highly attractive
                 to autonomous databases, especially because the
                 hardware architectures are becoming increasingly
                 complex and diverse.\par

                 In this article, we present our design, implementation,
                 and evaluation of the first cache-oblivious in-memory
                 query processor, EaseDB. Moreover, we discuss the
                 inherent limitations of the cache-oblivious approach as
                 well as the opportunities given by the upcoming
                 hardware architectures. Specifically, a cache-oblivious
                 technique usually requires sophisticated algorithm
                 design to achieve a comparable performance to its
                 cache-conscious counterpart. Nevertheless, this
                 development-time effort is compensated by the
                 automaticity of performance achievement and the reduced
                 ownership cost. Furthermore, this automaticity enables
                 cache-oblivious techniques to outperform their
                 cache-conscious counterparts in multi-threading
                 processors.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "cache-conscious; cache-oblivious; chip
                 multiprocessors; data caches; simultaneous
                 multithreading",
}

@Article{Cormode:2008:ACQ,
  author =       "Graham Cormode and Minos Garofalakis",
  title =        "Approximate continuous querying over distributed
                 streams",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366106",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "While traditional database systems optimize for
                 performance on one-shot query processing, emerging
                 large-scale monitoring applications require continuous
                 tracking of complex data-analysis queries over
                 collections of physically distributed streams. Thus,
                 effective solutions have to be simultaneously
                 space/time efficient (at each remote monitor site),
                 communication efficient (across the underlying
                 communication network), and provide continuous,
                 guaranteed-quality approximate query answers. In this
                 paper, we propose novel algorithmic solutions for the
                 problem of continuously tracking a broad class of
                 complex aggregate queries in such a distributed-streams
                 setting. Our tracking schemes maintain approximate
                 query answers with provable error guarantees, while
                 simultaneously optimizing the storage space and
                 processing time at each remote site, and the
                 communication cost across the network. In a nutshell,
                 our algorithms rely on tracking general-purpose
                 randomized sketch summaries of local streams at remote
                 sites along with concise prediction models of local
                 site behavior in order to produce highly communication-
                 and space/time-efficient solutions. The end result is a
                 powerful approximate query tracking framework that
                 readily incorporates several complex analysis queries
                 (including distributed join and multi-join aggregates,
                 and approximate wavelet representations), thus giving
                 the first known low-overhead tracking solution for such
                 queries in the distributed-streams model. Experiments
                 with real data validate our approach, revealing
                 significant savings over naive solutions as well as our
                 analytical worst-case guarantees.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate query processing; continuous distributed
                 monitoring; data stream algorithms; data synopses",
}

@Article{Eiter:2008:RLQ,
  author =       "Thomas Eiter and Michael Fink and Gianluigi Greco and
                 Domenico Lembo",
  title =        "Repair localization for query answering from
                 inconsistent databases",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366107",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Query answering from inconsistent databases amounts to
                 finding ``meaningful'' answers to queries posed over
                 database instances that do not satisfy integrity
                 constraints specified over their schema. A declarative
                 approach to this problem relies on the notion of
                 repair, that is, a database that satisfies integrity
                 constraints and is obtained from the original
                 inconsistent database by ``minimally'' adding and/or
                 deleting tuples. Consistent answers to a user query are
                 those answers that are in the evaluation of the query
                 over each repair. Motivated by the fact that computing
                 consistent answers from inconsistent databases is in
                 general intractable, the present paper investigates
                 techniques that allow to localize the difficult part of
                 the computation on a small fragment of the database at
                 hand, called ``affected'' part. Based on a number of
                 localization results, an approach to query answering
                 from inconsistent data is presented, in which the query
                 is evaluated over each of the repairs of the affected
                 part only, augmented with the part that is not
                 affected. Single query results are then suitably
                 recombined. For some relevant settings, techniques are
                 also discussed to factorize repairs into components
                 that can be processed independently of one another,
                 thereby guaranteeing exponential gain w.r.t. the basic
                 approach, which is not based on localization. The
                 effectiveness of the results is demonstrated for
                 consistent query answering over expressive schemas,
                 based on logic programming specifications as proposed
                 in the literature.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "consistent query answering; data integration; database
                 repairs; inconsistency management in databases; logic
                 programming; stable models",
}

@Article{Fagin:2008:QIS,
  author =       "Ronald Fagin and Phokion G. Kolaitis and Lucian Popa
                 and Wang-Chiew Tan",
  title =        "Quasi-inverses of schema mappings",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366108",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Schema mappings are high-level specifications that
                 describe the relationship between two database schemas.
                 Two operators on schema mappings, namely the
                 composition operator and the inverse operator, are
                 regarded as especially important. Progress on the study
                 of the inverse operator was not made until very
                 recently, as even finding the exact semantics of this
                 operator turned out to be a fairly delicate task.
                 Furthermore, this notion is rather restrictive, since
                 it is rare that a schema mapping possesses an
                 inverse.\par

                 In this article, we introduce and study the notion of a
                 quasi-inverse of a schema mapping. This notion is a
                 principled relaxation of the notion of an inverse of a
                 schema mapping; intuitively, it is obtained from the
                 notion of an inverse by not differentiating between
                 instances that are equivalent for data-exchange
                 purposes. For schema mappings specified by
                 source-to-target tuple-generating dependencies (s-t
                 tgds), we give a necessary and sufficient combinatorial
                 condition for the existence of a quasi-inverse, and
                 then use this condition to obtain both positive and
                 negative results about the existence of quasi-inverses.
                 In particular, we show that every LAV (local-as-view)
                 schema mapping has a quasi-inverse, but that there are
                 schema mappings specified by full s-t tgds that have no
                 quasi-inverse. After this, we study the language needed
                 to express quasi-inverses of schema mappings specified
                 by s-t tgds, and we obtain a complete characterization.
                 We also characterize the language needed to express
                 inverses of schema mappings, and thereby solve a
                 problem left open in the earlier study of the inverse
                 operator. Finally, we show that quasi-inverses can be
                 used in many cases to recover the data that was
                 exported by the original schema mapping when performing
                 data exchange.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "chase; data exchange; data integration; dependencies;
                 inverse; metadata model management; quasi-inverse;
                 schema mapping",
}

@Article{Zhang:2008:CTA,
  author =       "Donghui Zhang and Alexander Markowetz and Vassilis J.
                 Tsotras and Dimitrios Gunopulos and Bernhard Seeger",
  title =        "On computing temporal aggregates with range
                 predicates",
  journal =      j-TODS,
  volume =       "33",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1366102.1366109",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jun 25 08:39:17 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Computing temporal aggregates is an important but
                 costly operation for applications that maintain
                 time-evolving data (data warehouses, temporal
                 databases, etc.) Due to the large volume of such data,
                 performance improvements for temporal aggregate queries
                 are critical. Previous approaches have aggregate
                 predicates that involve only the time dimension. In
                 this article we examine techniques to compute temporal
                 aggregates that include key-range predicates as well
                 ({\em range-temporal aggregates\/}). In particular we
                 concentrate on the SUM aggregate, while COUNT is a
                 special case. To handle arbitrary key ranges, previous
                 methods would need to keep a separate index for every
                 possible key range. We propose an approach based on a
                 new index structure called the {\em Multiversion
                 SB-Tree}, which incorporates features from both the
                 SB-Tree and the Multiversion B+--tree, to handle
                 arbitrary key-range temporal aggregate queries. We
                 analyze the performance of our approach and present
                 experimental results that show its efficiency.
                 Furthermore, we address a novel and practical variation
                 called {\em functional\/} range-temporal aggregates.
                 Here, the value of any record is a function over time.
                 The meaning of aggregates is altered such that the
                 contribution of a record to the aggregate result is
                 proportional to the size of the intersection between
                 the record's time interval and the query time interval.
                 Both analytical and experimental results show the
                 efficiency of our result.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "functional aggregates; indexing; range predicates;
                 temporal aggregates",
}

@Article{Soliman:2008:PTR,
  author =       "Mohamed A. Soliman and Ihab F. Ilyas and Kevin
                 Chen--Chuan Chang",
  title =        "Probabilistic top-$k$ and ranking-aggregate queries",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "13:1--13:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386119",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Ranking and aggregation queries are widely used in
                 data exploration, data analysis, and decision-making
                 scenarios. While most of the currently proposed ranking
                 and aggregation techniques focus on deterministic data,
                 several emerging applications involve data that is
                 unclean or uncertain. Ranking and aggregating uncertain
                 (probabilistic) data raises new challenges in query
                 semantics and processing, making conventional methods
                 inapplicable. Furthermore, uncertainty imposes
                 probability as a new ranking dimension that does not
                 exist in the traditional settings.\par

                 In this article we introduce new probabilistic
                 formulations for top-$k$ and ranking-aggregate queries
                 in probabilistic databases. Our formulations are based
                 on marriage of traditional top-$k$ semantics with
                 possible worlds semantics. In the light of these
                 formulations, we construct a generic processing
                 framework supporting both query types, and leveraging
                 existing query processing and indexing capabilities in
                 current RDBMSs. The framework encapsulates a state
                 space model and efficient search algorithms to compute
                 query answers. Our proposed techniques minimize the
                 number of accessed tuples and the size of materialized
                 search space to compute query answers. Our experimental
                 study shows the efficiency of our techniques under
                 different data distributions with orders of magnitude
                 improvement over na{\"\i}ve methods.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregation; probabilistic data; Query processing;
                 ranking; top-k",
}

@Article{Ke:2008:CPM,
  author =       "Yiping Ke and James Cheng and Wilfred Ng",
  title =        "Correlated pattern mining in quantitative databases",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "14:1--14:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386120",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study mining correlations from quantitative
                 databases and show that this is a more effective
                 approach than mining associations to discover useful
                 patterns. We propose the novel notion of {\em
                 quantitative correlated pattern\/} (QCP), which is
                 founded on two formal concepts, mutual information and
                 all-confidence. We first devise a normalization on
                 mutual information and apply it to the problem of QCP
                 mining to capture the dependency between the
                 attributes. We further adopt all-confidence as a
                 quality measure to ensure, at a finer granularity, the
                 dependency between the attributes with specific
                 quantitative intervals. We also propose an effective
                 supervised method that combines the consecutive
                 intervals of the quantitative attributes based on
                 mutual information, such that the interval-combining is
                 guided by the dependency between the attributes. We
                 develop an algorithm, {\em QCoMine}, to mine QCPs
                 efficiently by utilizing normalized mutual information
                 and all-confidence to perform bilevel pruning. We also
                 identify the redundancy existing in the set of QCPs and
                 propose effective techniques to eliminate the
                 redundancy. Our extensive experiments on both real and
                 synthetic datasets verify the efficiency of {\em
                 QCoMine\/} and the quality of the QCPs. The
                 experimental results also justify the effectiveness of
                 our proposed techniques for redundancy elimination. To
                 further demonstrate the usefulness and the quality of
                 QCPs, we study an application of QCPs to
                 classification. We demonstrate that the classifier
                 built on the QCPs achieves higher classification
                 accuracy than the state-of-the-art classifiers built on
                 association rules.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "correlated patterns; information-theoretic approach;
                 mutual information; Quantitative databases",
}

@Article{Rusu:2008:SSJ,
  author =       "Florin Rusu and Alin Dobra",
  title =        "Sketches for size of join estimation",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386121",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Sketching techniques provide approximate answers to
                 aggregate queries both for data-streaming and
                 distributed computation. Small space summaries that
                 have linearity properties are required for both types
                 of applications. The prevalent method for analyzing
                 sketches uses moment analysis and
                 distribution-independent bounds based on moments. This
                 method produces clean, easy to interpret, theoretical
                 bounds that are especially useful for deriving
                 asymptotic results. However, the theoretical bounds
                 obscure fine details of the behavior of various
                 sketches and they are mostly not indicative of which
                 type of sketches should be used in practice. Moreover,
                 no significant empirical comparison between various
                 sketching techniques has been published, which makes
                 the choice even harder. In this article we take a close
                 look at the sketching techniques proposed in the
                 literature from a statistical point of view with the
                 goal of determining properties that indicate the actual
                 behavior and producing tighter confidence bounds.
                 Interestingly, the statistical analysis reveals that
                 two of the techniques, Fast-AGMS and Count-Min, provide
                 results that are in some cases orders of magnitude
                 better than the corresponding theoretical predictions.
                 We conduct an extensive empirical study that compares
                 the different sketching techniques in order to
                 corroborate the statistical analysis with the
                 conclusions we draw from it. The study indicates the
                 expected performance of various sketches, which is
                 crucial if the techniques are to be used by
                 practitioners. The overall conclusion of the study is
                 that Fast-AGMS sketches are, for the full spectrum of
                 problems, either the best, or close to the best,
                 sketching technique. We apply the insights obtained
                 from the statistical study and the experimental results
                 to design effective algorithms for sketching interval
                 data. We show how the two basic methods for sketching
                 interval data, DMAP and fast range-summation, can be
                 improved significantly with respect to the update time
                 without a significant loss in accuracy. The gain in
                 update time can be as large as two orders of magnitude,
                 thus making the improved methods practical. The
                 empirical study suggests that DMAP is preferable when
                 update time is the critical requirement and fast
                 range-summation is desirable for better accuracy.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "AGMS sketches; Count-Min sketches; DMAP; fast
                 range-summation; Fast-AGMS sketches; Fast-Count
                 sketches; Size of join estimation",
}

@Article{Xu:2008:CBS,
  author =       "Fei Xu and Christopher Jermaine and Alin Dobra",
  title =        "Confidence bounds for sampling-based group by
                 estimates",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386122",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Sampling is now a very important data management tool,
                 to such an extent that an interface for database
                 sampling is included in the latest SQL standard. In
                 this article we reconsider in depth what at first may
                 seem like a very simple problem --- computing the error
                 of a sampling-based guess for the answer to a GROUP BY
                 query over a multitable join. The difficulty when
                 sampling for the answer to such a query is that the
                 same sample will be used to guess the result of the
                 query for each group, which induces correlations among
                 the estimates. Thus, from a statistical point-of-view
                 it is very problematic and even dangerous to use
                 traditional methods such as confidence intervals for
                 communicating estimate accuracy to the user. We explore
                 ways to address this problem, and pay particular
                 attention to the computational aspects of computing
                 ``safe'' confidence intervals.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Approximate query processing; multiple hypothesis
                 testing; sampling",
}

@Article{LeFevre:2008:WAA,
  author =       "Kristen LeFevre and David J. DeWitt and Raghu
                 Ramakrishnan",
  title =        "Workload-aware anonymization techniques for
                 large-scale datasets",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386123",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Protecting individual privacy is an important problem
                 in microdata distribution and publishing. Anonymization
                 algorithms typically aim to satisfy certain privacy
                 definitions with minimal impact on the quality of the
                 resulting data. While much of the previous literature
                 has measured quality through simple one-size-fits-all
                 measures, we argue that quality is best judged with
                 respect to the workload for which the data will
                 ultimately be used.\par

                 This article provides a suite of anonymization
                 algorithms that incorporate a target class of
                 workloads, consisting of one or more data mining tasks
                 as well as selection predicates. An extensive empirical
                 evaluation indicates that this approach is often more
                 effective than previous techniques. In addition, we
                 consider the problem of scalability. The article
                 describes two extensions that allow us to scale the
                 anonymization algorithms to datasets much larger than
                 main memory. The first extension is based on ideas from
                 scalable decision trees, and the second is based on
                 sampling. A thorough performance evaluation indicates
                 that these techniques are viable in practice.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "anonymity; data mining; Databases; performance;
                 privacy; scalability",
}

@Article{Karras:2008:HSO,
  author =       "Panagiotis Karras and Nikos Mamoulis",
  title =        "Hierarchical synopses with optimal error guarantees",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386124",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Hierarchical synopsis structures offer a viable
                 alternative in terms of efficiency and flexibility in
                 relation to traditional summarization techniques such
                 as histograms. Previous research on such structures has
                 mostly focused on a single model, based on the Haar
                 wavelet decomposition. In previous work, we have
                 introduced a more refined, wavelet-inspired
                 hierarchical index structure for synopsis construction:
                 the Haar$^+$ tree. The chief advantages of this
                 structure are twofold. First, it achieves higher
                 synopsis quality at the task of summarizing data sets
                 with sharp discontinuities than state-of-the-art
                 histogram and Haar wavelet techniques. Second, thanks
                 to its search space delimitation capacity, Haar$^+$
                 synopsis construction operates in time {\em linear\/}
                 in the size of the data set for {\em any\/} monotonic
                 distributive error metric. Contemporaneous research has
                 introduced another hierarchical synopsis structure, the
                 compact hierarchical histogram (CHH). In this article,
                 we elaborate on both these structures. First, we
                 formally prove that the CHH, in its default
                 binary-hierarchy form, is a simplified variant of a
                 Haar$^+$ tree. We then focus on the summarization
                 problem, with both these hierarchical synopsis
                 structures, in which an error guarantee expressed by a
                 {\em maximum-error\/} metric is required. We show that
                 this problem is most efficiently solved through its
                 dual, space-minimization counterpart, which can also
                 achieve {\em optimal quality}. In this case, there is a
                 benefit to be gained by specializing the algorithm for
                 each structure; hence, our algorithm for
                 optimal-quality maximum-error CHH requires {\em low
                 polynomial\/} time; on the other hand, optimal-quality
                 Haar$^+$ synopses for maximum-error metrics are
                 constructed in exponential time; hence, we also develop
                 a low-polynomial-time approximation scheme for the
                 maximum-error Haar$^+$ case. Furthermore, we extend our
                 approach for both general-error and maximum-error
                 Haar$^+$ synopses to arbitrary dimensionality. In our
                 experimental study, (i) we confirm the theoretically
                 expected superiority of Haar$^+$ synopses over Haar
                 wavelet methods in both construction time and achieved
                 quality for representative error metrics; (ii) we
                 demonstrate that Haar$^+$ synopses are also constructed
                 faster than optimal plain histograms, and, moreover,
                 achieve higher synopsis quality with highly
                 discontinuous data sets; such an advantage of a
                 hierarchical synopsis structure over a histogram had
                 been intuitively expressed, but never experimentally
                 verified; and (iii) we show that Haar$^+$ synopsis
                 quality supersedes that of a CHH.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate query processing; data synopses;
                 Summarization",
}

@Article{Lester:2008:EOI,
  author =       "Nicholas Lester and Alistair Moffat and Justin Zobel",
  title =        "Efficient online index construction for text
                 databases",
  journal =      j-TODS,
  volume =       "33",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1386118.1386125",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 29 14:05:10 MDT 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Inverted index structures are a core element of
                 current text retrieval systems. They can be constructed
                 quickly using offline approaches, in which one or more
                 passes are made over a static set of input data, and,
                 at the completion of the process, an index is available
                 for querying. However, there are search environments in
                 which even a small delay in timeliness cannot be
                 tolerated, and the index must always be queryable and
                 up to date. Here we describe and analyze a {\em
                 geometric partitioning\/} mechanism for online index
                 construction that provides a range of tradeoffs between
                 costs, and can be adapted to different balances of
                 insertion and querying operations. Detailed
                 experimental results are provided that show the extent
                 of these tradeoffs, and that these new methods can
                 yield substantial savings in online indexing costs.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Index construction; index update; search engines; text
                 indexing",
}

@Article{Ozsoyoglu:2008:FTS,
  author =       "Z. Meral {\"O}zsoyo{\u{g}}lu",
  title =        "Foreword to {TODS SIGMOD\slash PODS\slash ICDT 2007}
                 special issue",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "20:1--20:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412332",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ooi:2008:IAS,
  author =       "Beng Chin Ooi",
  title =        "Introduction to {ACM SIGMOD 2007} special section",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "21:1--21:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412333",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Melnik:2008:CMB,
  author =       "Sergey Melnik and Atul Adya and Philip A. Bernstein",
  title =        "Compiling mappings to bridge applications and
                 databases",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "22:1--22:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412334",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Translating data and data access operations between
                 applications and databases is a longstanding data
                 management problem. We present a novel approach to this
                 problem, in which the relationship between the
                 application data and the persistent storage is
                 specified using a declarative mapping, which is
                 compiled into bidirectional views that drive the data
                 transformation engine. Expressing the application model
                 as a view on the database is used to answer queries,
                 while expressing the database schema as a view on the
                 application model allows us to leverage view
                 maintenance algorithms for update translation. This
                 approach has been implemented in a commercial product.
                 It enables developers to interact with a relational
                 database via a conceptual schema and an object-oriented
                 programming surface. We outline the implemented system
                 and focus on the challenges of mapping compilation,
                 which include rewriting queries under constraints and
                 supporting nonrelational constructs.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Mapping; query rewriting; updateable views",
}

@Article{Jermaine:2008:SAQ,
  author =       "Chris Jermaine and Subramanian Arumugam and Abhijit
                 Pol and Alin Dobra",
  title =        "Scalable approximate query processing with the {DBO}
                 engine",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "23:1--23:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412335",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article describes query processing in the DBO
                 database system. Like other database systems designed
                 for ad hoc analytic processing, DBO is able to compute
                 the exact answers to queries over a large relational
                 database in a scalable fashion. Unlike any other system
                 designed for analytic processing, DBO can constantly
                 maintain a guess as to the final answer to an aggregate
                 query throughout execution, along with statistically
                 meaningful bounds for the guess's accuracy. As DBO
                 gathers more and more information, the guess gets more
                 and more accurate, until it is 100\% accurate as the
                 query is completed. This allows users to stop the
                 execution as soon as they are happy with the query
                 accuracy, and thus encourages exploratory data
                 analysis.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Online aggregation; randomized algorithms; sampling",
}

@Article{Libkin:2008:IPS,
  author =       "Leonid Libkin",
  title =        "Introduction to the {PODS 2007} special section",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "24:1--24:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412336",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2008:ECX,
  author =       "Wenfei Fan and Floris Geerts and Frank Neven",
  title =        "Expressiveness and complexity of {XML} publishing
                 transducers",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "25:1--25:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412337",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A number of languages have been developed for
                 specifying XML publishing, that is, transformations of
                 relational data into XML trees. These languages
                 generally describe the behaviors of a middleware
                 controller that builds an output tree iteratively,
                 issuing queries to a relational source and expanding
                 the tree with the query results at each step. To study
                 the complexity and expressive power of XML publishing
                 languages, this article proposes a notion of {\em
                 publishing transducers}, which generate XML trees from
                 relational data. We study a variety of publishing
                 transducers based on what relational queries a
                 transducer can issue, what temporary stores a
                 transducer can use during tree generation, and whether
                 or not some tree nodes are allowed to be virtual, that
                 is, excluded from the output tree. We first show how
                 existing XML publishing languages can be characterized
                 by such transducers, and thus provide a synergy between
                 theory and practice. We then study the membership,
                 emptiness, and equivalence problems for various classes
                 of transducers. We establish lower and upper bounds,
                 all matching, ranging from PTIME to undecidable.
                 Finally, we investigate the expressive power of these
                 transducers and existing languages. We show that when
                 treated as relational query languages, different
                 classes of transducers capture either complexity
                 classes (e.g., PSPACE) or fragments of datalog (e.g.,
                 linear datalog). For tree generation, we establish
                 connections between publishing transducers and logical
                 transductions, among other things.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "complexity; data exchange; expressiveness; transducer;
                 XML publishing",
}

@Article{Jayram:2008:ESA,
  author =       "T. S. Jayram and Andrew McGregor and S. Muthukrishnan
                 and Erik Vee",
  title =        "Estimating statistical aggregates on probabilistic
                 data streams",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "26:1--26:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412338",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The probabilistic stream model was introduced by
                 Jayram et al. [2007]. It is a generalization of the
                 data stream model that is suited to handling {\em
                 probabilistic\/} data, where each item of the stream
                 represents a probability distribution over a set of
                 possible events. Therefore, a probabilistic stream
                 determines a distribution over a potentially
                 exponential number of classical {\em deterministic\/}
                 streams, where each item is deterministically one of
                 the domain values.\par

                 We present algorithms for computing commonly used
                 aggregates on a probabilistic stream. We present the
                 first one pass streaming algorithms for estimating the
                 expected mean of a probabilistic stream. Next, we
                 consider the problem of estimating frequency moments
                 for probabilistic data. We propose a general approach
                 to obtain unbiased estimators working over
                 probabilistic data by utilizing unbiased estimators
                 designed for standard streams. Applying this approach,
                 we extend a classical data stream algorithm to obtain a
                 one-pass algorithm for estimating $ F_2 $, the second
                 frequency moment. We present the first known streaming
                 algorithms for estimating $ F_0 $, the number of
                 distinct items on probabilistic streams. Our work also
                 gives an efficient one-pass algorithm for estimating
                 the median, and a two-pass algorithm for estimating the
                 range.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "frequency moments; mean; median; OLAP; Probabilistic
                 streams",
}

@Article{Schwentick:2008:IIS,
  author =       "Thomas Schwentick and Dan Suciu",
  title =        "Introduction to {ICDT 2007} special section",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "27:1--27:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412339",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Buneman:2008:EIP,
  author =       "Peter Buneman and James Cheney and Stijn Vansummeren",
  title =        "On the expressiveness of implicit provenance in query
                 and update languages",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "28:1--28:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412340",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Information describing the origin of data, generally
                 referred to as {\em provenance}, is important in
                 scientific and curated databases where it is the basis
                 for the trust one puts in their contents. Since such
                 databases are constructed using operations of both
                 query and update languages, it is of paramount
                 importance to describe the effect of these languages on
                 provenance.\par

                 In this article we study provenance for query and
                 update languages that are closely related to SQL, and
                 compare two ways in which they can manipulate
                 provenance so that elements of the input are rearranged
                 to elements of the output: {\em implicit provenance},
                 where a query or update only provides the rearranged
                 output, and provenance is provided implicitly by a
                 default provenance semantics; and {\em explicit
                 provenance}, where a query or update provides both the
                 output and the description of the provenance of each
                 component of the output. Although explicit provenance
                 is in general more expressive, we show that the classes
                 of implicit provenance operations expressible by query
                 and update languages correspond to natural semantic
                 subclasses of the explicit provenance queries.\par

                 One of the consequences of this study is that
                 provenance separates the expressive power of query and
                 update languages. The model is also relevant to
                 annotation propagation schemes in which annotations on
                 the input to a query or update have to be transferred
                 to the output or vice versa.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "conservativity; nested relational calculus; nested
                 update language; Provenance",
}

@Article{Ghelli:2008:CAX,
  author =       "Giorgio Ghelli and Kristoffer Rose and J{\'e}r{\^o}me
                 Sim{\'e}on",
  title =        "Commutativity analysis for {XML} updates",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "29:1--29:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412341",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "An effective approach to support XML updates is to use
                 XQuery extended with update operations. This approach
                 results in very expressive languages which are
                 convenient for users but are difficult to optimize or
                 reason about. A crucial question underlying many static
                 analysis problems for such languages, from optimization
                 to view maintenance, is whether two expressions
                 commute. Unfortunately, commutativity is undecidable
                 for most existing XML update languages. In this
                 article, we propose a conservative analysis for an
                 expressive XML update language that can be used to
                 determine commutativity. The approach relies on a form
                 of path analysis that computes upper bounds for the
                 nodes that are accessed or modified in a given
                 expression. Our main result is a theorem that can be
                 used to identify commuting expressions. We illustrate
                 how the technique applies to concrete examples of query
                 optimization in the presence of updates.",
  acknowledgement = ack-nhfb,
  articleno =    "29",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "commutativity; optimization; updates; XML; XQuery",
}

@Article{Pavlou:2008:FAD,
  author =       "Kyriacos E. Pavlou and Richard T. Snodgrass",
  title =        "Forensic analysis of database tampering",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "30:1--30:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412342",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Regulations and societal expectations have recently
                 expressed the need to mediate access to valuable
                 databases, even by insiders. One approach is tamper
                 detection via cryptographic hashing. This article shows
                 how to determine when the tampering occurred, what data
                 was tampered with, and perhaps, ultimately, who did the
                 tampering, via forensic analysis. We present four
                 successively more sophisticated forensic analysis
                 algorithms: the Monochromatic, RGBY, Tiled Bitmap, and
                 a3D algorithms, and characterize their ``forensic
                 cost'' under worst-case, best-case, and average-case
                 assumptions on the distribution of corruption sites. A
                 lower bound on forensic cost is derived, with RGBY and
                 a3D being shown optimal for a large number of
                 corruptions. We also provide validated cost formul{\ae}
                 for these algorithms and recommendations for the
                 circumstances in which each algorithm is indicated.",
  acknowledgement = ack-nhfb,
  articleno =    "30",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "a3D algorithm; compliant records; forensic analysis
                 algorithm; forensic cost; Monochromatic algorithm;
                 Polychromatic algorithm; RGBY algorithm; Tiled Bitmap
                 algorithm",
}

@Article{Bartolini:2008:ESB,
  author =       "Ilaria Bartolini and Paolo Ciaccia and Marco Patella",
  title =        "Efficient sort-based skyline evaluation",
  journal =      j-TODS,
  volume =       "33",
  number =       "4",
  pages =        "31:1--31:??",
  month =        nov,
  year =         "2008",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1412331.1412343",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 23 11:45:08 MST 2008",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Skyline queries compute the set of Pareto-optimal
                 tuples in a relation, that is, those tuples that are
                 not {\em dominated\/} by any other tuple in the same
                 relation. Although several algorithms have been
                 proposed for efficiently evaluating skyline queries,
                 they either necessitate the relation to have been
                 indexed or have to perform the dominance tests on {\em
                 all\/} the tuples in order to determine the result. In
                 this article we introduce salsa, a novel skyline
                 algorithm that exploits the idea of presorting the
                 input data so as to effectively {\em limit\/} the
                 number of tuples to be read and compared. This makes
                 salsa also attractive when skyline queries are executed
                 on top of systems that do not understand skyline
                 semantics, or when the skyline logic runs on clients
                 with limited power and/or bandwidth. We prove that, if
                 one considers symmetric sorting functions, the number
                 of tuples to be read is minimized by sorting data
                 according to a ``minimum coordinate,'' minC, criterion,
                 and that performance can be further improved if data
                 distribution is known and an asymmetric sorting
                 function is used. Experimental results obtained on
                 synthetic and real datasets show that salsa
                 consistently outperforms state-of-the-art sequential
                 skyline algorithms and that its performance can be
                 accurately predicted.",
  acknowledgement = ack-nhfb,
  articleno =    "31",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Monotone functions; Skyline query",
}

@Article{Mishra:2009:DQM,
  author =       "Chaitanya Mishra and Nick Koudas",
  title =        "The design of a query monitoring system",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "1:1--1:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508858",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Query monitoring refers to the problem of observing
                 and predicting various parameters related to the
                 execution of a query in a database system. In addition
                 to being a useful tool for database users and
                 administrators, it can also serve as an information
                 collection service for resource allocation and adaptive
                 query processing techniques. In this article, we
                 present a query monitoring system from the ground up,
                 describing various new techniques for query monitoring,
                 their implementation inside a real database system, and
                 a novel interface that presents the observed and
                 predicted information in an accessible manner. To
                 enable this system, we introduce several lightweight
                 online techniques for progressively estimating and
                 refining the cardinality of different relational
                 operators using information collected at query
                 execution time. These include binary and multiway joins
                 as well as typical grouping operations and combinations
                 thereof. We describe the various algorithms used to
                 efficiently implement estimators and present the
                 results of an evaluation of a prototype implementation
                 of our framework in an open-source data management
                 system. Our results demonstrate the feasibility and
                 practical utility of the approach presented herein.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "progress estimation; Query monitoring",
}

@Article{Cheng:2009:EQP,
  author =       "James Cheng and Yiping Ke and Wilfred Ng",
  title =        "Efficient query processing on graph databases",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "2:1--2:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508859",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problem of processing {\em subgraph
                 queries\/} on a database that consists of a set of
                 graphs. The answer to a subgraph query is the set of
                 graphs in the database that are supergraphs of the
                 query. In this article, we propose an efficient index,
                 {\em FG*-index}, to solve this problem.\par

                 The cost of processing a subgraph query using most
                 existing indexes mainly consists of two parts: the {\em
                 index probing\/} cost and the {\em candidate
                 verification\/} cost. Index probing is to find the
                 query in the index, or to find the graphs from which we
                 can generate a candidate answer set for the query.
                 Candidate verification is to test whether each graph in
                 the candidate set is indeed a supergraph of the query.
                 We design FG*-index to minimize these two costs as
                 follows.\par

                 FG*-index consists of three components: the {\em
                 FG-index}, the {\em feature-index}, and the {\em
                 FAQ-index}. First, the FG-index employs the concept of
                 {\em Frequent subGraph\/} ({\em FG\/}) to allow the set
                 of queries that are FGs to be answered without
                 candidate verification. We call this set of queries
                 {\em FG-queries}. We can enlarge the set of FG-queries
                 so that more queries can be answered without candidate
                 verification; however, a larger set of FG-queries
                 implies a larger FG-index and hence the index probing
                 cost also increases. We propose the feature-index to
                 reduce the index probing cost. The feature-index uses
                 features to filter false results that are matched in
                 the FG-index, so that we can quickly find the truly
                 matching graphs for a query. For processing
                 non-FG-queries, we propose the FAQ-index, which is
                 dynamically constructed from the set of {\em Frequently
                 Asked non-FG-Queries\/} ({\em FAQs\/}). Using the
                 FAQ-index, verification is not required for processing
                 FAQs and only a small number of candidates need to be
                 verified for processing non-FG-queries that are {\em
                 not frequently asked}. Finally, a comprehensive set of
                 experiments verifies that query processing using
                 FG*-index is up to orders of magnitude more efficient
                 than state-of-the-art indexes and it is also more
                 scalable.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "frequent subgraphs; Graph databases; graph indexing;
                 graph query processing",
}

@Article{Spiegel:2009:TSA,
  author =       "Joshua Spiegel and Neoklis Polyzotis",
  title =        "{TuG} synopses for approximate query answering",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "3:1--3:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508860",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article introduces the Tuple Graph (TuG)
                 synopses, a new class of data summaries that enable
                 accurate approximate answers for complex relational
                 queries. The proposed summarization framework adopts a
                 ``semi-structured'' view of the relational database,
                 modeling a relational data set as a graph of tuples and
                 join queries as graph traversals, respectively. The key
                 idea is to approximate the structure of the induced
                 data graph in a concise synopsis, and to approximate
                 the answer to a query by performing the corresponding
                 traversal over the summarized graph. We detail the
                 (TuG) synopsis model that is based on this novel
                 approach, and we describe an efficient and scalable
                 construction algorithm for building accurate (TuG)
                 within a specific storage budget. We validate the
                 performance of (TuG) with an extensive experimental
                 study on real-life and synthetic datasets. Our results
                 verify the effectiveness of (TuG) in generating
                 accurate approximate answers for complex join queries,
                 and demonstrate their benefits over existing
                 summarization techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate query processing; Data synopses;
                 selectivity estimation",
}

@Article{Kramer:2009:SIC,
  author =       "J{\"u}rgen Kr{\"a}mer and Bernhard Seeger",
  title =        "Semantics and implementation of continuous sliding
                 window queries over data streams",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "4:1--4:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508861",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In recent years the processing of continuous queries
                 over potentially infinite data streams has attracted a
                 lot of research attention. We observed that the
                 majority of work addresses individual stream operations
                 and system-related issues rather than the development
                 of a general-purpose basis for stream processing
                 systems. Furthermore, example continuous queries are
                 often formulated in some declarative query language
                 without specifying the underlying semantics precisely
                 enough. To overcome these deficiencies, this article
                 presents a consistent and powerful operator algebra for
                 data streams which ensures that continuous queries have
                 well-defined, deterministic results. In analogy to
                 traditional database systems, we distinguish between a
                 logical and a physical operator algebra. While the
                 logical algebra specifies the semantics of the
                 individual operators in a descriptive but concrete way
                 over temporal multisets, the physical algebra provides
                 efficient implementations in the form of
                 stream-to-stream operators. By adapting and enhancing
                 research from temporal databases to meet the
                 challenging requirements in streaming applications, we
                 are able to carry over the conventional transformation
                 rules from relational databases to stream processing.
                 For this reason, our approach not only makes it
                 possible to express continuous queries with a sound
                 semantics, but also provides a solid foundation for
                 query optimization, one of the major research topics in
                 the stream community. Since this article seamlessly
                 explains the steps from query formulation to query
                 execution, it outlines the innovative features and
                 operational functionality implemented in our
                 state-of-the-art stream processing infrastructure.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "continuous queries; data streams; query optimization;
                 Semantics",
}

@Article{Jain:2009:QAO,
  author =       "Alpa Jain and Panagiotis G. Ipeirotis",
  title =        "A quality-aware optimizer for information extraction",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "5:1--5:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508862",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A large amount of structured information is buried in
                 unstructured text. Information extraction systems can
                 extract structured relations from the documents and
                 enable sophisticated, SQL-like queries over
                 unstructured text. Information extraction systems are
                 not perfect and their output has imperfect precision
                 and recall (i.e., contains spurious tuples and misses
                 good tuples). Typically, an extraction system has a set
                 of parameters that can be used as ``knobs'' to tune the
                 system to be either precision- or recall-oriented.
                 Furthermore, the choice of documents processed by the
                 extraction system also affects the quality of the
                 extracted relation. So far, estimating the output
                 quality of an information extraction task has been an
                 ad hoc procedure, based mainly on heuristics. In this
                 article, we show how to use Receiver Operating
                 Characteristic (ROC) curves to estimate the extraction
                 quality in a statistically robust way and show how to
                 use ROC analysis to select the extraction parameters in
                 a principled manner. Furthermore, we present analytic
                 models that reveal how different document retrieval
                 strategies affect the quality of the extracted
                 relation. Finally, we present our maximum likelihood
                 approach for estimating, on the fly, the parameters
                 required by our analytic models to predict the runtime
                 and the output quality of each execution plan. Our
                 experimental evaluation demonstrates that our
                 optimization approach predicts accurately the output
                 quality and selects the fastest execution plan that
                 satisfies the output quality restrictions.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Information Extraction; ROC curves",
}

@Article{Considine:2009:RAA,
  author =       "Jeffrey Considine and Marios Hadjieleftheriou and
                 Feifei Li and John Byers and George Kollios",
  title =        "Robust approximate aggregation in sensor data
                 management systems",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "6:1--6:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508863",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In the emerging area of sensor-based systems, a
                 significant challenge is to develop scalable,
                 fault-tolerant methods to extract useful information
                 from the data the sensors collect. An approach to this
                 data management problem is the use of sensor database
                 systems, which allow users to perform aggregation
                 queries such as MIN, COUNT, and AVG on the readings of
                 a sensor network. In addition, more advanced queries
                 such as frequency counting and quantile estimation can
                 be supported. Due to energy limitations in sensor-based
                 networks, centralized data collection is generally
                 impractical, so most systems use in-network aggregation
                 to reduce network traffic. However, even these
                 aggregation strategies remain bandwidth-intensive when
                 combined with the fault-tolerant, multipath routing
                 methods often used in these environments. To avoid this
                 expense, we investigate the use of approximate
                 in-network aggregation using small sketches. We present
                 duplicate-insensitive sketching techniques that can be
                 implemented efficiently on small sensor devices with
                 limited hardware support and we analyze both their
                 performance and accuracy. Finally, we present an
                 experimental evaluation that validates the
                 effectiveness of our methods.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregation; approximation algorithms; Sensor
                 databases; sketches; synopses",
}

@Article{Angiulli:2009:DOP,
  author =       "Fabrizio Angiulli and Fabio Fassetti and Luigi
                 Palopoli",
  title =        "Detecting outlying properties of exceptional objects",
  journal =      j-TODS,
  volume =       "34",
  number =       "1",
  pages =        "7:1--7:??",
  month =        apr,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1508857.1508864",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:22:33 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Assume you are given a data population characterized
                 by a certain number of attributes. Assume, moreover,
                 you are provided with the information that one of the
                 individuals in this data population is abnormal, but no
                 reason whatsoever is given to you as to why this
                 particular individual is to be considered abnormal. In
                 several cases, you will be indeed interested in
                 discovering such reasons. This article is precisely
                 concerned with this problem of discovering sets of
                 attributes that account for the (a priori stated)
                 abnormality of an individual within a given dataset. A
                 criterion is presented to measure the abnormality of
                 combinations of attribute values featured by the given
                 abnormal individual with respect to the reference
                 population. In this respect, each subset of attributes
                 is intended to somehow represent a ``property'' of
                 individuals. We distinguish between global and local
                 properties. Global properties are subsets of attributes
                 explaining the given abnormality with respect to the
                 entire data population. With local ones, instead, two
                 subsets of attributes are singled out, where the former
                 one justifies the abnormality within the data
                 subpopulation selected using the values taken by the
                 exceptional individual on those attributes included in
                 the latter one. The problem of individuating abnormal
                 properties with associated explanations is formally
                 stated and analyzed. Such a formal characterization is
                 then exploited in order to devise efficient algorithms
                 for detecting both global and local forms of most
                 abnormal properties. The experimental evidence, which
                 is accounted for in the article, shows that the
                 algorithms are both able to mine meaningful information
                 and to accomplish the computational task by examining a
                 negligible fraction of the search space.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data mining; knowledge discovery; outlier
                 characterization",
}

@Article{Wong:2009:ABA,
  author =       "Raymond Chi-Wing Wong and Ada Wai-Chee Fu and Ke Wang
                 and Jian Pei",
  title =        "Anonymization-based attacks in privacy-preserving data
                 publishing",
  journal =      j-TODS,
  volume =       "34",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1538909.1538910",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:23:25 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data publishing generates much concern over the
                 protection of individual privacy. Recent studies
                 consider cases where the adversary may possess
                 different kinds of knowledge about the data. In this
                 article, we show that knowledge of the mechanism or
                 algorithm of anonymization for data publication can
                 also lead to extra information that assists the
                 adversary and jeopardizes individual privacy. In
                 particular, all known mechanisms try to minimize
                 information loss and such an attempt provides a
                 loophole for attacks. We call such an attack a
                 minimality attack. In this article, we introduce a
                 model called $m$-confidentiality which deals with
                 minimality attacks, and propose a feasible solution.
                 Our experiments show that minimality attacks are
                 practical concerns on real datasets and that our
                 algorithm can prevent such attacks with very little
                 overhead and information loss.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data publishing; k -anonymity; l -diversity;
                 minimality attack; Privacy preservation",
}

@Article{Ghinita:2009:FED,
  author =       "Gabriel Ghinita and Panagiotis Karras and Panos Kalnis
                 and Nikos Mamoulis",
  title =        "A framework for efficient data anonymization under
                 privacy and accuracy constraints",
  journal =      j-TODS,
  volume =       "34",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1538909.1538911",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:23:25 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Recent research studied the problem of publishing
                 microdata without revealing sensitive information,
                 leading to the privacy-preserving paradigms of
                 $k$-anonymity and $l$-diversity. $k$-anonymity protects
                 against the identification of an individual's record.
                 $l$-diversity, in addition, safeguards against the
                 association of an individual with specific sensitive
                 information. However, existing approaches suffer from
                 at least one of the following drawbacks: (i)
                 $l$-diversification is solved by techniques developed
                 for the simpler $k$-anonymization problem, causing
                 unnecessary information loss. (ii) The anonymization
                 process is inefficient in terms of computational and
                 I/O cost. (iii) Previous research focused exclusively
                 on the privacy-constrained problem and ignored the
                 equally important accuracy-constrained (or dual)
                 anonymization problem.\par

                 In this article, we propose a framework for efficient
                 anonymization of microdata that addresses these
                 deficiencies. First, we focus on one-dimensional (i.e.,
                 single-attribute) quasi-identifiers, and study the
                 properties of optimal solutions under the $k$-anonymity
                 and $l$-diversity models for the privacy-constrained
                 (i.e., direct) and the accuracy-constrained (i.e.,
                 dual) anonymization problems. Guided by these
                 properties, we develop efficient heuristics to solve
                 the one-dimensional problems in linear time. Finally,
                 we generalize our solutions to multidimensional
                 quasi-identifiers using space-mapping techniques.
                 Extensive experimental evaluation shows that our
                 techniques clearly outperform the existing approaches
                 in terms of execution time and information loss.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "anonymity; Privacy",
}

@Article{Hartmann:2009:ERA,
  author =       "Sven Hartmann and Sebastian Link",
  title =        "Efficient reasoning about a robust {XML} key
                 fragment",
  journal =      j-TODS,
  volume =       "34",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1538909.1538912",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:23:25 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We review key constraints in the context of XML as
                 introduced by Buneman et al. We demonstrate
                 that:\par

                 (1) one of the proposed inference rules is not sound in
                 general, and\par

                 (2) the inference rules are incomplete for XML key
                 implication, even for nonempty sets of simple key
                 paths.\par

                 This shows, in contrast to earlier statements, that the
                 axiomatizability of XML keys is still open, and
                 efficient algorithms for deciding their implication
                 still need to be developed. Solutions to these problems
                 have a wide range of applications including consistency
                 validation, XML schema design, data exchange and
                 integration, consistent query answering, XML query
                 optimization and rewriting, and indexing.\par

                 In this article, we investigate the axiomatizability
                 and implication problem for XML keys with nonempty sets
                 of simple key paths. In particular, we propose a set of
                 inference rules that is indeed sound and complete for
                 the implication of such XML keys. We demonstrate that
                 this fragment is robust by showing the duality of XML
                 key implication to the reachability problem of fixed
                 nodes in a suitable digraph. This enables us to develop
                 a quadratic-time algorithm for deciding implication,
                 and shows that reasoning about this XML key fragment is
                 practically efficient. Therefore, XML applications can
                 be unlocked effectively since they benefit not only
                 from those XML keys specified explicitly by the data
                 designer but also from those that are specified
                 implicitly.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "axiomatization; implication; reachability; XML data;
                 XML key",
}

@Article{Lin:2009:SII,
  author =       "Yi Lin and Bettina Kemme and Ricardo Jim{\'e}nez-Peris
                 and Marta Pati{\~n}o-Mart{\'\i}nez and Jos{\'e} Enrique
                 Armend{\'a}riz-I{\~n}igo",
  title =        "Snapshot isolation and integrity constraints in
                 replicated databases",
  journal =      j-TODS,
  volume =       "34",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1538909.1538913",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:23:25 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Database replication is widely used for fault
                 tolerance and performance. However, it requires replica
                 control to keep data copies consistent despite updates.
                 The traditional correctness criterion for the
                 concurrent execution of transactions in a replicated
                 database is 1-copy-serializability. It is based on
                 serializability, the strongest isolation level in a
                 nonreplicated system. In recent years, however,
                 Snapshot Isolation (SI), a slightly weaker isolation
                 level, has become popular in commercial database
                 systems. There exist already several replica control
                 protocols that provide SI in a replicated system.
                 However, most of the correctness reasoning for these
                 protocols has been rather informal. Additionally, most
                 of the work so far ignores the issue of integrity
                 constraints. In this article, we provide a formal
                 definition of 1-copy-SI using and extending a
                 well-established definition of SI in a nonreplicated
                 system. Our definition considers integrity constraints
                 in a way that conforms to the way integrity constraints
                 are handled in commercial systems. We discuss a set of
                 necessary and sufficient conditions for a replicated
                 history to be producible under 1-copy-SI. This makes
                 our formalism a convenient tool to prove the
                 correctness of replica control algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "integrity constraints; Replication; snapshot
                 isolation",
}

@Article{Su:2009:OOA,
  author =       "Weifeng Su and Jiying Wang and Frederick H.
                 Lochovsky",
  title =        "{ODE}: Ontology-assisted data extraction",
  journal =      j-TODS,
  volume =       "34",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1538909.1538914",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jul 2 12:23:25 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Online databases respond to a user query with result
                 records encoded in HTML files. Data extraction, which
                 is important for many applications, extracts the
                 records from the HTML files automatically. We present a
                 novel data extraction method, ODE (Ontology-assisted
                 Data Extraction), which automatically extracts the
                 query result records from the HTML pages. ODE first
                 constructs an ontology for a domain according to
                 information matching between the query interfaces and
                 query result pages from different Web sites within the
                 same domain. Then, the constructed domain ontology is
                 used during data extraction to identify the query
                 result section in a query result page and to align and
                 label the data values in the extracted records. The
                 ontology-assisted data extraction method is fully
                 automatic and overcomes many of the deficiencies of
                 current automatic data extraction methods. Experimental
                 results show that ODE is extremely accurate for
                 identifying the query result section in an HTML page,
                 segmenting the query result section into query result
                 records, and aligning and labeling the data values in
                 the query result records.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data value alignment; Domain ontology; label
                 assignment",
}

@Article{Agarwal:2009:ISS,
  author =       "Pankaj K. Agarwal and Junyi Xie and Jun Yang and Hai
                 Yu",
  title =        "Input-sensitive scalable continuous join query
                 processing",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "13:1--13:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567275",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article considers the problem of scalably
                 processing a large number of continuous queries. Our
                 approach, consisting of novel data structures and
                 algorithms and a flexible processing framework,
                 advances the state-of-the-art in several ways. First,
                 our approach is query sensitive in the sense that it
                 exploits potential overlaps in query predicates for
                 efficient group processing. We partition the collection
                 of continuous queries into groups based on the
                 clustering patterns of the query predicates, and apply
                 specialized processing strategies to heavily clustered
                 groups (or {\em hotspots\/}). We show how to maintain
                 the hotspots efficiently, and use them to scalably
                 process continuous select-join, band-join, and
                 window-join queries. Second, our approach is also data
                 sensitive, in the sense that it makes cost-based
                 decisions on how to process each incoming tuple based
                 on its characteristics. Experiments demonstrate that
                 our approach can improve the processing throughput by
                 orders of magnitude.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Continuous queries; data streams; event matching;
                 publish/subscribe",
}

@Article{Sharifzadeh:2009:PSS,
  author =       "Mehdi Sharifzadeh and Cyrus Shahabi and Leyla Kazemi",
  title =        "Processing spatial skyline queries in both vector
                 spaces and spatial network databases",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "14:1--14:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567276",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we first introduce the concept of
                 Spatial Skyline Queries (SSQ). Given a set of data
                 points $P$ and a set of query points $Q$, each data
                 point has a number of {\em derived spatial\/}
                 attributes each of which is the point's distance to a
                 query point. An SSQ retrieves those points of $P$ which
                 are not dominated by any other point in $P$ considering
                 their derived spatial attributes. The main difference
                 with the regular skyline query is that this {\em
                 spatial domination\/} depends on the location of the
                 query points $Q$. SSQ has application in several
                 domains such as emergency response and online maps. The
                 main intuition and novelty behind our approaches is
                 that we exploit the geometric properties of the SSQ
                 problem space to avoid the exhaustive examination of
                 all the point pairs in $P$ and $Q$. Consequently, we
                 reduce the complexity of SSQ search from $ O(|P|^2 |Q|)
                 $ to $ O(|S|^2 |C| + \sqrt {|P|}) $, where $ |S| $ and
                 $ |C| $ are the solution size and the number of
                 vertices of the convex hull of $Q$,
                 respectively.\par

                 Considering Euclidean distance, we propose two
                 algorithms, $ B^2 S^2 $ and VS$^2$, for static query
                 points and one algorithm, VCS$^2$, for streaming $Q$
                 whose points change location over time (e.g., are
                 mobile). VCS$^2$ exploits the pattern of change in $Q$
                 to avoid unnecessary recomputation of the skyline and
                 hence efficiently perform updates. We also propose two
                 algorithms, SNS$^2$ and VSNS$^2$, that compute the
                 spatial skyline with respect to the network distance in
                 a spatial network database. Our extensive experiments
                 using real-world datasets verify that both R-tree-based
                 $ B^2 S^2 $ and Voronoi-based VS$^2$ outperform the
                 best competitor approach in terms of both processing
                 time and I/O cost. Furthermore, their output computed
                 based on Euclidean distance is a good approximation of
                 the spatial skyline in network space. For accurate
                 computation of spatial skylines in network space, our
                 experiments showed the superiority of VSNS$^2$ over
                 SNS$^2$.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "spatial databases; Spatial skyline; Voronoi diagrams",
}

@Article{Yi:2009:SSG,
  author =       "Ke Yi and Feifei Li and Graham Cormode and Marios
                 Hadjieleftheriou and George Kollios and Divesh
                 Srivastava",
  title =        "Small synopses for group-by query verification on
                 outsourced data streams",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567277",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Due to the overwhelming flow of information in many
                 data stream applications, data outsourcing is a natural
                 and effective paradigm for individual businesses to
                 address the issue of scale. In the standard data
                 outsourcing model, the data owner outsources streaming
                 data to one or more third-party servers, which answer
                 queries posed by a potentially large number of clients
                 on the data owner's behalf. Data outsourcing
                 intrinsically raises issues of trust, making outsourced
                 query assurance on data streams a problem with
                 important practical implications. Existing solutions
                 proposed in this model all build upon cryptographic
                 primitives such as signatures and collision-resistant
                 hash functions, which only work for certain types of
                 queries, for example, simple selection/aggregation
                 queries.\par

                 In this article, we consider another common type of
                 queries, namely, ``GROUP BY, SUM'' queries, which
                 previous techniques fail to support. Our new solutions
                 are not based on cryptographic primitives, but instead
                 use algebraic and probabilistic techniques to compute a
                 small synopsis on the true query result, which is then
                 communicated to the client so as to verify the
                 correctness of the query result returned by the server.
                 The synopsis uses a constant amount of space
                 irrespective of the result size, has an extremely small
                 probability of failure, and can be maintained using no
                 extra space when the query result changes as elements
                 stream by. We then generalize our synopsis to allow
                 some tolerance on the number of erroneous groups, in
                 order to support semantic load shedding on the server.
                 When the number of erroneous groups is indeed
                 tolerable, the synopsis can be strengthened so that we
                 can locate and even correct these errors. Finally, we
                 implement our techniques and perform an empirical
                 evaluation using live network traffic.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data streams; outsourcing; Synopses",
}

@Article{Perez:2009:SCS,
  author =       "Jorge P{\'e}rez and Marcelo Arenas and Claudio
                 Gutierrez",
  title =        "Semantics and complexity of {SPARQL}",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567278",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "SPARQL is the standard language for querying RDF data.
                 In this article, we address systematically the formal
                 study of the database aspects of SPARQL, concentrating
                 in its graph pattern matching facility. We provide a
                 compositional semantics for the core part of SPARQL,
                 and study the complexity of the evaluation of several
                 fragments of the language. Among other complexity
                 results, we show that the evaluation of general SPARQL
                 patterns is PSPACE-complete. We identify a large class
                 of SPARQL patterns, defined by imposing a simple and
                 natural syntactic restriction, where the query
                 evaluation problem can be solved more efficiently. This
                 restriction gives rise to the class of well-designed
                 patterns. We show that the evaluation problem is
                 coNP-complete for well-designed patterns. Moreover, we
                 provide several rewriting rules for well-designed
                 patterns whose application may have a considerable
                 impact in the cost of evaluating SPARQL queries.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Complexity; query language; RDF; semantic Web;
                 SPARQL",
}

@Article{Markowetz:2009:KSR,
  author =       "Alexander Markowetz and Yin Yang and Dimitris
                 Papadias",
  title =        "Keyword search over relational tables and streams",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567279",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "{\em Relational Keyword Search\/} (R-KWS) provides an
                 intuitive way to query relational data without
                 requiring SQL, or knowledge of the underlying schema.
                 In this article we describe a comprehensive framework
                 for R-KWS covering snapshot queries on conventional
                 tables and continuous queries on relational streams.
                 Our contributions are summarized as follows: (i) We
                 provide formal semantics, addressing the temporal
                 validity and order of results, spanning uniformly over
                 tables and streams; (ii) we investigate two general
                 methodologies for query processing, {\em graph based\/}
                 and {\em operator based}, that resolve several problems
                 of previous approaches; and (iii) we develop a range of
                 algorithms and optimizations covering both
                 methodologies. We demonstrate the effectiveness of
                 R-KWS, as well as the significant performance benefits
                 of the proposed techniques, through extensive
                 experiments with static and streaming datasets.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "data graph; data streams; query processing; relational
                 databases; Search",
}

@Article{Cohen:2009:ICP,
  author =       "Sara Cohen and Benny Kimelfeld and Yehoshua Sagiv",
  title =        "Incorporating constraints in probabilistic {XML}",
  journal =      j-TODS,
  volume =       "34",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2009",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1567274.1567280",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Aug 31 16:11:01 MDT 2009",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Constraints are important, not only for maintaining
                 data integrity, but also because they capture natural
                 probabilistic dependencies among data items. A {\em
                 probabilistic XML database\/} (PXDB) is the probability
                 subspace comprising the instances of a {\em
                 p-document\/} that satisfy a set of constraints. In
                 contrast to existing models that can express
                 probabilistic dependencies, it is shown that query
                 evaluation is tractable in PXDBs. The problems of
                 sampling and determining well-definedness (i.e.,
                 whether the aforesaid subspace is nonempty) are also
                 tractable. Furthermore, queries and constraints can
                 include the aggregate functions {\em count, max,
                 min,\/} and {\em ratio.\/} Finally, this approach can
                 be easily extended to allow a probabilistic
                 interpretation of constraints.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "constraints; Probabilistic databases; probabilistic
                 XML; sampling probabilistic data",
}

@Article{Sasha:2009:FTS,
  author =       "Dennis Sasha and Maurizio Lenzerini and Z. Meral
                 {\"O}zsoyo{\u{g}}lu",
  title =        "Foreword to {TODS SIGMOD\slash PODS 2008} special
                 issue",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "19:1--19:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cahill:2009:SIS,
  author =       "Michael J. Cahill and Uwe R{\"o}hm and Alan D.
                 Fekete",
  title =        "Serializable isolation for snapshot databases",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "20:1--20:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{He:2009:RQC,
  author =       "Bingsheng He and Mian Lu and Ke Yang and Rui Fang and
                 Naga K. Govindaraju and Qiong Luo and Pedro V. Sander",
  title =        "Relational query coprocessing on graphics processors",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "21:1--21:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2009:RSM,
  author =       "Marcelo Arenas and Jorge P{\'e}rez and Cristian
                 Riveros",
  title =        "The recovery of a schema mapping: {Bringing} exchanged
                 data back",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "22:1--22:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Abiteboul:2009:SAA,
  author =       "Serge Abiteboul and Luc Segoufin and Victor Vianu",
  title =        "Static analysis of active {XML} systems",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "23:1--23:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chow:2009:CQP,
  author =       "Chi-Yin Chow and Mohamed F. Mokbel and Walid G. Aref",
  title =        "{Casper*}: {Query} processing for location services
                 without compromising privacy",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "24:1--24:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Benedikt:2009:XRL,
  author =       "Michael Benedikt and Christoph Koch",
  title =        "From {XQuery} to relational logics",
  journal =      j-TODS,
  volume =       "34",
  number =       "4",
  pages =        "25:1--25:??",
  month =        dec,
  year =         "2009",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:50 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ghanem:2010:SVD,
  author =       "Thanaa M. Ghanem and Ahmed K. Elmagarmid and
                 Per-{\AA}ke Larson and Walid G. Aref",
  title =        "Supporting views in data stream management systems",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "1:1--1:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wu:2010:AML,
  author =       "Kesheng Wu and Arie Shoshani and Kurt Stockinger",
  title =        "Analyses of multi-level and multi-component compressed
                 bitmap indexes",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "2:1--2:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lian:2010:RSS,
  author =       "Xiang Lian and Lei Chen",
  title =        "Reverse skyline search in uncertain databases",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "3:1--3:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Augsten:2010:GDB,
  author =       "Nikolaus Augsten and Michael B{\"o}hlen and Johann
                 Gamper",
  title =        "The $ p q $-gram distance between ordered labeled
                 trees",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "4:1--4:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kolahi:2010:ITA,
  author =       "Solmaz Kolahi and Leonid Libkin",
  title =        "An information-theoretic analysis of worst-case
                 redundancy in database design",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "5:1--5:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schnaitter:2010:OAE,
  author =       "Karl Schnaitter and Neoklis Polyzotis",
  title =        "Optimal algorithms for evaluating rank joins in
                 database systems",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "6:1--6:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Soror:2010:AVM,
  author =       "Ahmed A. Soror and Umar Farooq Minhas and Ashraf
                 Aboulnaga and Kenneth Salem and Peter Kokosielis and
                 Sunil Kamath",
  title =        "Automatic virtual machine configuration for database
                 workloads",
  journal =      j-TODS,
  volume =       "35",
  number =       "1",
  pages =        "7:1--7:??",
  month =        feb,
  year =         "2010",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Mar 15 12:22:52 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Xiao:2010:TAT,
  author =       "Xiaokui Xiao and Yufei Tao and Nick Koudas",
  title =        "Transparent anonymization: {Thwarting} adversaries who
                 know the algorithm",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "8:1--8:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735887",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Numerous generalization techniques have been proposed
                 for privacy-preserving data publishing. Most existing
                 techniques, however, implicitly assume that the
                 adversary knows little about the anonymization
                 algorithm adopted by the data publisher. Consequently,
                 they cannot guard against privacy attacks that exploit
                 various characteristics of the anonymization mechanism.
                 This article provides a practical solution to this
                 problem. First, we propose an analytical model for
                 evaluating disclosure risks, when an adversary knows
                 {\em everything\/} in the anonymization process, except
                 the sensitive values. Based on this model, we develop a
                 privacy principle, {\em transparent l-diversity}, which
                 ensures privacy protection against such powerful
                 adversaries. We identify three algorithms that achieve
                 transparent $l$-diversity, and verify their
                 effectiveness and efficiency through extensive
                 experiments with real data.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "generalization; l -diversity; Privacy-preserving data
                 publishing",
}

@Article{U:2010:OMB,
  author =       "Leong Hou U. and Kyriakos Mouratidis and Man Lung Yiu
                 and Nikos Mamoulis",
  title =        "Optimal matching between spatial datasets under
                 capacity constraints",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "9:1--9:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735888",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Consider a set of {\em customers\/} (e.g., WiFi
                 receivers) and a set of {\em service providers\/}
                 (e.g., wireless access points), where each provider has
                 a {\em capacity\/} and the quality of service offered
                 to its customers is anti-proportional to their
                 distance. The {\em Capacity Constrained Assignment\/}
                 (CCA) is a matching between the two sets such that (i)
                 each customer is assigned to at most one provider, (ii)
                 every provider serves no more customers than its
                 capacity, (iii) the maximum possible number of
                 customers are served, and (iv) the sum of Euclidean
                 distances within the assigned provider-customer pairs
                 is minimized. Although max-flow algorithms are
                 applicable to this problem, they require the complete
                 distance-based bipartite graph between the customer and
                 provider sets. For large spatial datasets, this graph
                 is expensive to compute and it may be too large to fit
                 in main memory. Motivated by this fact, we propose
                 efficient algorithms for {\em optimal assignment\/}
                 that employ novel edge-pruning strategies, based on the
                 spatial properties of the problem. Additionally, we
                 develop incremental techniques that maintain an optimal
                 assignment (in the presence of updates) with a
                 processing cost several times lower than CCA
                 recomputation from scratch. Finally, we present {\em
                 approximate\/} (i.e., suboptimal) CCA solutions that
                 provide a tunable trade-off between result accuracy and
                 computation cost, abiding by theoretical quality
                 guarantees. A thorough experimental evaluation
                 demonstrates the efficiency and practicality of the
                 proposed techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Optimal assignment; spatial databases",
}

@Article{Liu:2010:RSI,
  author =       "Ziyang Liu and Yi Chen",
  title =        "Return specification inference and result clustering
                 for keyword search on {XML}",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "10:1--10:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735889",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Keyword search enables Web users to easily access XML
                 data without the need to learn a structured query
                 language and to study possibly complex data schemas.
                 Existing work has addressed the problem of selecting
                 qualified data nodes that match keywords and connecting
                 them in a meaningful way, in the spirit of inferring
                 the {\em where clause\/} in XQuery. However, how to
                 infer the {\em return clause\/} for keyword searches is
                 an open problem.\par

                 To address this challenge, we present a keyword search
                 engine for data-centric XML, XSeek, to infer the
                 semantics of the search and identify return nodes
                 effectively. XSeek recognizes possible entities and
                 attributes inherently represented in the data. It also
                 distinguishes between predicates and return
                 specifications in query keywords. Then based on the
                 analysis of both XML data structures and keyword
                 patterns, XSeek generates return nodes. Furthermore,
                 when the query is ambiguous and it is hard or
                 impossible to determine the desirable return
                 information, XSeek clusters the query results according
                 to their semantics based on the user-specified
                 granularity, and enables the user to easily browse and
                 select the desired ones. Extensive experimental studies
                 show the effectiveness and efficiency of XSeek.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "keyword search; result clustering; XML",
}

@Article{Bex:2010:ICR,
  author =       "Geert Jan Bex and Frank Neven and Thomas Schwentick
                 and Stijn Vansummeren",
  title =        "Inference of concise regular expressions and {DTDs}",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "11:1--11:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735890",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider the problem of inferring a concise
                 Document Type Definition (DTD) for a given set of
                 XML-documents, a problem that basically reduces to
                 learning {\em concise\/} regular expressions from
                 positive examples strings. We identify two classes of
                 concise regular expressions --- the single occurrence
                 regular expressions (SOREs) and the chain regular
                 expressions (CHAREs) --- that capture the far majority
                 of expressions used in practical DTDs. For the
                 inference of SOREs we present several algorithms that
                 first infer an automaton for a given set of example
                 strings and then translate that automaton to a
                 corresponding SORE, possibly repairing the automaton
                 when no equivalent SORE can be found. In the process,
                 we introduce a novel automaton to regular expression
                 rewrite technique which is of independent interest.
                 When only a very small amount of XML data is available,
                 however (for instance when the data is generated by Web
                 service requests or by answers to queries), these
                 algorithms produce regular expressions that are too
                 specific. Therefore, we introduce a novel learning
                 algorithm crx that directly infers CHAREs (which form a
                 subclass of SOREs) without going through an automaton
                 representation. We show that crx performs very well
                 within its target class on very small datasets.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Regular expressions; schema inference; XML",
}

@Article{DeCapitaniDiVimercati:2010:EPR,
  author =       "Sabrina {De Capitani Di Vimercati} and Sara Foresti
                 and Sushil Jajodia and Stefano Paraboschi and
                 Pierangela Samarati",
  title =        "Encryption policies for regulating access to
                 outsourced data",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "12:1--12:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735891",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Current access control models typically assume that
                 resources are under the strict custody of a trusted
                 party which monitors each access request to verify if
                 it is compliant with the specified access control
                 policy. There are many scenarios where this approach is
                 becoming no longer adequate. Many clear trends in Web
                 technology are creating a need for owners of sensitive
                 information to manage access to it by legitimate users
                 using the services of {\em honest but curious\/} third
                 parties, that is, parties trusted with providing the
                 required service but not authorized to read the actual
                 data content. In this scenario, the data owner encrypts
                 the data before outsourcing and stores them at the
                 server. Only the data owner and users with knowledge of
                 the key will be able to decrypt the data. Possible
                 access authorizations are to be enforced by the owner.
                 In this article, we address the problem of enforcing
                 selective access on outsourced data without need of
                 involving the owner in the access control process. The
                 solution puts forward a novel approach that combines
                 cryptography with authorizations, thus enforcing access
                 control via {\em selective encryption}. The article
                 presents a formal model for access control management
                 and illustrates how an authorization policy can be
                 translated into an equivalent encryption policy while
                 minimizing the amount of keys and cryptographic tokens
                 to be managed. The article also introduces a two-layer
                 encryption approach that allows the data owner to
                 outsource, besides the data, the complete management of
                 the authorization policy itself, thus providing
                 efficiency and scalability in dealing with policy
                 updates. We also discuss experimental results showing
                 that our approach is able to efficiently manage complex
                 scenarios.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data outsourcing; encryption policy; privacy",
}

@Article{Koutrika:2010:PQB,
  author =       "Georgia Koutrika and Yannis Ioannidis",
  title =        "Personalizing queries based on networks of composite
                 preferences",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "13:1--13:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735892",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "People's preferences are expressed at varying levels
                 of granularity and detail as a result of partial or
                 imperfect knowledge. One may have some preference for a
                 general class of entities, for example, liking
                 comedies, and another one for a fine-grained, specific
                 class, such as disliking recent thrillers with Al
                 Pacino. In this article, we are interested in capturing
                 such complex, multi-granular preferences for
                 personalizing database queries and in studying their
                 impact on query results. We organize the collection of
                 one's preferences in a {\em preference network\/} (a
                 directed acyclic graph), where each node refers to a
                 subclass of the entities that its parent refers to, and
                 whenever they both apply, more specific preferences
                 override more generic ones. We study query
                 personalization based on networks of preferences and
                 provide efficient algorithms for identifying relevant
                 preferences, modifying queries accordingly, and
                 processing personalized queries. Finally, we present
                 results of both synthetic and real-user experiments,
                 which: (a) demonstrate the efficiency of our
                 algorithms, (b) provide insight as to the
                 appropriateness of the proposed preference model, and
                 (c) show the benefits of query personalization based on
                 composite preferences compared to simpler preference
                 representations.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "personalization; Preference modeling; preference
                 networks",
}

@Article{Flesca:2010:QRI,
  author =       "Sergio Flesca and Filippo Furfaro and Francesco
                 Parisi",
  title =        "Querying and repairing inconsistent numerical
                 databases",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "14:1--14:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735893",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The problem of extracting consistent information from
                 relational databases violating integrity constraints on
                 numerical data is addressed. In particular, aggregate
                 constraints defined as linear inequalities on
                 aggregate-sum queries on input data are considered. The
                 notion of repair as consistent set of updates at
                 attribute-value level is exploited, and the
                 characterization of several data-complexity issues
                 related to repairing data and computing consistent
                 query answers is provided. Moreover, a method for
                 computing ``reasonable'' repairs of inconsistent
                 numerical databases is provided, for a restricted but
                 expressive class of aggregate constraints. Several
                 experiments are presented which assess the
                 effectiveness of the proposed approach in real-life
                 application scenarios.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "aggregate constraints; consistent query answer;
                 Inconsistent databases; repairs",
}

@Article{Liu:2010:CIE,
  author =       "Hongyan Liu and Xiaoyu Wang and Yinghui Yang",
  title =        "Comments on {``An integrated efficient solution for
                 computing frequent and top-$k$ elements in data
                 streams''}",
  journal =      j-TODS,
  volume =       "35",
  number =       "2",
  pages =        "15:1--15:??",
  month =        apr,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1735886.1735894",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 28 13:44:08 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Metwally:2006:IES}.",
  abstract =     "We investigate a well-known algorithm, {\em
                 Space-Saving\/} [Metwally et al. 2006], which has been
                 proven efficient and effective at mining frequent
                 elements in data streams. We discovered an error in one
                 of the theorems in Metwally et al. [2006]. Experiments
                 are conducted to illustrate the error.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "approximate queries; data streams; frequent elements;
                 top-$k$ elements; Zipfian distributions",
}

@Article{Graefe:2010:SBT,
  author =       "Goetz Graefe",
  title =        "A survey of {B}-tree locking techniques",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "16:1--16:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806908",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "B-trees have been ubiquitous in database management
                 systems for several decades, and they are used in other
                 storage systems as well. Their basic structure and
                 basic operations are well and widely understood
                 including search, insertion, and deletion. Concurrency
                 control of operations in B-trees, however, is perceived
                 as a difficult subject with many subtleties and special
                 cases. The purpose of this survey is to clarify,
                 simplify, and structure the topic of concurrency
                 control in B-trees by dividing it into two subtopics
                 and exploring each of them in depth.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chen:2010:COI,
  author =       "Su Chen and Mario A. Nascimento and Beng Chin Ooi and
                 Kian-Lee Tan",
  title =        "Continuous online index tuning in moving object
                 databases",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "17:1--17:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806909",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In a {\em Moving Object Database\/} (MOD), the
                 dataset, for example, the location of objects and their
                 distribution, and the workload change frequently.
                 Traditional static indexes are not able to cope well
                 with such changes, that is, their effectiveness and
                 efficiency are seriously affected. This calls for the
                 development of novel indexes that can be reconfigured
                 automatically based on the state of the system. In this
                 article, we design and present the ST$^2$ B-tree, a
                 {\em S\/}elf-{\em T\/}unable {\em S\/}patio-{\em
                 T\/}emporal {\em B\/}$^+$ -tree index for MODs. In
                 ST$^2$ B-tree, the data space is partitioned into
                 regions of different density with respect to a set of
                 reference points. Based on the density, objects in a
                 region are managed using a grid of appropriate
                 granularity; intuitively, a dense region employs a grid
                 with fine granularity, while a sparse region uses a
                 grid with coarse granularity. In this way, the ST$^2$
                 B-tree adapts itself to workload diversity in space. To
                 enable online tuning, the ST$^2$ B-tree employs a
                 ``multitree'' indexing technique. The underlying
                 B$^+$-tree is logically divided into two subtrees.
                 Objects are dispatched to either subtree depending on
                 their last update time. The two subtrees are rebuilt
                 periodically and alternately. Whenever a subtree is
                 rebuilt, it is tuned to optimize performance by picking
                 an appropriate setting (e.g., the set of reference
                 points and grid granularity) based on the most recent
                 data and workload. To cut down the overhead of
                 rebuilding, we propose an eager update technique to
                 construct the subtree. Finally, we present a tuning
                 framework for the ST$^2$ B-tree, where the tuning is
                 conducted online and automatically without human
                 intervention, and without interfering with the regular
                 functions of the MOD. We have implemented the tuning
                 framework and the ST$^2$ B-tree, and conducted
                 extensive performance evaluations. The results show
                 that the self-tuning mechanism minimizes the
                 degradation of performance caused by workload changes
                 without any noticeable overhead.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Data distribution; index tuning; location-based
                 services; moving object indexing; self-tuning",
}

@Article{Hu:2010:PAL,
  author =       "Haibo Hu and Jianliang Xu and Sai Tung On and Jing Du
                 and Joseph Kee-Yin Ng",
  title =        "Privacy-aware location data publishing",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "18:1--18:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806910",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article examines a new problem of $k$-anonymity
                 with respect to a reference dataset in privacy-aware
                 location data publishing: given a user dataset and a
                 sensitive event dataset, we want to generalize the user
                 dataset such that by joining it with the event dataset
                 through location, each event is covered by at least $k$
                 users. Existing $k$ -anonymity algorithms generalize
                 every $k$ user locations to the same vague value,
                 regardless of the events. Therefore, they tend to
                 overprotect against the privacy compromise and make the
                 published data less useful. In this article, we propose
                 a new generalization paradigm called {\em local
                 enlargement}, as opposed to conventional hierarchy- or
                 partition-based generalization. Local enlargement
                 guarantees that user locations are enlarged just enough
                 to cover all events $k$ times, and thus maximize the
                 usefulness of the published data. We develop an $
                 O(H_n) $-approximate algorithm under the local
                 enlargement paradigm, where $n$ is the maximum number
                 of events a user could possibly cover and $ H_n $ is
                 the Harmonic number of $n$. With strong pruning
                 techniques and mathematical analysis, we show that it
                 runs efficiently and that the generalized user
                 locations are up to several orders of magnitude smaller
                 than those by the existing algorithms. In addition, it
                 is robust enough to protect against various privacy
                 attacks.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "k -anonymity; location privacy",
}

@Article{Liu:2010:IXS,
  author =       "Ziyang Liu and Yu Huang and Yi Chen",
  title =        "Improving {XML} search by generating and utilizing
                 informative result snippets",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "19:1--19:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806911",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Snippets are used by almost every text search engine
                 to complement the ranking scheme in order to
                 effectively handle user searches, which are inherently
                 ambiguous and whose relevance semantics are difficult
                 to assess. Despite the fact that XML is a standard
                 representation format of Web data, research on
                 generating result snippets for XML search remains
                 limited.\par

                 To tackle this important yet open problem, in this
                 article, we present a system extract which generates
                 snippets for XML search results. We identify that a
                 good XML result snippet should be a meaningful
                 information unit of a small size that effectively
                 summarizes this query result and differentiates it from
                 others, according to which users can quickly assess the
                 relevance of the query result. We have designed and
                 implemented a novel algorithm to satisfy these
                 requirements. Furthermore, we propose to cluster the
                 query results based on their snippets. Since XML result
                 clustering can only be done at query time,
                 snippet-based clustering significantly improves the
                 efficiency while compromising little clustering
                 accuracy. We verified the efficiency and effectiveness
                 of our approach through experiments.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "clustering; keyword search; snippets; XML",
}

@Article{Tao:2010:EAN,
  author =       "Yufei Tao and Ke Yi and Cheng Sheng and Panos Kalnis",
  title =        "Efficient and accurate nearest neighbor and closest
                 pair search in high-dimensional space",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "20:1--20:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806912",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Nearest Neighbor (NN) search in high-dimensional space
                 is an important problem in many applications. From the
                 database perspective, a good solution needs to have two
                 properties: (i) it can be easily incorporated in a
                 relational database, and (ii) its query cost should
                 increase {\em sublinearly\/} with the dataset size,
                 regardless of the data and query distributions. {\em
                 Locality-Sensitive Hashing\/} (LSH) is a well-known
                 methodology fulfilling both requirements, but its
                 current implementations either incur expensive space
                 and query cost, or abandon its theoretical guarantee on
                 the quality of query results.\par

                 Motivated by this, we improve LSH by proposing an
                 access method called the {\em Locality-Sensitive
                 B-tree\/} (LSB-tree) to enable fast, accurate,
                 high-dimensional NN search in relational databases. The
                 combination of several LSB-trees forms a {\em
                 LSB-forest\/} that has strong quality guarantees, but
                 improves dramatically the efficiency of the previous
                 LSH implementation having the same guarantees. In
                 practice, the LSB-tree itself is also an effective
                 index which consumes linear space, supports efficient
                 updates, and provides accurate query results. In our
                 experiments, the LSB-tree was faster than: (i)
                 iDistance (a famous technique for exact NN search) by
                 two orders of magnitude, and (ii) MedRank (a recent
                 approximate method with nontrivial quality guarantees)
                 by one order of magnitude, and meanwhile returned much
                 better results.\par

                 As a second step, we extend our LSB technique to solve
                 another classic problem, called Closest Pair (CP)
                 search, in high-dimensional space. The long-term
                 challenge for this problem has been to achieve {\em
                 subquadratic\/} running time at very high
                 dimensionalities, which fails most of the existing
                 solutions. We show that, using a LSB-forest, CP search
                 can be accomplished in (worst-case) time significantly
                 lower than the quadratic complexity, yet still ensuring
                 very good quality. In practice, accurate answers can be
                 found using just two LSB-trees, thus giving a
                 substantial reduction in the space and running time. In
                 our experiments, our technique was faster: (i) than
                 distance browsing (a well-known method for solving the
                 problem exactly) by several orders of magnitude, and
                 (ii) than D-shift (an approximate approach with
                 theoretical guarantees in low-dimensional space) by one
                 order of magnitude, and at the same time, outputs
                 better results.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "closest pair search; Locality-sensitive hashing;
                 nearest neighbor search",
}

@Article{Nash:2010:VQD,
  author =       "Alan Nash and Luc Segoufin and Victor Vianu",
  title =        "Views and queries: {Determinacy} and rewriting",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "21:1--21:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806913",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We investigate the question of whether a query $Q$ can
                 be answered using a set $V$ of views. We first define
                 the problem in information-theoretic terms: we say that
                 $V$ determines $Q$ if $V$ provides enough information
                 to uniquely determine the answer to $Q$. Next, we look
                 at the problem of rewriting $Q$ in terms of $V$ using a
                 specific language. Given a view language $V$ and query
                 language $Q$, we say that a rewriting language $R$ is
                 complete for $V$-to-$Q$ rewritings if every $ Q \in Q $
                 can be rewritten in terms of $ V \in V $ using a query
                 in $R$, whenever $V$ determines $Q$. While query
                 rewriting using views has been extensively investigated
                 for some specific languages, the connection to the
                 information-theoretic notion of determinacy, and the
                 question of completeness of a rewriting language have
                 received little attention. In this article we
                 investigate systematically the notion of determinacy
                 and its connection to rewriting. The results concern
                 decidability of determinacy for various view and query
                 languages, as well as the power required of complete
                 rewriting languages. We consider languages ranging from
                 first-order to conjunctive queries.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "Queries; rewriting; views",
}

@Article{Denecker:2010:TLR,
  author =       "Marc Denecker and {\'A}lvaro Cort{\'e}s-Calabuig and
                 Maurice Bruynooghes and Ofer Arieli",
  title =        "Towards a logical reconstruction of a theory for
                 locally closed databases",
  journal =      j-TODS,
  volume =       "35",
  number =       "3",
  pages =        "22:1--22:??",
  month =        jul,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1806907.1806914",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jul 28 15:53:01 MDT 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The {\em Closed World Assumption\/} (CWA) on databases
                 expresses the assumption that an atom not in the
                 database is false. This assumption is applicable only
                 in cases where the database has complete knowledge
                 about the domain of discourse. In this article, we
                 investigate {\em locally closed\/} databases, that is:
                 databases that are sound but partially incomplete about
                 their domain. Such databases consist of a standard
                 database instance, augmented with a collection of {\em
                 Local Closed World Assumptions\/} (LCWAs). A LCWA is a
                 ``local'' form of the CWA, expressing that a database
                 relation is complete in a certain area, called a {\em
                 window of expertise}. In this work, we study locally
                 closed databases both from a knowledge representation
                 and from a computational perspective. At the
                 representation level, the approach taken in this
                 article distinguishes between the data that is conveyed
                 by a database and the metaknowledge about the area in
                 which the data is complete. We study the semantics of
                 the LCWA's and relate it to several knowledge
                 representation formalisms. At the reasoning level, we
                 study the complexity of, and algorithms for two basic
                 reasoning tasks: computing {\em certain\/} and {\em
                 possible\/} answers to queries and determining whether
                 a database has complete knowledge on a query. As the
                 complexity of these tasks is unacceptably high, we
                 develop efficient {\em approximate\/} methods for query
                 answering. We also prove that for useful classes of
                 queries and locally closed databases, these methods are
                 {\em optimal}, and thus they solve the original query
                 in a tractable way. As a result, we obtain classes of
                 queries and locally closed databases for which query
                 answering is tractable.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
  keywords =     "closed world assumption; Databases; locally closed
                 databases",
}

@Article{Ozsoyoglu:2010:FTI,
  author =       "Z. Meral {\"O}zsoyoglu",
  title =        "Foreword to {TODS} invited papers issue",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "23:1--23:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862920",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ivanova:2010:ARI,
  author =       "Milena G. Ivanova and Martin L. Kersten and Niels J.
                 Nes and Romulo A. P. Gon{\c{c}}alves",
  title =        "An architecture for recycling intermediates in a
                 column-store",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "24:1--24:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862921",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Automatic recycling of intermediate results to improve
                 both query response time and throughput is a grand
                 challenge for state-of-the-art databases. Tuples are
                 loaded and streamed through a tuple-at-a-time
                 processing pipeline, avoiding materialization of
                 intermediates as much as possible. This limits the
                 opportunities for reuse of overlapping computations to
                 DBA-defined materialized views and function/result
                 cache tuning. In contrast, the operator-at-a-time
                 execution paradigm produces fully materialized results
                 in each step of the query plan. To avoid resource
                 contention, these intermediates are evicted as soon as
                 possible. In this article we study an architecture that
                 harvests the byproducts of the operator-at-a-time
                 paradigm in a column-store system using a lightweight
                 mechanism, the recycler.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ghoting:2010:EAS,
  author =       "Amol Ghoting and Konstantin Makarychev",
  title =        "{I/O} efficient algorithms for serial and parallel
                 suffix tree construction",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "25:1--25:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862922",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Over the past three decades, the suffix tree has
                 served as a fundamental data structure in string
                 processing. However, its widespread applicability has
                 been hindered due to the fact that suffix tree
                 construction does not scale well with the size of the
                 input string. With advances in data collection and
                 storage technologies, large strings have become
                 ubiquitous, especially across emerging applications
                 involving text, time series, and biological sequence
                 data. To benefit from these advances, it is imperative
                 that we have a scalable suffix tree construction
                 algorithm. The past few years have seen the emergence
                 of several disk-based suffix tree construction
                 algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Berinde:2010:SOH,
  author =       "Radu Berinde and Piotr Indyk and Graham Cormode and
                 Martin J. Strauss",
  title =        "Space-optimal heavy hitters with strong error bounds",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "26:1--26:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862923",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The problem of finding heavy hitters and approximating
                 the frequencies of items is at the heart of many
                 problems in data stream analysis. It has been observed
                 that several proposed solutions to this problem can
                 outperform their worst-case guarantees on real data.
                 This leads to the question of whether some stronger
                 bounds can be guaranteed. We answer this in the
                 positive by showing that a class of counter-based
                 algorithms (including the popular and very
                 space-efficient Frequent and SpacesSaving algorithms)
                 provides much stronger approximation guarantees than
                 previously known. Specifically, we show that errors in
                 the approximation of individual elements do not depend
                 on the frequencies of the most frequent elements, but
                 only on the frequency of the remaining tail.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2010:RIC,
  author =       "Wenfei Fan and Floris Geerts",
  title =        "Relative information completeness",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "27:1--27:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862924",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article investigates the question of whether a
                 partially closed database has complete information to
                 answer a query. In practice an enterprise often
                 maintains master data $ D m $, a closed-world database.
                 We say that a database $D$ is partially closed if it
                 satisfies a set $V$ of containment constraints of the
                 form $ q(D) \subset p(D m) $, where $q$ is a query in a
                 language $ L C $ and $p$ is a projection query. The
                 part of $D$ not constrained by $ (D m, V) $ is open,
                 from which some tuples may be missing. The database $D$
                 is said to be complete for a query $Q$ relative to $ (D
                 m, V) $ if for all partially closed extensions $ D' $
                 of $D$, $ Q(D') = Q(D) $, i.e., adding tuples to $D$
                 either violates some constraints in $V$ or does not
                 change the answer to $Q$.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fletcher:2010:TTS,
  author =       "George H. L. Fletcher and Jan {Van Den Bussche} and
                 Dirk {Van Gucht} and Stijn Vansummeren",
  title =        "Towards a theory of search queries",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "28:1--28:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862925",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The need to manage diverse information sources has
                 triggered the rise of very loosely structured data
                 models, known as dataspace models. Such information
                 management systems must allow querying in simple ways,
                 mostly by a form of searching. Motivated by these
                 developments, we propose a theory of search queries in
                 a general model of dataspaces. In this model, a
                 dataspace is a collection of data objects, where each
                 data object is a collection of data items. Basic search
                 queries are expressed using filters on data items,
                 following the basic model of Boolean search in
                 information retrieval. We characterize semantically the
                 class of queries that can be expressed by searching.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bjorklund:2010:IXE,
  author =       "Henrik Bj{\"o}rklund and Wouter Gelade and Wim
                 Martens",
  title =        "Incremental {XPath} evaluation",
  journal =      j-TODS,
  volume =       "35",
  number =       "4",
  pages =        "29:1--29:??",
  month =        nov,
  year =         "2010",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1862919.1862926",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Dec 15 10:34:39 MST 2010",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Incremental view maintenance for XPath queries asks to
                 maintain a materialized XPath view over an XML
                 database. It assumes an underlying XML database D and a
                 query Q. One is given a sequence of updates U to D, and
                 the problem is to compute the result of Q(U(D)): the
                 result of evaluating query Q on database D after having
                 applied updates U. This article initiates a systematic
                 study of the Boolean version of this problem. In the
                 Boolean version, one only wants to know whether Q(U(D))
                 is empty or not. In order to quickly answer this
                 question, we are allowed to maintain an auxiliary data
                 structure.",
  acknowledgement = ack-nhfb,
  articleno =    "29",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wang:2011:OPD,
  author =       "Ting Wang and Ling Liu",
  title =        "Output privacy in data mining",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929935",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Privacy has been identified as a vital requirement in
                 designing and implementing data mining systems. In
                 general, privacy preservation demands protecting both
                 input and output privacy: the former refers to
                 sanitizing the raw data itself before performing
                 mining; while the latter refers to preventing the
                 mining output (models or patterns) from malicious
                 inference attacks.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nergiz:2011:IA,
  author =       "Mehmet Ercan Nergiz and Acar Tamersoy and Yucel
                 Saygin",
  title =        "Instant anonymization",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929936",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Anonymization-based privacy protection ensures that
                 data cannot be traced back to individuals. Researchers
                 working in this area have proposed a wide variety of
                 anonymization algorithms, many of which require a
                 considerable number of database accesses. This is a
                 problem of efficiency, especially when the released
                 data is subject to visualization or when the algorithm
                 needs to be run many times to get an acceptable ratio
                 of privacy/utility. In this paper, we present two
                 instant anonymization algorithms for the privacy
                 metrics k-anonymity and $ \ell $-diversity. Proposed
                 algorithms minimize the number of data accesses by
                 utilizing the summary structure already maintained by
                 the database management system for query selectivity.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gross-Amblard:2011:QPW,
  author =       "David Gross-Amblard",
  title =        "Query-preserving watermarking of relational databases
                 and {XML} documents",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929937",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Watermarking allows robust and unobtrusive insertion
                 of information in a digital document. During the last
                 few years, techniques have been proposed for
                 watermarking relational databases or Xml documents,
                 where information insertion must preserve a specific
                 measure on data (for example the mean and variance of
                 numerical attributes). In this article we investigate
                 the problem of watermarking databases or Xml while
                 preserving a set of parametric queries in a specified
                 language, up to an acceptable distortion. We first show
                 that unrestricted databases can not be watermarked
                 while preserving trivial parametric queries. We then
                 exhibit query languages and classes of structures that
                 allow guaranteed watermarking capacity, namely (1)
                 local query languages on structures with bounded degree
                 Gaifman graph, and (2) monadic second-order queries on
                 trees or treelike structures.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Termehchy:2011:USI,
  author =       "Arash Termehchy and Marianne Winslett",
  title =        "Using structural information in {XML} keyword search
                 effectively",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "4:1--4:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929938",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The popularity of XML has exacerbated the need for an
                 easy-to-use, high precision query interface for XML
                 data. When traditional document-oriented keyword search
                 techniques do not suffice, natural language interfaces
                 and keyword search techniques that take advantage of
                 XML structure make it very easy for ordinary users to
                 query XML databases. Unfortunately, current approaches
                 to processing these queries rely heavily on heuristics
                 that are intuitively appealing but ultimately ad hoc.
                 These approaches often retrieve false positive answers,
                 overlook correct answers, and cannot rank answers
                 appropriately. To address these problems for
                 data-centric XML, we propose coherency ranking (CR), a
                 domain- and database design-independent ranking method
                 for XML keyword queries that is based on an extension
                 of the concepts of data dependencies and mutual
                 information.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cautis:2011:QXD,
  author =       "Bogdan Cautis and Alin Deutsch and Nicola Onose and
                 Vasilis Vassalos",
  title =        "Querying {XML} data sources that export very large
                 sets of views",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "5:1--5:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929939",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problem of querying XML data sources that
                 accept only a limited set of queries, such as sources
                 accessible by Web services which can implement very
                 large (potentially infinite) families of XPath queries.
                 To compactly specify such families of queries we adopt
                 the Query Set Specifications, a formalism close to
                 context-free grammars. We say that query Q is
                 expressible by the specification P if it is equivalent
                 to some expansion of P. Q is supported by P if it has
                 an equivalent rewriting using some finite set of P's
                 expansions. We study the complexity of expressibility
                 and support and identify large classes of XPath queries
                 for which there are efficient (PTIME) algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Liu:2011:GSW,
  author =       "Ziyang Liu and Susan B. Davidson and Yi Chen",
  title =        "Generating sound workflow views for correct provenance
                 analysis",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "6:1--6:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929940",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Workflow views abstract groups of tasks in a workflow
                 into high level composite tasks, in order to reuse
                 subworkflows and facilitate provenance analysis.
                 However, unless a view is carefully designed, it may
                 not preserve the dataflow between tasks in the
                 workflow, that is, it may not be sound. Unsound views
                 can be misleading and cause incorrect provenance
                 analysis. This article studies the problem of
                 efficiently identifying and correcting unsound workflow
                 views with minimal changes, and constructing minimal
                 sound and elucidative workflow views with a set of
                 user-specified relevant tasks. In particular, two
                 related problems are investigated. First, given a
                 workflow view, we wish to split each unsound composite
                 task into the minimal number of tasks, such that the
                 resulting view is sound.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jin:2011:PTE,
  author =       "Ruoming Jin and Ning Ruan and Yang Xiang and Haixun
                 Wang",
  title =        "Path-tree: {An} efficient reachability indexing scheme
                 for large directed graphs",
  journal =      j-TODS,
  volume =       "36",
  number =       "1",
  pages =        "7:1--7:??",
  month =        mar,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1929934.1929941",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 16 09:42:23 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Reachability query is one of the fundamental queries
                 in graph database. The main idea behind answering
                 reachability queries is to assign vertices with certain
                 labels such that the reachability between any two
                 vertices can be determined by the labeling information.
                 Though several approaches have been proposed for
                 building these reachability labels, it remains open
                 issues on how to handle increasingly large number of
                 vertices in real-world graphs, and how to find the best
                 tradeoff among the labeling size, the query answering
                 time, and the construction time. In this article, we
                 introduce a novel graph structure, referred to as
                 path-tree, to help labeling very large graphs.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Law:2011:RLD,
  author =       "Yan-Nei Law and Haixun Wang and Carlo Zaniolo",
  title =        "Relational languages and data models for continuous
                 queries on sequences and data streams",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "8:1--8:??",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966386",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Most data stream management systems are based on
                 extensions of the relational data model and query
                 languages, but rigorous analyses of the problems and
                 limitations of this approach, and how to overcome them,
                 are still wanting. In this article, we elucidate the
                 interaction between stream-oriented extensions of the
                 relational model and continuous query language
                 constructs, and show that the resulting expressive
                 power problems are even more serious for data streams
                 than for databases. In particular, we study the loss of
                 expressive power caused by the loss of blocking query
                 operators, and characterize nonblocking queries as
                 monotonic functions on the database. Thus we introduce
                 the notion of NB-completeness to assure that a query
                 language is as suitable for continuous queries as it is
                 for traditional database queries.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gao:2011:CNN,
  author =       "Yunjun Gao and Baihua Zheng and Gang Chen and Chun
                 Chen and Qing Li",
  title =        "Continuous nearest-neighbor search in the presence of
                 obstacles",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "9:1--9:??",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966387",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Despite the ubiquity of physical obstacles (e.g.,
                 buildings, hills, and blindages, etc.) in the real
                 world, most of spatial queries ignore the obstacles. In
                 this article, we study a novel form of continuous
                 nearest-neighbor queries in the presence of obstacles,
                 namely continuous obstructed nearest-neighbor (CONN)
                 search, which considers the impact of obstacles on the
                 distance between objects. Given a data set $P$, an
                 obstacle set $O$, and a query line segment $q$, in a
                 two-dimensional space, a CONN query retrieves the
                 nearest neighbor $ p \in P $ of each point $ p^\prime $
                 on $q$ according to the obstructed distance, the
                 shortest path between $p$ and $ p^\prime $ without
                 crossing any obstacle in $O$.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yiu:2011:DAR,
  author =       "Man Lung Yiu and Christian S. Jensen and Jesper
                 M{\o}ller and Hua Lu",
  title =        "Design and analysis of a ranking approach to private
                 location-based services",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "10:1--10:??",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966388",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Users of mobile services wish to retrieve nearby
                 points of interest without disclosing their locations
                 to the services. This article addresses the challenge
                 of optimizing the query performance while satisfying
                 given location privacy and query accuracy requirements.
                 The article's proposal, SpaceTwist, aims to offer
                 location privacy for k nearest neighbor (kNN) queries
                 at low communication cost without requiring a trusted
                 anonymizer. The solution can be used with a
                 conventional DBMS as well as with a server optimized
                 for location-based services. In particular, we believe
                 that this is the first solution that expresses the
                 server-side functionality in a single SQL statement. In
                 its basic form, SpaceTwist utilizes well-known
                 incremental NN query processing on the server.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fagin:2011:RDE,
  author =       "Ronald Fagin and Phokion G. Kolaitis and Lucian Popa
                 and Wang-Chiew Tan",
  title =        "Reverse data exchange: {Coping} with nulls",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "11:1--11:??",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966389",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "An inverse of a schema mapping M is intended to undo
                 what M does, thus providing a way to perform reverse
                 data exchange. In recent years, three different
                 formalizations of this concept have been introduced and
                 studied, namely the notions of an inverse of a schema
                 mapping, a quasi-inverse of a schema mapping, and a
                 maximum recovery of a schema mapping. The study of
                 these notions has been carried out in the context in
                 which source instances are restricted to consist
                 entirely of constants, while target instances may
                 contain both constants and labeled nulls. This
                 restriction on source instances is crucial for
                 obtaining some of the main technical results about
                 these three notions, but, at the same time, limits
                 their usefulness, since reverse data exchange naturally
                 leads to source instances that may contain both
                 constants and labeled nulls.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Atallah:2011:AEA,
  author =       "Mikhail J. Atallah and Yinian Qi and Hao Yuan",
  title =        "Asymptotically efficient algorithms for skyline
                 probabilities of uncertain data",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "12:1--12:28",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966390",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Skyline computation is widely used in multicriteria
                 decision making. As research in uncertain databases
                 draws increasing attention, skyline queries with
                 uncertain data have also been studied. Some earlier
                 work focused on probabilistic skylines with a given
                 threshold; Atallah and Qi [2009] studied the problem to
                 compute skyline probabilities for all instances of
                 uncertain objects without the use of thresholds, and
                 proposed an algorithm with subquadratic time
                 complexity. In this work, we propose a new algorithm
                 for computing all skyline probabilities that is
                 asymptotically faster: worst-case $ O(n \sqrt {n} \log
                 n) $ time and $ O(n) $ space for 2D data; $ O(n^{2 - 1
                 / d} \log^{d - 1} n) $ time and $ O(n \log^{d - 2} n) $
                 space for $d$-dimensional data. Furthermore, we study
                 the online version of the problem: Given any query
                 point $p$ (unknown until the query time), return the
                 probability that no instance in the given data set
                 dominates $p$. We propose an algorithm for answering
                 such an online query for $d$-dimensional data in $
                 O(n^{1 1 / d} \log^{d 1} n) $ time after preprocessing
                 the data in $ O(n^{2 1 / d} \log^{d 1}) $ time and
                 space.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lee:2011:RAF,
  author =       "Tony T. Lee and Tong Ye",
  title =        "A relational approach to functional decomposition of
                 logic circuits",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "13:1--13:??",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966391",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Functional decomposition of Boolean functions has a
                 profound influence on all quality aspects of
                 cost-effectively implementing modern digital systems
                 and data-mining. The relational databases are
                 multivalued tables, which include any truth tables of
                 logic functions as special cases. In this article, we
                 propose a relational database approach to the
                 decomposition of logic circuits. The relational algebra
                 consists of a set of well-defined algebraic operations
                 that can be performed on multivalued tables. Our
                 approach shows that the functional decomposition of
                 logic circuits is similar to the normalization of
                 relational databases; they are governed by the same
                 concepts of functional dependency (FD) and multivalued
                 dependency (MVD).",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hernich:2011:CWD,
  author =       "Andr{\'e} Hernich and Leonid Libkin and Nicole
                 Schweikardt",
  title =        "Closed world data exchange",
  journal =      j-TODS,
  volume =       "36",
  number =       "2",
  pages =        "14:1--14:40",
  month =        may,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/1966385.1966392",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 3 18:41:49 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data exchange deals with translating data structured
                 in some source format into data structured in some
                 target format, given a specification of the
                 relationship between the source and the target and
                 possibly constraints on the target; and answering
                 queries over the target in a way that is semantically
                 consistent with the information in the source.
                 Theoretical foundations of data exchange have been
                 actively explored recently. It was also noticed that
                 the standard semantics for query answering in data
                 exchange may lead to counterintuitive or anomalous
                 answers. In the present article, we explain that this
                 behavior is due to the fact that solutions can contain
                 invented information (information that is not related
                 to the source instance), and that the presence of
                 incomplete information in target instances has been
                 ignored.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Xiao:2011:ESJ,
  author =       "Chuan Xiao and Wei Wang and Xuemin Lin and Jeffrey Xu
                 Yu and Guoren Wang",
  title =        "Efficient similarity joins for near-duplicate
                 detection",
  journal =      j-TODS,
  volume =       "36",
  number =       "3",
  pages =        "15:1--15:41",
  month =        aug,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2000824.2000825",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Aug 23 18:27:45 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "With the increasing amount of data and the need to
                 integrate data from multiple data sources, one of the
                 challenging issues is to identify near-duplicate
                 records efficiently. In this article, we focus on
                 efficient algorithms to find a pair of records such
                 that their similarities are no less than a given
                 threshold. Several existing algorithms rely on the
                 prefix filtering principle to avoid computing
                 similarity values for all possible pairs of records. We
                 propose new filtering techniques by exploiting the
                 token ordering information; they are integrated into
                 the existing methods and drastically reduce the
                 candidate sizes and hence improve the efficiency.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Song:2011:DDR,
  author =       "Shaoxu Song and Lei Chen",
  title =        "Differential dependencies: Reasoning and discovery",
  journal =      j-TODS,
  volume =       "36",
  number =       "3",
  pages =        "16:1--16:41",
  month =        aug,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2000824.2000826",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Aug 23 18:27:45 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See comments \cite{Vincent:2015:TCD} and response
                 \cite{Song:2017:RDD}.",
  abstract =     "The importance of difference semantics (e.g.,
                 ``similar'' or ``dissimilar'') has been recently
                 recognized for declaring dependencies among various
                 types of data, such as numerical values or text values.
                 We propose a novel form of Differential Dependencies
                 (dds), which specifies constraints on difference,
                 called differential functions, instead of
                 identification functions in traditional dependency
                 notations like functional dependencies. Informally, a
                 differential dependency states that if two tuples have
                 distances on attributes X agreeing with a certain
                 differential function, then their distances on
                 attributes Y should also agree with the corresponding
                 differential function on Y. For example, [date($ \leq 7
                 $ )] $ \rightarrow $ [price($ < 100 $ )] states that
                 the price difference of any two days within a week
                 length should be no greater than 100 dollars.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papapetrou:2011:EBS,
  author =       "Panagiotis Papapetrou and Vassilis Athitsos and
                 Michalis Potamias and George Kollios and Dimitrios
                 Gunopulos",
  title =        "Embedding-based subsequence matching in time-series
                 databases",
  journal =      j-TODS,
  volume =       "36",
  number =       "3",
  pages =        "17:1--17:39",
  month =        aug,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2000824.2000827",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Aug 23 18:27:45 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We propose an embedding-based framework for
                 subsequence matching in time-series databases that
                 improves the efficiency of processing subsequence
                 matching queries under the Dynamic Time Warping (DTW)
                 distance measure. This framework partially reduces
                 subsequence matching to vector matching, using an
                 embedding that maps each query sequence to a vector and
                 each database time series into a sequence of vectors.
                 The database embedding is computed offline, as a
                 preprocessing step. At runtime, given a query object,
                 an embedding of that object is computed online.
                 Relatively few areas of interest are efficiently
                 identified in the database sequences by comparing the
                 embedding of the query with the database vectors.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jampani:2011:MCD,
  author =       "Ravi Jampani and Fei Xu and Mingxi Wu and Luis Perez
                 and Chris Jermaine and Peter J. Haas",
  title =        "The {Monte Carlo} database system: Stochastic analysis
                 close to the data",
  journal =      j-TODS,
  volume =       "36",
  number =       "3",
  pages =        "18:1--18:41",
  month =        aug,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2000824.2000828",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Aug 23 18:27:45 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The application of stochastic models and analysis
                 techniques to large datasets is now commonplace.
                 Unfortunately, in practice this usually means
                 extracting data from a database system into an external
                 tool (such as SAS, R, Arena, or Matlab), and then
                 running the analysis there. This extract-and-model
                 paradigm is typically error-prone, slow, does not
                 support fine-grained modeling, and discourages what-if
                 and sensitivity analyses. In this article we describe
                 MCDB, a database system that permits a wide spectrum of
                 stochastic models to be used in conjunction with the
                 data stored in a large database, without ever
                 extracting the data. MCDB facilitates in-database
                 execution of tasks such as risk assessment, prediction,
                 and imputation of missing data, as well as management
                 of errors due to data integration, information
                 extraction, and privacy-preserving data
                 anonymization.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Stefanidis:2011:SRC,
  author =       "Kostas Stefanidis and Georgia Koutrika and Evaggelia
                 Pitoura",
  title =        "A survey on representation, composition and
                 application of preferences in database systems",
  journal =      j-TODS,
  volume =       "36",
  number =       "3",
  pages =        "19:1--19:45",
  month =        aug,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2000824.2000829",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Aug 23 18:27:45 MDT 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Preferences have been traditionally studied in
                 philosophy, psychology, and economics and applied to
                 decision making problems. Recently, they have attracted
                 the attention of researchers in other fields, such as
                 databases where they capture soft criteria for queries.
                 Databases bring a whole fresh perspective to the study
                 of preferences, both computational and
                 representational. From a representational perspective,
                 the central question is how we can effectively
                 represent preferences and incorporate them in database
                 querying. From a computational perspective, we can look
                 at how we can efficiently process preferences in the
                 context of database queries. Several approaches have
                 been proposed but a systematic study of these works is
                 missing.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ozsoyoglu:2011:FTI,
  author =       "Z. Meral {\"O}zsoyoglu",
  title =        "Foreword to {TODS} invited papers issue 2011",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "20:1--20:2",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043653",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cheng:2011:FMC,
  author =       "James Cheng and Yiping Ke and Ada Wai-Chee Fu and
                 Jeffrey Xu Yu and Linhong Zhu",
  title =        "Finding maximal cliques in massive networks",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "21:1--21:34",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043654",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Maximal clique enumeration is a fundamental problem in
                 graph theory and has important applications in many
                 areas such as social network analysis and
                 bioinformatics. The problem is extensively studied;
                 however, the best existing algorithms require memory
                 space linear in the size of the input graph. This has
                 become a serious concern in view of the massive volume
                 of today's fast-growing networks. We propose a general
                 framework for designing external-memory algorithms for
                 maximal clique enumeration in large graphs. The general
                 framework enables maximal clique enumeration to be
                 processed recursively in small subgraphs of the input
                 graph, thus allowing in-memory computation of maximal
                 cliques without the costly random disk access.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kim:2011:DFA,
  author =       "Changkyu Kim and Jatin Chhugani and Nadathur Satish
                 and Eric Sedlar and Anthony D. Nguyen and Tim Kaldewey
                 and Victor W. Lee and Scott A. Brandt and Pradeep
                 Dubey",
  title =        "Designing fast architecture-sensitive tree search on
                 modern multicore\slash many-core processors",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "22:1--22:34",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043655",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In-memory tree structured index search is a
                 fundamental database operation. Modern processors
                 provide tremendous computing power by integrating
                 multiple cores, each with wide vector units. There has
                 been much work to exploit modern processor
                 architectures for database primitives like scan, sort,
                 join, and aggregation. However, unlike other
                 primitives, tree search presents significant challenges
                 due to irregular and unpredictable data accesses in
                 tree traversal. In this article, we present FAST, an
                 extremely fast architecture-sensitive layout of the
                 index tree. FAST is a binary tree logically organized
                 to optimize for architecture features like page size,
                 cache line size, and Single Instruction Multiple Data
                 (SIMD) width of the underlying hardware.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Alexe:2011:CSM,
  author =       "Bogdan Alexe and Balder ten Cate and Phokion G.
                 Kolaitis and Wang-Chiew Tan",
  title =        "Characterizing schema mappings via data examples",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "23:1--23:48",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043656",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Schema mappings are high-level specifications that
                 describe the relationship between two database schemas;
                 they are considered to be the essential building blocks
                 in data exchange and data integration, and have been
                 the object of extensive research investigations. Since
                 in real-life applications schema mappings can be quite
                 complex, it is important to develop methods and tools
                 for understanding, explaining, and refining schema
                 mappings. A promising approach to this effect is to use
                 ``good'' data examples that illustrate the schema
                 mapping at hand. We develop a foundation for the
                 systematic investigation of data examples and obtain a
                 number of results on both the capabilities and the
                 limitations of data examples in explaining and
                 understanding schema mappings.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cohen:2011:BET,
  author =       "Sara Cohen and Yaacov Y. Weiss",
  title =        "Bag equivalence of tree patterns",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "24:1--24:35",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043657",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "When a query is evaluated under bag semantics, each
                 answer is returned as many times as it has derivations.
                 Bag semantics has long been recognized as important,
                 especially when aggregation functions will be applied
                 to query results. This article is the first to focus on
                 bag semantics for tree pattern queries. In particular,
                 the problem of bag equivalence of a large class of tree
                 pattern queries (which can be used to model XPath) is
                 explored. The queries can contain unions, branching,
                 label wildcards, the vertical child and descendant
                 axes, the horizontal following and following-sibling
                 axes, as well as positional (i.e., first and last)
                 axes.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Abiteboul:2011:CCD,
  author =       "Serge Abiteboul and T.-H. Hubert Chan and Evgeny
                 Kharlamov and Werner Nutt and Pierre Senellart",
  title =        "Capturing continuous data and answering aggregate
                 queries in probabilistic {XML}",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "25:1--25:45",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043658",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Sources of data uncertainty and imprecision are
                 numerous. A way to handle this uncertainty is to
                 associate probabilistic annotations to data. Many such
                 probabilistic database models have been proposed, both
                 in the relational and in the semi-structured setting.
                 The latter is particularly well adapted to the
                 management of uncertain data coming from a variety of
                 automatic processes. An important problem, in the
                 context of probabilistic XML databases, is that of
                 answering aggregate queries (count, sum, avg, etc.),
                 which has received limited attention so far. In a model
                 unifying the various (discrete) semi-structured
                 probabilistic models studied up to now, we present
                 algorithms to compute the distribution of the
                 aggregation values (exploiting some regularity
                 properties of the aggregate functions) and
                 probabilistic moments (especially expectation and
                 variance) of this distribution.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bai:2011:CPT,
  author =       "Xiao Bai and Rachid Guerraoui and Anne-Marie Kermarrec
                 and Vincent Leroy",
  title =        "Collaborative personalized top-$k$ processing",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "26:1--26:38",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043659",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article presents P4Q, a fully decentralized
                 gossip-based protocol to personalize query processing
                 in social tagging systems. P4Q dynamically associates
                 each user with social acquaintances sharing similar
                 tagging behaviors. Queries are gossiped among such
                 acquaintances, computed on-the-fly in a collaborative,
                 yet partitioned manner, and results are iteratively
                 refined and returned to the querier. Analytical and
                 experimental evaluations convey the scalability of P4Q
                 for top-k query processing, as well its inherent
                 ability to cope with users updating profiles and
                 departing.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Goncalves:2011:DCQ,
  author =       "Romulo Goncalves and Martin Kersten",
  title =        "The {Data Cyclotron} query processing scheme",
  journal =      j-TODS,
  volume =       "36",
  number =       "4",
  pages =        "27:1--27:35",
  month =        dec,
  year =         "2011",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2043652.2043660",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Dec 20 07:23:02 MST 2011",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A grand challenge of distributed query processing is
                 to devise a self-organizing architecture which exploits
                 all hardware resources optimally to manage the database
                 hot set, minimize query response time, and maximize
                 throughput without single point global coordination.
                 The Data Cyclotron architecture [Goncalves and Kersten
                 2010] addresses this challenge using turbulent data
                 movement through a storage ring built from distributed
                 main memory and capitalizing on the functionality
                 offered by modern remote-DMA network facilities.
                 Queries assigned to individual nodes interact with the
                 storage ring by picking up data fragments, which are
                 continuously flowing around, that is, the hot set. The
                 storage ring is steered by the Level Of Interest (LOI)
                 attached to each data fragment, which represents the
                 cumulative query interest as it passes around the ring
                 multiple times.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Graefe:2012:SBT,
  author =       "Goetz Graefe",
  title =        "A survey of {B}-tree logging and recovery techniques",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "1:1--1:35",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109197",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "B-trees have been ubiquitous in database management
                 systems for several decades, and they serve in many
                 other storage systems as well. Their basic structure
                 and their basic operations are well understood
                 including search, insertion, and deletion. However,
                 implementation of transactional guarantees such as
                 all-or-nothing failure atomicity and durability in
                 spite of media and system failures seems to be
                 difficult. High-performance techniques such as
                 pseudo-deleted records, allocation-only logging, and
                 transaction processing during crash recovery are widely
                 used in commercial B-tree implementations but not
                 widely understood. This survey collects many of these
                 techniques as a reference for students, researchers,
                 system architects, and software developers. Central in
                 this discussion are physical data independence,
                 separation of logical database contents and physical
                 representation, and the concepts of user transactions
                 and system transactions. Many of the techniques
                 discussed are applicable beyond B-trees.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Martinenghi:2012:PMR,
  author =       "Davide Martinenghi and Marco Tagliasacchi",
  title =        "Proximity measures for rank join",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "2:1--2:46",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109198",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We introduce the proximity rank join problem, where we
                 are given a set of relations whose tuples are equipped
                 with a score and a real-valued feature vector. Given a
                 target feature vector, the goal is to return the K
                 combinations of tuples with high scores that are as
                 close as possible to the target and to each other,
                 according to some notion of distance or dissimilarity.
                 The setting closely resembles that of traditional rank
                 join, but the geometry of the vector space plays a
                 distinctive role in the computation of the overall
                 score of a combination. Also, the input relations
                 typically return their results either by distance from
                 the target or by score. Because of these aspects, it
                 turns out that traditional rank join algorithms, such
                 as the well-known HRJN, have shortcomings in solving
                 the proximity rank join problem, as they may read more
                 input than needed. To overcome this weakness, we define
                 a tight bound (used as a stopping criterion) that
                 guarantees instance optimality, that is, an I/O cost is
                 achieved that is always within a constant factor of
                 optimal. The tight bound can also be used to drive an
                 adaptive pulling strategy, deciding at each step which
                 relation to access next. For practically relevant
                 classes of problems, we show how to compute the tight
                 bound efficiently. An extensive experimental study
                 validates our results and demonstrates significant
                 gains over existing solutions.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nuray-Turan:2012:AOS,
  author =       "Rabia Nuray-Turan and Dmitri V. Kalashnikov and Sharad
                 Mehrotra and Yaming Yu",
  title =        "Attribute and object selection queries on objects with
                 probabilistic attributes",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "3:1--3:??",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109199",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Modern data processing techniques such as entity
                 resolution, data cleaning, information extraction, and
                 automated tagging often produce results consisting of
                 objects whose attributes may contain uncertainty. This
                 uncertainty is frequently captured in the form of a set
                 of multiple mutually exclusive value choices for each
                 uncertain attribute along with a measure of probability
                 for alternative values. However, the lay end-user, as
                 well as some end-applications, might not be able to
                 interpret the results if outputted in such a form.
                 Thus, the question is how to present such results to
                 the user in practice, for example, to support
                 attribute-value selection and object selection queries
                 the user might be interested in. Specifically, in this
                 article we study the problem of maximizing the quality
                 of these selection queries on top of such a
                 probabilistic representation. The quality is measured
                 using the standard and commonly used set-based quality
                 metrics. We formalize the problem and then develop
                 efficient approaches that provide high-quality answers
                 for these queries. The comprehensive empirical
                 evaluation over three different domains demonstrates
                 the advantage of our approach over existing
                 techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Liu:2012:DSR,
  author =       "Ziyang Liu and Yi Chen",
  title =        "Differentiating search results on structured data",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "4:1--4:??",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109200",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Studies show that about 50\% of Web search is for
                 information exploration purposes, where a user would
                 like to investigate, compare, evaluate, and synthesize
                 multiple relevant results. Due to the absence of
                 general tools that can effectively analyze and
                 differentiate multiple results, a user has to manually
                 read and comprehend potential large results in an
                 exploratory search. Such a process is time consuming,
                 labor intensive and error prone. Interestingly, we find
                 that the metadata information embedded in structured
                 data provides a potential for automating or
                 semi-automating the comparison of multiple results. In
                 this article we present an approach for structured data
                 search result differentiation. We define the
                 differentiability of query results and quantify the
                 degree of difference. Then we define the problem of
                 identifying a limited number of valid features in a
                 result that can maximally differentiate this result
                 from the others, which is proved NP-hard. We propose
                 two local optimality conditions, namely single-swap and
                 multi-swap, and design efficient algorithms to achieve
                 local optimality. We then present a feature type-based
                 approach, which further improves the quality of the
                 features identified for result differentiation. To show
                 the usefulness of our approach, we implemented a system
                 CompareIt, which can be used to compare structured
                 search results as well as any objects. Our empirical
                 evaluation verifies the effectiveness and efficiency of
                 the proposed approach.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yang:2012:SES,
  author =       "Di Yang and Elke A. Rundensteiner and Matthew O.
                 Ward",
  title =        "Shared execution strategy for neighbor-based pattern
                 mining requests over streaming windows",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "5:1--5:??",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109201",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In diverse applications ranging from stock trading to
                 traffic monitoring, data streams are continuously
                 monitored by multiple analysts for extracting patterns
                 of interest in real time. These analysts often submit
                 similar pattern mining requests yet customized with
                 different parameter settings. In this work, we present
                 shared execution strategies for processing a large
                 number of neighbor-based pattern mining requests of the
                 same type yet with arbitrary parameter settings. Such
                 neighbor-based pattern mining requests cover a broad
                 range of popular mining query types, including
                 detection of clusters, outliers, and nearest neighbors.
                 Given the high algorithmic complexity of the mining
                 process, serving multiple such queries in a single
                 system is extremely resource intensive. The naive
                 method of detecting and maintaining patterns for
                 different queries independently is often infeasible in
                 practice, as its demands on system resources increase
                 dramatically with the cardinality of the query
                 workload. In order to maximize the efficiency of the
                 system resource utilization for executing multiple
                 queries simultaneously, we analyze the commonalities of
                 the neighbor-based pattern mining queries, and identify
                 several general optimization principles which lead to
                 significant system resource sharing among multiple
                 queries. In particular, as a preliminary sharing
                 effort, we observe that the computation needed for the
                 range query searches (the process of searching the
                 neighbors for each object) can be shared among multiple
                 queries and thus saves the CPU consumption. Then we
                 analyze the interrelations between the patterns
                 identified by queries with different parameters
                 settings, including both pattern-specific and
                 window-specific parameters. For that, we first
                 introduce an incremental pattern representation, which
                 represents the patterns identified by queries with
                 different pattern-specific parameters within a single
                 compact structure. This enables integrated pattern
                 maintenance for multiple queries. Second, by leveraging
                 the potential overlaps among sliding windows, we
                 propose a metaquery strategy which utilizes a single
                 query to answer multiple queries with different
                 window-specific parameters. By combining these three
                 techniques, namely the range query search sharing,
                 integrated pattern maintenance, and metaquery strategy,
                 our framework realizes fully shared execution of
                 multiple queries with arbitrary parameter settings. It
                 achieves significant savings of computational and
                 memory resources due to shared execution. Our
                 comprehensive experimental study, using real data
                 streams from domains of stock trades and moving object
                 monitoring, demonstrates that our solution is
                 significantly faster than the independent execution
                 strategy, while using only a small portion of memory
                 space compared to the independent execution. We also
                 show that our solution scales in handling large numbers
                 of queries in the order of hundreds or even thousands
                 under high input data rates.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Re:2012:UCE,
  author =       "Christopher R{\'e} and D. Suciu",
  title =        "Understanding cardinality estimation using entropy
                 maximization",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "6:1--6:??",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109202",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Cardinality estimation is the problem of estimating
                 the number of tuples returned by a query; it is a
                 fundamentally important task in data management, used
                 in query optimization, progress estimation, and
                 resource provisioning. We study cardinality estimation
                 in a principled framework: given a set of statistical
                 assertions about the number of tuples returned by a
                 fixed set of queries, predict the number of tuples
                 returned by a new query. We model this problem using
                 the probability space, over possible worlds, that
                 satisfies all provided statistical assertions and
                 maximizes entropy. We call this the Entropy
                 Maximization model for statistics (MaxEnt). In this
                 article we develop the mathematical techniques needed
                 to use the MaxEnt model for predicting the cardinality
                 of conjunctive queries.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nuray-Turan:2012:EWQ,
  author =       "Rabia Nuray-Turan and Dmitri V. Kalashnikov and Sharad
                 Mehrotra",
  title =        "Exploiting {Web} querying for {Web} people search",
  journal =      j-TODS,
  volume =       "37",
  number =       "1",
  pages =        "7:1--7:??",
  month =        feb,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2109196.2109203",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:17 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Searching for people on the Web is one of the most
                 common query types submitted to Web search engines
                 today. However, when a person name is queried, the
                 returned Webpages often contain documents related to
                 several distinct namesakes who have the queried name.
                 The task of disambiguating and finding the Webpages
                 related to the specific person of interest is left to
                 the user. Many Web People Search (WePS) approaches have
                 been developed recently that attempt to automate this
                 disambiguation process. Nevertheless, the
                 disambiguation quality of these techniques leaves major
                 room for improvement. In this article, we present a new
                 WePS approach. It is based on issuing additional
                 auxiliary queries to the Web to gain additional
                 knowledge about the Webpages that need to be
                 disambiguated. Thus, the approach uses the Web as an
                 external data source by issuing queries to collect
                 co-occurrence statistics. These statistics are used to
                 assess the overlap of the contextual entities extracted
                 from the Webpages. The article also proposes a
                 methodology to make this Web querying technique
                 efficient. Further, the article proposes an approach
                 that is capable of combining various types of
                 disambiguating information, including other common
                 types of similarities, by applying a correlation
                 clustering approach with after-clustering of singleton
                 clusters. These properties allow the framework to get
                 an advantage in terms of result quality over other
                 state-of-the-art WePS techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schneider:2012:CDB,
  author =       "Markus Schneider and Tao Chen and Ganesh Viswanathan
                 and Wenjie Yuan",
  title =        "Cardinal directions between complex regions",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "8:1--8:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188350",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Besides topological relationships and approximate
                 relationships, cardinal directions like north and
                 southwest have turned out to be an important class of
                 qualitative spatial relationships. They are of
                 interdisciplinary interest in fields like cognitive
                 science, robotics, artificial intelligence, and
                 qualitative spatial reasoning. In spatial databases and
                 Geographic Information Systems (GIS) they are
                 frequently used as join and selection criteria in
                 spatial queries. However, the available computational
                 models of cardinal directions suffer a number of
                 problems like the use of too coarse approximations of
                 the two spatial operand objects in terms of single
                 representative points or minimum bounding rectangles,
                 the lacking property of converseness of the cardinal
                 directions computed, and the limited applicability to
                 simple instead of complex regions only. This article
                 proposes and formally defines a novel two-phase model,
                 called the Objects Interaction Matrix (OIM) model, that
                 solves these problems, and determines cardinal
                 directions for even complex regions. The model consists
                 of a tiling phase and an interpretation phase. In the
                 tiling phase, a tiling strategy first determines the
                 zones belonging to the nine cardinal directions of each
                 individual region object and then intersects them. The
                 result leads to a bounded grid called objects
                 interaction grid. For each grid cell the information
                 about the region objects that intersect it is stored in
                 an objects interaction matrix. In the subsequent
                 interpretation phase, a well-defined interpretation
                 method is applied to such a matrix and determines the
                 cardinal direction. Spatial example queries illustrate
                 our new cardinal direction concept that is embedded in
                 a spatial extension of SQL and provides user-defined
                 cardinal direction predicates.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wijsen:2012:CCQ,
  author =       "Jef Wijsen",
  title =        "Certain conjunctive query answering in first-order
                 logic",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "9:1--9:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188351",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Primary key violations provide a natural means for
                 modeling uncertainty in the relational data model. A
                 repair (or possible world) of a database is then
                 obtained by selecting a maximal number of tuples
                 without ever selecting two distinct tuples that have
                 the same primary key value. For a Boolean query q, the
                 problem CERTAINTY( q ) takes as input a database db and
                 asks whether q evaluates to true on every repair of db.
                 We are interested in determining queries q for which
                 CERTAINTY( q ) is first-order expressible (and hence in
                 the low complexity class AC ${}^\circ $ ). For queries
                 q in the class of conjunctive queries without
                 self-join, we provide a necessary syntactic condition
                 for first-order expressibility of CERTAINTY( q ). For
                 acyclic queries (in the sense of Beeri et al. [1983]),
                 this necessary condition is also a sufficient
                 condition. So we obtain a decision procedure for
                 first-order expressibility of CERTAINTY( q ) when q is
                 acyclic and without self-join. We also show that if
                 CERTAINTY( q ) is first-order expressible, its
                 first-order definition, commonly called certain
                 first-order rewriting, can be constructed in a rather
                 straightforward way.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Abiteboul:2012:CWS,
  author =       "Serge Abiteboul and Pierre Bourhis and Victor Vianu",
  title =        "Comparing workflow specification languages: a matter
                 of views",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "10:1--10:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188352",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We address the problem of comparing the expressiveness
                 of workflow specification formalisms using a notion of
                 view of a workflow. Views allow to compare widely
                 different workflow systems by mapping them to a common
                 representation capturing the observables relevant to
                 the comparison. Using this framework, we compare the
                 expressiveness of several workflow specification
                 mechanisms, including automata, temporal constraints,
                 and pre-and postconditions, with XML and relational
                 databases as underlying data models. One surprising
                 result shows the considerable power of static
                 constraints to simulate apparently much richer workflow
                 control mechanisms.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tassa:2012:SDC,
  author =       "Tamir Tassa and Ehud Gudes",
  title =        "Secure distributed computation of anonymized views of
                 shared databases",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "11:1--11:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188353",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider the problem of computing efficient
                 anonymizations of partitioned databases. Given a
                 database that is partitioned between several sites,
                 either horizontally or vertically, we devise secure
                 distributed algorithms that allow the different sites
                 to obtain a k -anonymized and l-diverse view of the
                 union of their databases, without disclosing sensitive
                 information. Our algorithms are based on the sequential
                 algorithm [Goldberger and Tassa 2010] that offers
                 anonymizations with utility that is significantly
                 better than other anonymization algorithms, and in
                 particular those that were implemented so far in the
                 distributed setting. Our algorithms can apply to
                 different generalization techniques and utility
                 measures and to any number of sites. While previous
                 distributed algorithms depend on costly cryptographic
                 primitives, the cryptographic assumptions of our
                 solution are surprisingly minimal.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sheng:2012:EAA,
  author =       "Cheng Sheng and Yufei Tao and Jianzhong Li",
  title =        "Exact and approximate algorithms for the most
                 connected vertex problem",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "12:1--12:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188354",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "An (edge) hidden graph is a graph whose edges are
                 notexplicitly given. Detecting the presence of an edge
                 requires an expensive edge probing query. We consider
                 the $k$ Most Connected Vertex ($k$-MCV) problem on
                 hidden bipartite graphs. Given a bipartite graph $G$
                 with independent vertex sets $B$ and $W$, the goal is
                 to find the $k$ vertices in $B$ with the largest
                 degrees using the minimum number of queries. This
                 problem can be regarded as a top-$k$ extension of
                 semi-join, and is encountered in several applications
                 in practice. If $B$ and $W$ have $n$ and $m$ vertices,
                 respectively, the number of queries needed to solve the
                 problem is $ n m $ in the worst case. This, however, is
                 a pessimistic estimate on how many queries are
                 necessary on practical data. In fact, on some inputs,
                 the problem may be settled with only $ k m + n $
                 queries, which is significantly lower than $ n m $ for
                 $ k \ll n $. The huge difference between $ k m + n $
                 and $ n m $ makes it interesting to design an adaptive
                 algorithm that is guaranteed to achieve the best
                 possible performance on every input $G$. For $ k \leq n
                 / 2 $, we give an algorithm that is instance optimal
                 among a broad class of solutions. This means that, for
                 any $G$, our algorithm can perform more queries than
                 the optimal solution (which is unknown) by only a
                 constant factor, which can be shown at most $2$. As a
                 second step, we study an $ \epsilon $-approximate
                 version of the $k$-MCV problem, where $ \epsilon $ is a
                 parameter satisfying $ 0 < \epsilon < 1 $. The goal is
                 to return $k$ black vertices $ b_1, \ldots, b_k $ such
                 that the degree of $ b_i (i \leq k) $ can be smaller
                 than $ t_i $ by a factor of at most $ \epsilon $, where
                 $ t_i, \ldots, t_k $ (in nonascending order) are the
                 degrees of the $k$ most connected black vertices. We
                 give an efficient randomized algorithm that
                 successfully finds the correct answer with high
                 probability. In particular, for a fixed $ \epsilon $
                 and a fixed success probability, our algorithm performs
                 $ o(n m) $ queries in expectation for $ t_k = \omega
                 (\log n) $. In other words, whenever $ t_k $ is greater
                 than $ \log n $ by more than a constant, our algorithm
                 beats the $ \Omega (n m) $ lower bound for solving the
                 $k$-MCV problem exactly. All the proposed algorithms,
                 despite the complication of their underlying theory,
                 are simple enough for easy implementation in practice.
                 Extensive experiments have confirmed that their
                 performance in reality agrees with our theoretical
                 findings very well.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hartmann:2012:IPD,
  author =       "Sven Hartmann and Sebastian Link",
  title =        "The implication problem of data dependencies over
                 {SQL} table definitions: {Axiomatic}, algorithmic and
                 logical characterizations",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "13:1--13:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188355",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We investigate the implication problem for classes of
                 data dependencies over SQL table definitions. Under
                 Zaniolo's ``no information'' interpretation of null
                 markers we establish an axiomatization and algorithms
                 to decide the implication problem for the combined
                 class of functional and multivalued dependencies in the
                 presence of NOT NULL constraints. The resulting theory
                 subsumes three previously orthogonal frameworks. We
                 further show that the implication problem of this class
                 is equivalent to that in a propositional fragment of
                 Schaerf and Cadoli's [1995] family of para-consistent
                 S-3 logics. In particular, S is the set of variables
                 that correspond to attributes declared NOT NULL. We
                 also show how our equivalences for multivalued
                 dependencies can be extended to Delobel's class of full
                 first-order hierarchical decompositions, and the
                 equivalences for functional dependencies can be
                 extended to arbitrary Boolean dependencies. These
                 dualities allow us to transfer several findings from
                 the propositional fragments to the corresponding
                 classes of data dependencies, and vice versa. We show
                 that our results also apply to Codd's null
                 interpretation ``value unknown at present'', but not to
                 Imielinski's [1989] or-relations utilizing Levene and
                 Loizou's weak possible world semantics [Levene and
                 Loizou 1998]. Our findings establish NOT NULL
                 constraints as an effective mechanism to balance not
                 only the certainty in database relations but also the
                 expressiveness with the efficiency of entailment
                 relations. They also control the degree by which the
                 implication of data dependencies over total relations
                 is soundly approximated in SQL table definitions.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhang:2012:SS,
  author =       "Wenjie Zhang and Xuemin Lin and Ying Zhang and
                 Muhammad Aamir Cheema and Qing Zhang",
  title =        "Stochastic skylines",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "14:1--14:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188356",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In many applications involving multiple criteria
                 optimal decision making, users may often want to make a
                 personal trade-off among all optimal solutions for
                 selecting one object that fits best their personal
                 needs. As a key feature, the skyline in a
                 multidimensional space provides the minimum set of
                 candidates for such purposes by removing all points not
                 preferred by any (monotonic) utility/scoring functions;
                 that is, the skyline removes all objects not preferred
                 by any user no matter how their preferences vary.
                 Driven by many recent applications with uncertain data,
                 the probabilistic skyline model is proposed to retrieve
                 uncertain objects based on skyline probabilities.
                 Nevertheless, skyline probabilities cannot capture the
                 preferences of monotonic utility functions. Motivated
                 by this, in this article we propose a novel skyline
                 operator, namely stochastic skylines. In the light of
                 the expected utility principle, stochastic skylines
                 guarantee to provide the minimum set of candidates to
                 optimal solutions over a family of utility functions.
                 We first propose the lskyline operator based on the
                 lower orthant orders. lskyline guarantees to provide
                 the minimum set of candidates to the optimal solutions
                 for the family of monotonic multiplicative utility
                 functions. While lskyline works very effectively for
                 the family of multiplicative functions, it may miss
                 optimal solutions for other utility /scoring functions
                 (e.g., linear functions). To resolve this, we also
                 propose a general stochastic skyline operator,
                 gskyline, based on the usual orders. gskyline provides
                 the minimum candidate set to the optimal solutions for
                 all monotonic functions. For the first time regarding
                 the existing literature, we investigate the
                 complexities of determining a stochastic order between
                 two uncertain objects whose probability distributions
                 are described discretely. We firstly show that
                 determining the lower orthant order is NP-complete with
                 respect to the dimensionality; consequently the problem
                 of computing lskyline is NP-complete. We also show an
                 interesting result as follows. While the usual order
                 involves more complicated geometric forms than the
                 lower orthant order, the usual order may be determined
                 in polynomial time regarding all the inputs, including
                 the dimensionality; this implies that gskyline can be
                 computed in polynomial time. A general framework is
                 developed for efficiently and effectively retrieving
                 lskyline and gskyline from a set of uncertain objects,
                 respectively, together with efficient and effective
                 filtering techniques. Novel and efficient verification
                 algorithms are developed to efficiently compute
                 lskyline over multidimensional uncertain data, which
                 run in polynomial time if the dimensionality is fixed,
                 and to efficiently compute gskyline in polynomial time
                 regarding all inputs. We also show, by theoretical
                 analysis and experiments, that the sizes of lskyline
                 and gskyline are both quite similar to that of
                 conventional skyline over certain data. Comprehensive
                 experiments demonstrate that our techniques are
                 efficient and scalable regarding both CPU and IO
                 costs.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Xia:2012:OSS,
  author =       "Tian Xia and Donghui Zhang and Zheng Fang and Cindy
                 Chen and Jie Wang",
  title =        "Online subspace skyline query processing using the
                 {Compressed SkyCube}",
  journal =      j-TODS,
  volume =       "37",
  number =       "2",
  pages =        "15:1--15:??",
  month =        may,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2188349.2188357",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 1 17:45:19 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The skyline query can help identify the ``best''
                 objects in a multi-attribute dataset. During the past
                 decade, this query has received considerable attention
                 in the database research community. Most research
                 focused on computing the ``skyline'' of a dataset, or
                 the set of ``skyline objects'' that are not dominated
                 by any other object. Such algorithms are not
                 appropriate in an online system, which should respond
                 in real time to skyline query requests with arbitrary
                 subsets of the attributes (also called subspaces). To
                 guarantee real-time response, an online system should
                 precompute the skylines for all subspaces, and look up
                 a skyline upon query. Unfortunately, because the number
                 of subspaces is exponential to the number of
                 attributes, such pre computation has very expensive
                 storage cost and update cost. We propose the Compressed
                 SkyCube (CSC) that is much more compact, yet can still
                 return the skyline of any subspace without consulting
                 the base table. The CSC therefore combines the
                 advantage of precomputation in that it can respond to
                 queries in real time, and the advantage of
                 no-precomputation in that it has efficient space cost
                 and update cost. This article presents the CSC data
                 structures, the CSC query algorithm, the CSC update
                 algorithm, and the CSC initial computation scheme. A
                 solution to extend to high-dimensional data is also
                 proposed.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ozsoyoglu:2012:Fa,
  author =       "Z. Meral {\"O}zsoyoglu",
  title =        "Foreword",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338627",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Graefe:2012:FBT,
  author =       "Goetz Graefe and Hideaki Kimura and Harumi Kuno",
  title =        "{Foster} {B}-trees",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338630",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Foster B-trees are a new variant of B-trees that
                 combines advantages of prior B-tree variants optimized
                 for many-core processors and modern memory hierarchies
                 with flash storage and nonvolatile memory. Specific
                 goals include: (i) minimal concurrency control
                 requirements for the data structure, (ii) efficient
                 migration of nodes to new storage locations, and (iii)
                 support for continuous and comprehensive self-testing.
                 Like B$^{\rm link}$-trees, Foster B-trees optimize
                 latching without imposing restrictions or specific
                 designs on transactional locking, for example, key
                 range locking. Like write-optimized B-trees, and unlike
                 B$^{\rm link}$-trees, Foster B-trees enable large
                 writes on RAID and flash devices as well as wear
                 leveling and efficient defragmentation. Finally, they
                 support continuous and inexpensive yet comprehensive
                 verification of all invariants, including all
                 cross-node invariants of the B-tree structure. An
                 implementation and a performance evaluation show that
                 the Foster B-tree supports high concurrency and high
                 update rates without compromising consistency,
                 correctness, or read performance.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wang:2012:RAT,
  author =       "Junhu Wang and Jeffrey Xu Yu",
  title =        "Revisiting answering tree pattern queries using
                 views",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338631",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We revisit the problem of answering tree pattern
                 queries using views. We first show that, for queries
                 and views that do not have nodes labeled with the
                 wildcard *, there is an approach which does not require
                 us to find any rewritings explicitly, yet which
                 produces the same answers as the maximal contained
                 rewriting. Then, using the new approach, we give simple
                 conditions and a corresponding algorithm for
                 identifying redundant view answers, which are view
                 answers that can be ignored when evaluating the maximal
                 contained rewriting. We also consider redundant view
                 answers in the case where there are multiple views, the
                 relationship between redundant views and redundant view
                 answers, and discuss how to combine the removal of
                 redundant view answers and redundant rewritings. We
                 show that the aforesaid results can be extended to a
                 number of other special cases. Finally, for arbitrary
                 queries and views in P$^{{/, / /, ., []}}$, we provide
                 a method to find the maximal contained rewriting and
                 show how to answer the query using views without
                 explicitly finding the rewritings.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{David:2012:ERA,
  author =       "Claire David and Leonid Libkin and Tony Tan",
  title =        "Efficient reasoning about data trees via integer
                 linear programming",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338632",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data trees provide a standard abstraction of XML
                 documents with data values: they are trees whose nodes,
                 in addition to the usual labels, can carry labels from
                 an infinite alphabet (data). Therefore, one is
                 interested in decidable formalisms for reasoning about
                 data trees. While some are known-such as the
                 two-variable logic-they tend to be of very high
                 complexity, and most decidability proofs are highly
                 nontrivial. We are therefore interested in reasonable
                 complexity formalisms as well as better techniques for
                 proving decidability. Here we show that many decidable
                 formalisms for data trees are subsumed-fully or
                 partially-by the power of tree automata together with
                 set constraints and linear constraints on cardinalities
                 of various sets of data values. All these constraints
                 can be translated into instances of integer linear
                 programming, giving us an NP upper bound on the
                 complexity of the reasoning tasks. We prove that this
                 bound, as well as the key encoding technique, remain
                 very robust, and allow the addition of features such as
                 counting of paths and patterns, and even a concise
                 encoding of constraints, without increasing the
                 complexity. The NP bound is tight, as we also show that
                 the satisfiability of a single set constraint is
                 already NP-hard. We then relate our results to several
                 reasoning tasks over XML documents, such as
                 satisfiability of schemas and data dependencies and
                 satisfiability of the two-variable logic. As a final
                 contribution, we describe experimental results based on
                 the implementation of some reasoning tasks using the
                 SMT solver Z3.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lemire:2012:RRB,
  author =       "Daniel Lemire and Owen Kaser and Eduardo Gutarra",
  title =        "Reordering rows for better compression: {Beyond} the
                 lexicographic order",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "20:1--20:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338633",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Sorting database tables before compressing them
                 improves the compression rate. Can we do better than
                 the lexicographical order? For minimizing the number of
                 runs in a run-length encoding compression scheme, the
                 best approaches to row-ordering are derived from
                 traveling salesman heuristics, although there is a
                 significant trade-off between running time and
                 compression. A new heuristic, Multiple Lists, which is
                 a variant on Nearest Neighbor that trades off
                 compression for a major running-time speedup, is a good
                 option for very large tables. However, for some
                 compression schemes, it is more important to generate
                 long runs rather than few runs. For this case, another
                 novel heuristic, Vortex, is promising. We find that we
                 can improve run-length encoding up to a factor of 3
                 whereas we can improve prefix coding by up to 80\%:
                 these gains are on top of the gains due to
                 lexicographically sorting the table. We prove that the
                 new row reordering is optimal (within 10\%) at
                 minimizing the runs of identical values within columns,
                 in a few cases.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gupta:2012:EQE,
  author =       "Nitin Gupta and Lucja Kot and Sudip Roy and Gabriel
                 Bender and Johannes Gehrke and Christoph Koch",
  title =        "Entangled queries: {Enabling} declarative data-driven
                 coordination",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "21:1--21:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338629",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many data-driven social and Web applications involve
                 collaboration and coordination. The vision of
                 Declarative Data-Driven Coordination (D3C), proposed in
                 Kot et al. [2010], is to support coordination in the
                 spirit of data management: to make it data-centric and
                 to specify it using convenient declarative languages.
                 This article introduces entangled queries, a language
                 that extends SQL by constraints that allow for the
                 coordinated choice of result tuples across queries
                 originating from different users or applications. It is
                 nontrivial to define a declarative coordination
                 formalism without arriving at the general (NP-complete)
                 Constraint Satisfaction Problem from AI. In this
                 article, we propose an efficiently enforceable
                 syntactic safety condition that we argue is at the
                 sweet spot where interesting declarative power meets
                 applicability in large-scale data management systems
                 and applications. The key computational problem of D3C
                 is to match entangled queries to achieve coordination.
                 We present an efficient matching algorithm which
                 statically analyzes query workloads and merges
                 coordinating entangled queries into compound SQL
                 queries. These can be sent to a standard database
                 system and return only coordinated results. We present
                 the overall architecture of an implemented system that
                 contains our evaluation algorithm. We also describe a
                 proof-of-concept Facebook application we have built on
                 top of this system to allow friends to coordinate
                 flight plans. Finally, we evaluate the performance of
                 the matching algorithm experimentally on realistic
                 coordination workloads.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Damaggio:2012:ASD,
  author =       "Elio Damaggio and Alin Deutsch and Victor Vianu",
  title =        "Artifact systems with data dependencies and
                 arithmetic",
  journal =      j-TODS,
  volume =       "37",
  number =       "3",
  pages =        "22:1--22:??",
  month =        aug,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2338626.2338628",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Sep 6 09:52:39 MDT 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the static verification problem for
                 data-centric business processes, specified in a variant
                 of IBM's ``business artifact'' model. Artifacts are
                 records of variables that correspond to
                 business-relevant objects and are updated by a set of
                 services equipped with pre- and postconditions, that
                 implement business process tasks. The verification
                 problem consists in statically checking whether all
                 runs of an artifact system satisfy desirable properties
                 expressed in a first-order extension of linear-time
                 temporal logic. Previous work identified the class of
                 guarded artifact systems and properties, for which
                 verification is decidable. However, the results suffer
                 an important limitation: they fail in the presence of
                 even very simple data dependencies or arithmetic, both
                 crucial to real-life business processes. In this
                 article, we extend the artifact model and verification
                 results to alleviate this limitation. We identify a
                 practically significant class of business artifacts
                 with data dependencies and arithmetic, for which
                 verification is decidable. The technical machinery
                 needed to establish the results is fundamentally
                 different from previous work. While the worst-case
                 complexity of verification is nonelementary, we
                 identify various realistic restrictions yielding more
                 palatable upper bounds.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ozsoyoglu:2012:Fb,
  author =       "Z. Meral {\"O}zsoyoglu",
  title =        "Foreword",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "23:1--23:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389242",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kimelfeld:2012:MCV,
  author =       "Benny Kimelfeld and Jan Vondr{\'a}k and Ryan
                 Williams",
  title =        "Maximizing Conjunctive Views in Deletion Propagation",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "24:1--24:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389243",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In deletion propagation, tuples from the database are
                 deleted in order to reflect the deletion of a tuple
                 from the view. Such an operation may result in the
                 (often necessary) deletion of additional tuples from
                 the view, besides the intentionally deleted one. The
                 article studies the complexity of deletion propagation,
                 where the view is defined by a conjunctive query (CQ),
                 and the goal is to maximize the number of tuples that
                 remain in the view. Buneman et al. showed that for some
                 simple CQs, this problem can be solved by a
                 straightforward algorithm, which is called here the
                 unidimensional algorithm. The article identifies
                 additional cases of CQs where the unidimensional
                 algorithm succeeds, and in contrast, shows that for
                 some other CQs the problem is NP-hard to approximate
                 better than some constant ratio. In fact, it is shown
                 here that among the CQs without self joins, the hard
                 CQs are exactly the ones that the unidimensional
                 algorithm fails on. In other words, the following
                 dichotomy result is proved: for every CQ without self
                 joins, deletion propagation is either APX-hard or
                 solvable (in polynomial time) by the unidimensional
                 algorithm. The article then presents approximation
                 algorithms for certain CQs where deletion propagation
                 is APX-hard. Specifically, two constant-ratio (and
                 polynomial-time) approximation algorithms are given for
                 the class of sunflower CQs (i.e., CQs having a
                 sunflower hypergraph) without self joins. The first
                 algorithm, providing the approximation ratio $ 1 - 1 /
                 e $, is obtained by formulating the problem at hand as
                 that of maximizing a monotone submodular function
                 subject to a matroid constraint, and then using a known
                 algorithm for such maximization. The second algorithm
                 gives a smaller approximation ratio, $ 1 / 2 $, yet in
                 polynomial time even under combined complexity.
                 Finally, it is shown that self joins can significantly
                 harden approximation in deletion propagation.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2012:DCD,
  author =       "Wenfei Fan and Floris Geerts and Jef Wijsen",
  title =        "Determining the Currency of Data",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "25:1--25:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389244",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data in real-life databases become obsolete rapidly.
                 One often finds that multiple values of the same entity
                 reside in a database. While all of these values were
                 once correct, most of them may have become stale and
                 inaccurate. Worse still, the values often do not carry
                 reliable timestamps. With this comes the need for
                 studying data currency, to identify the current value
                 of an entity in a database and to answer queries with
                 the current values, in the absence of reliable
                 timestamps. This article investigates the currency of
                 data. (1) We propose a model that specifies partial
                 currency orders in terms of simple constraints. The
                 model also allows us to express what values are copied
                 from other data sources, bearing currency orders in
                 those sources, in terms of copy functions defined on
                 correlated attributes. (2) We study fundamental
                 problems for data currency, to determine whether a
                 specification is consistent, whether a value is more
                 current than another, and whether a query answer is
                 certain no matter how partial currency orders are
                 completed. (3) Moreover, we identify several problems
                 associated with copy functions, to decide whether a
                 copy function imports sufficient current data to answer
                 a query, whether a copy function can be extended to
                 import necessary current data for a query while
                 respecting the constraints, and whether it suffices to
                 copy data of a bounded size. (4) We establish upper and
                 lower bounds of these problems, all matching, for
                 combined complexity and data complexity, and for a
                 variety of query languages. We also identify special
                 cases that warrant lower complexity.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sheng:2012:WCE,
  author =       "Cheng Sheng and Yufei Tao",
  title =        "Worst-Case {I/O}-Efficient Skyline Algorithms",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "26:1--26:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389245",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider the skyline problem (aka the maxima
                 problem), which has been extensively studied in the
                 database community. The input is a set $P$ of $d$
                 dimensional points. A point dominates another if the
                 coordinate of the former is at most that of the latter
                 on every dimension. The goal is to find the skyline,
                 which is the set of points $ p \in P $ such that $p$ is
                 not dominated by any other point in $P$. The main
                 result of this article is that, for any fixed
                 dimensionality $ d \geq 3 $, in external memory the
                 skyline problem can be settled by performing $ O((N /
                 B) \log_{M / B}^{d - 2}(N / B)) $ I/Os in the worst
                 case, where $N$ is the cardinality of $P$, $B$ the size
                 of a disk block, and $M$ the capacity of main memory.
                 Similar bounds can also be achieved for computing
                 several skyline variants, including the $k$-dominant
                 skyline, $k$-skyband, and $ \alpha $ skyline.
                 Furthermore, the performance can be improved if some
                 dimensions of the data space have small domains. When
                 the dimensionality $d$ is not fixed, the challenge is
                 to outperform the naive algorithm that simply checks
                 all pairs of points in $ P \times P $. We give an
                 algorithm that terminates in $ O((N / B) \log^{d - 2}
                 N) $ I/Os, thus beating the naive solution for any $ d
                 = O(\log N / \log \log N) $.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2012:SPS,
  author =       "Boduo Li and Edward Mazur and Yanlei Diao and Andrew
                 McGregor and Prashant Shenoy",
  title =        "{SCALLA}: a Platform for Scalable One-Pass Analytics
                 Using {MapReduce}",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "27:1--27:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389246",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Today's one-pass analytics applications tend to be
                 data-intensive in nature and require the ability to
                 process high volumes of data efficiently. MapReduce is
                 a popular programming model for processing large
                 datasets using a cluster of machines. However, the
                 traditional MapReduce model is not well-suited for
                 one-pass analytics, since it is geared towards batch
                 processing and requires the dataset to be fully loaded
                 into the cluster before running analytical queries.
                 This article examines, from a systems standpoint, what
                 architectural design changes are necessary to bring the
                 benefits of the MapReduce model to incremental one-pass
                 analytics. Our empirical and theoretical analyses of
                 Hadoop-based MapReduce systems show that the widely
                 used sort-merge implementation for partitioning and
                 parallel processing poses a fundamental barrier to
                 incremental one-pass analytics, despite various
                 optimizations. To address these limitations, we propose
                 a new data analysis platform that employs hash
                 techniques to enable fast in-memory processing, and a
                 new frequent key based technique to extend such
                 processing to workloads that require a large key-state
                 space. Evaluation of our Hadoop-based prototype using
                 real-world workloads shows that our new platform
                 significantly improves the progress of map tasks,
                 allows the reduce progress to keep up with the map
                 progress, with up to 3 orders of magnitude reduction of
                 internal data spills, and enables results to be
                 returned continuously during the job.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2012:WHT,
  author =       "Yinan Li and Jignesh M. Patel and Allison Terrell",
  title =        "{WHAM}: a High-Throughput Sequence Alignment Method",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "28:1--28:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389247",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Over the last decade, the cost of producing genomic
                 sequences has dropped dramatically due to the current
                 so-called next-generation sequencing methods. However,
                 these next-generation sequencing methods are critically
                 dependent on fast and sophisticated data processing
                 methods for aligning a set of query sequences to a
                 reference genome using rich string matching models. The
                 focus of this work is on the design, development and
                 evaluation of a data processing system for this crucial
                 ``short read alignment'' problem. Our system, called
                 WHAM, employs hash-based indexing methods and bitwise
                 operations for sequence alignments. It allows rich
                 match models and it is significantly faster than the
                 existing state-of-the-art methods. In addition, its
                 relative speedup over the existing method is poised to
                 increase in the future in which read sequence lengths
                 will increase.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Xie:2012:FAS,
  author =       "Kexin Xie and Ke Deng and Shuo Shang and Xiaofang Zhou
                 and Kai Zheng",
  title =        "Finding Alternative Shortest Paths in Spatial
                 Networks",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "29:1--29:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389248",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Shortest path query is one of the most fundamental
                 queries in spatial network databases. There exist
                 algorithms that can process shortest path queries in
                 real time. However, many complex applications require
                 more than just the calculation of a single shortest
                 path. For example, one of the common ways to determine
                 the importance (or price) of a vertex or an edge in
                 spatial network is to use Vickrey pricing, which
                 intuitively values the vertex $v$ (or edge $e$ ) based
                 on how much harder for travelling from the sources to
                 the destinations without using $v$ (or $e$ ). In such
                 cases, the alternative shortest paths without using $v$
                 (or $e$ ) are required. In this article, we propose
                 using a precomputation based approach for both single
                 pair alternative shortest path and all pairs shortest
                 paths processing. To compute the alternative shortest
                 path between a source and a destination efficiently, a
                 na{\"\i}ive way is to precompute and store all
                 alternative shortest paths between every pair of
                 vertices avoiding every possible vertex (or edge),
                 which requires $ O(n^4) $ space. Currently, the state
                 of the art approach for reducing the storage cost is to
                 choose a subset of the vertices as center points, and
                 only store the single-source alternative shortest paths
                 from those center points. Such approach has the space
                 complexity of $ O(n^2 \log n) $. We propose a storage
                 scheme termed iSPQF, which utilizes shortest path
                 quadtrees by observing the relationships between each
                 avoiding vertex and its corresponding alternative
                 shortest paths. We have reduced the space complexity
                 from the na{\"\i}ive $ O(n^4) $ (or the state of the
                 art $ O(n^4 \log n) $) to $ O(\min (\gamma, L) n^{1.5})
                 $ with comparable query performance of $ O(K) $, where
                 $K$ is the number of vertices in the returned paths,
                 $L$ is the diameter of the spatial network, and $
                 \gamma $ is a value that depends on the structure of
                 the spatial network, which is empirically estimated to
                 be 40 for real road networks. Experiments on real road
                 networks have shown that the space cost of the proposed
                 iSPQF is scalable, and both the algorithms based on
                 iSPQF are efficient.",
  acknowledgement = ack-nhfb,
  articleno =    "29",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Amsterdamer:2012:PM,
  author =       "Yael Amsterdamer and Daniel Deutch and Tova Milo and
                 Val Tannen",
  title =        "On Provenance Minimization",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "30:1--30:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389249",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Provenance information has been proved to be very
                 effective in capturing the computational process
                 performed by queries, and has been used extensively as
                 the input to many advanced data management tools (e.g.,
                 view maintenance, trust assessment, or query answering
                 in probabilistic databases). We observe here that while
                 different (set-)equivalent queries may admit different
                 provenance expressions when evaluated on the same
                 database, there is always some part of these
                 expressions that is common to all. We refer to this
                 part as the core provenance. In addition to being
                 informative, the core provenance is also useful as a
                 compact input to the aforementioned data management
                 tools. We formally define the notion of core
                 provenance. We study algorithms that, given a query,
                 compute an equivalent (called p-minimal) query that for
                 every input database, the provenance of every result
                 tuple is the core provenance. We study such algorithms
                 for queries of varying expressive power (namely
                 conjunctive queries with disequalities and unions
                 thereof). Finally, we observe that, in general, one
                 would not want to require database systems to execute a
                 specific p-minimal query, but instead to be able to
                 find, possibly off-line, the core provenance of a given
                 tuple in the output (computed by an arbitrary
                 equivalent query), without reevaluating the query. We
                 provide algorithms for such direct computation of the
                 core provenance.",
  acknowledgement = ack-nhfb,
  articleno =    "30",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Barcelo:2012:ELP,
  author =       "Pablo Barcel{\'o} and Leonid Libkin and Anthony W. Lin
                 and Peter T. Wood",
  title =        "Expressive Languages for Path Queries over
                 Graph-Structured Data",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "31:1--31:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389250",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "For many problems arising in the setting of graph
                 querying (such as finding semantic associations in RDF
                 graphs, exact and approximate pattern matching,
                 sequence alignment, etc.), the power of standard
                 languages such as the widely studied conjunctive
                 regular path queries (CRPQs) is insufficient in at
                 least two ways. First, they cannot output paths and
                 second, more crucially, they cannot express
                 relationships among paths. We thus propose a class of
                 extended CRPQs, called ECRPQs, which add regular
                 relations on tuples of paths, and allow path variables
                 in the heads of queries. We provide several examples of
                 their usefulness in querying graph structured data, and
                 study their properties. We analyze query evaluation and
                 representation of tuples of paths in the output by
                 means of automata. We present a detailed analysis of
                 data and combined complexity of queries, and consider
                 restrictions that lower the complexity of ECRPQs to
                 that of relational conjunctive queries. We study the
                 containment problem, and look at further extensions
                 with first-order features, and with nonregular
                 relations that add arithmetic constraints on the
                 lengths of paths and numbers of occurrences of
                 labels.",
  acknowledgement = ack-nhfb,
  articleno =    "31",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cong:2012:PED,
  author =       "Gao Cong and Wenfei Fan and Anastasios Kementsietsidis
                 and Jianzhong Li and Xianmin Liu",
  title =        "Partial Evaluation for Distributed {XPath} Query
                 Processing and Beyond",
  journal =      j-TODS,
  volume =       "37",
  number =       "4",
  pages =        "32:1--32:??",
  month =        dec,
  year =         "2012",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2389241.2389251",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Dec 20 19:03:29 MST 2012",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article proposes algorithms for evaluating XPath
                 queries over an XML tree that is partitioned
                 horizontally and vertically, and is distributed across
                 a number of sites. The key idea is based on partial
                 evaluation: it is to send the whole query to each site
                 that partially evaluates the query, in parallel, and
                 sends the results as compact (Boolean) functions to a
                 coordinator that combines these to obtain the result.
                 This approach possesses the following performance
                 guarantees. First, each site is visited at most twice
                 for data-selecting XPath queries, and only once for
                 Boolean XPath queries. Second, the network traffic is
                 determined by the answer to the query, rather than the
                 size of the tree. Third, the total computation is
                 comparable to that of centralized algorithms on the
                 tree stored in a single site, regardless of how the
                 tree is fragmented and distributed. We also present a
                 MapReduce algorithm for evaluating Boolean XPath
                 queries, based on partial evaluation. In addition, we
                 provide algorithms to evaluate XPath queries on very
                 large XML trees, in a centralized setting. We show both
                 analytically and empirically that our techniques are
                 scalable with large trees and complex XPath queries.
                 These results, we believe, illustrate the usefulness
                 and potential of partial evaluation in distributed
                 systems as well as centralized XML stores for
                 evaluating XPath queries and beyond.",
  acknowledgement = ack-nhfb,
  articleno =    "32",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papavasileiou:2013:HLC,
  author =       "Vicky Papavasileiou and Giorgos Flouris and Irini
                 Fundulaki and Dimitris Kotzinos and Vassilis
                 Christophides",
  title =        "High-level change detection in {RDF(S) KBs}",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "1:1--1:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "With the increasing use of Web 2.0 to create,
                 disseminate, and consume large volumes of data, more
                 and more information is published and becomes available
                 for potential data consumers, that is,
                 applications/services, individual users and
                 communities, outside their production site. The most
                 representative example of this trend is Linked Open
                 Data (LOD), a set of interlinked data and knowledge
                 bases. The main challenge in this context is data
                 governance within loosely coordinated organizations
                 that are publishing added-value interlinked data on the
                 Web, bringing together issues related to data
                 management and data quality, in order to support the
                 full lifecycle of data production, consumption, and
                 management. In this article, we are interested in
                 curation issues for RDF(S) data, which is the default
                 data model for LOD. In particular, we are addressing
                 change management for RDF(S) data maintained by large
                 communities (scientists, librarians, etc.) which act as
                 curators to ensure high quality of data. Such curated
                 Knowledge Bases (KBs) are constantly evolving for
                 various reasons, such as the inclusion of new
                 experimental evidence or observations, or the
                 correction of erroneous conceptualizations. Managing
                 such changes poses several research problems, including
                 the problem of detecting the changes (delta) between
                 versions of the same KB developed and maintained by
                 different groups of curators, a crucial task for
                 assisting them in understanding the involved changes.
                 This becomes all the more important as curated KBs are
                 interconnected (through copying or referencing) and
                 thus changes need to be propagated from one KB to
                 another either within or across communities. This
                 article addresses this problem by proposing a change
                 language which allows the formulation of concise and
                 intuitive deltas. The language is expressive enough to
                 describe unambiguously any possible change encountered
                 in curated KBs expressed in RDF(S), and can be
                 efficiently and deterministically detected in an
                 automated way. Moreover, we devise a change detection
                 algorithm which is sound and complete with respect to
                 the aforementioned language, and study appropriate
                 semantics for executing the deltas expressed in our
                 language in order to move backwards and forwards in a
                 multiversion repository, using only the corresponding
                 deltas. Finally, we evaluate through experiments the
                 effectiveness and efficiency of our algorithms using
                 real ontologies from the cultural, bioinformatics, and
                 entertainment domains.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jaiswal:2013:SME,
  author =       "Anuj Jaiswal and David J. Miller and Prasenjit Mitra",
  title =        "Schema matching and embedded value mapping for
                 databases with opaque column names and mixed continuous
                 and discrete-valued data fields",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "2:1--2:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Schema matching and value mapping across two
                 information sources, such as databases, are critical
                 information aggregation tasks. Before data can be
                 integrated from multiple tables, the columns and values
                 within the tables must be matched. The complexities of
                 both these problems grow quickly with the number of
                 attributes to be matched and due to multiple semantics
                 of data values. Traditional research has mostly tackled
                 schema matching and value mapping independently, and
                 for categorical (discrete-valued) attributes. We
                 propose novel methods that leverage value mappings to
                 enhance schema matching in the presence of opaque
                 column names for schemas consisting of both continuous
                 and discrete-valued attributes. An additional source of
                 complexity is that a discrete-valued attribute in one
                 schema could in fact be a quantized, encoded version of
                 a continuous-valued attribute in the other schema. In
                 our approach, which can tackle both ``onto'' and
                 bijective schema matching, the fitness objective for
                 matching a pair of attributes from two schemas exploits
                 the statistical distribution over values within the two
                 attributes. Suitable fitness objectives are based on
                 Euclidean-distance and the data log-likelihood, both of
                 which are applied in our experimental study. A
                 heuristic local descent optimization strategy that uses
                 two-opt switching to optimize attribute matches, while
                 simultaneously embedding value mappings, is applied for
                 our matching methods. Our experiments show that the
                 proposed techniques matched mixed continuous and
                 discrete-valued attribute schemas with high accuracy
                 and, thus, should be a useful addition to a framework
                 of (semi) automated tools for data alignment.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Grust:2013:OSQ,
  author =       "Torsten Grust and Jan Rittinger",
  title =        "Observing {SQL} queries in their natural habitat",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "3:1--3:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We describe Habitat, a declarative observational
                 debugger for SQL. Habitat facilitates true
                 language-level (not: plan-level) debugging of, probably
                 flawed, SQL queries that yield unexpected results.
                 Users mark SQL subexpressions of arbitrary size and
                 then observe whether these evaluate as expected.
                 Habitat understands query nesting and free row
                 variables in correlated subqueries, and generally aims
                 to not constrain users while suspect subexpressions are
                 marked for observation. From the marked SQL text,
                 Habitat's algebraic compiler derives a new query whose
                 result represents the values of the desired
                 observations. These observations are generated by the
                 target SQL database host itself and are derived from
                 the original data: Habitat does not require prior data
                 extraction or extra debugging middleware. Experiments
                 with TPC-H database instances indicate that
                 observations impose a runtime overhead sufficiently low
                 to allow for interactive debugging sessions.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Benzaken:2013:OXQ,
  author =       "V{\'e}ronique Benzaken and Giuseppe Castagna and Dario
                 Colazzo and Kim Nguy{\'{\^e}}n",
  title =        "Optimizing {XML} querying using type-based document
                 projection",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "4:1--4:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "XML data projection (or pruning) is a natural
                 optimization for main memory query engines: given a
                 query Q over a document D, the subtrees of D that are
                 not necessary to evaluate Q are pruned, thus producing
                 a smaller document D'; the query Q is then executed on
                 D', hence avoiding to allocate and process nodes that
                 will never be reached by Q. In this article, we propose
                 a new approach, based on types, that greatly improves
                 current solutions. Besides providing comparable or
                 greater precision and far lesser pruning overhead, our
                 solution-unlike current approaches-takes into account
                 backward axes, predicates, and can be applied to
                 multiple queries rather than just to single ones. A
                 side contribution is a new type system for XPath able
                 to handle backward axes. The soundness of our approach
                 is formally proved. Furthermore, we prove that the
                 approach is also complete (i.e., yields the best
                 possible type-driven pruning) for a relevant class of
                 queries and Schemas. We further validate our approach
                 using the XMark and XPathMark benchmarks and show that
                 pruning not only improves the main memory query
                 engine's performances (as expected) but also those of
                 state of the art native XML databases.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Das:2013:EES,
  author =       "Sudipto Das and Divyakant Agrawal and Amr {El
                 Abbadi}",
  title =        "{ElasTraS}: an elastic, scalable, and self-managing
                 transactional database for the cloud",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "5:1--5:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A database management system (DBMS) serving a cloud
                 platform must handle large numbers of application
                 databases (or tenants) that are characterized by
                 diverse schemas, varying footprints, and unpredictable
                 load patterns. Scaling out using clusters of commodity
                 servers and sharing resources among tenants (i.e.,
                 multitenancy) are important features of such systems.
                 Moreover, when deployed on a pay-per-use
                 infrastructure, minimizing the system's operating cost
                 while ensuring good performance is also an important
                 goal. Traditional DBMSs were not designed for such
                 scenarios and hence do not possess the mentioned
                 features critical for DBMSs in the cloud. We present
                 ElasTraS, which combines three design principles to
                 build an elastically-scalable multitenant DBMS for
                 transaction processing workloads. These design
                 principles are gleaned from a careful analysis of the
                 years of research in building scalable key-value stores
                 and decades of research in high performance transaction
                 processing systems. ElasTraS scales to thousands of
                 tenants, effectively consolidates tenants with small
                 footprints while scaling-out large tenants across
                 multiple servers in a cluster. ElasTraS also supports
                 low-latency multistep ACID transactions, is
                 fault-tolerant, self-managing, and highly available to
                 support mission critical applications. ElasTraS
                 leverages Albatross, a low overhead on-demand live
                 database migration technique, for elastic load
                 balancing by adding more servers during high load and
                 consolidating to fewer servers during usage troughs.
                 This elastic scaling minimizes the operating cost and
                 ensures good performance even in the presence of
                 unpredictable changes to the workload. We elucidate the
                 design principles, explain the architecture, describe a
                 prototype implementation, present the detailed design
                 and implementation of Albatross, and experimentally
                 evaluate the implementation using a variety of
                 transaction processing workloads. On a cluster of 20
                 commodity servers, our prototype serves thousands of
                 tenants and serves more than 1 billion transactions per
                 day while migrating tenant databases with minimal
                 overhead to allow lightweight elastic scaling. Using a
                 cluster of 30 commodity servers, ElasTraS can scale-out
                 a terabyte TPC-C database serving an aggregate
                 throughput of approximately one quarter of a million
                 TPC-C transactions per minute.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Franceschet:2013:GTA,
  author =       "Massimo Franceschet and Donatella Gubiani and Angelo
                 Montanari and Carla Piazza",
  title =        "A graph-theoretic approach to map conceptual designs
                 to {XML} schemas",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "6:1--6:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We propose a mapping from a database conceptual design
                 to a schema for XML that produces highly connected and
                 nested XML structures. We first introduce two
                 alternative definitions of the mapping, one modeling
                 entities as global XML elements and expressing
                 relationships among them in terms of keys and key
                 references (flat design), the other one encoding
                 relationships by properly including the elements for
                 some entities into the elements for other entities
                 (nest design). Then we provide a benchmark evaluation
                 of the two solutions showing that the nest approach,
                 compared to the flat one, leads to improvements in both
                 query and validation performances. This motivates us to
                 systematically investigate the best way to nest XML
                 structures. We identify two different nesting
                 solutions: a maximum depth nesting, that keeps low the
                 number of costly join operations that are necessary to
                 reconstruct information at query time using the mapped
                 schema, and a maximum density nesting, that minimizes
                 the number of schema constraints used in the mapping of
                 the conceptual schema, thus reducing the validation
                 overhead. On the one hand, the problem of finding a
                 maximum depth nesting turns out to be NP-complete and,
                 moreover, it admits no constant ratio approximation
                 algorithm. On the other hand, we devise a
                 graph-theoretic algorithm, NiduX, that solves the
                 maximum density problem in linear time. Interestingly,
                 NiduX finds the optimal solution for the harder maximum
                 depth problem whenever the conceptual design graph is
                 either acyclic or complete. In randomly generated
                 intermediate cases of the graph topology, we
                 experimentally show that NiduX finds a good
                 approximation of the optimal solution.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wu:2013:MSK,
  author =       "Dingming Wu and Man Lung Yiu and Christian S. Jensen",
  title =        "Moving spatial keyword queries: Formulation, methods,
                 and analysis",
  journal =      j-TODS,
  volume =       "38",
  number =       "1",
  pages =        "7:1--7:??",
  month =        apr,
  year =         "2013",
  CODEN =        "ATDSD3",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Apr 23 18:06:18 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Web users and content are increasingly being
                 geo-positioned. This development gives prominence to
                 spatial keyword queries, which involve both the
                 locations and textual descriptions of content. We study
                 the efficient processing of continuously moving top- k
                 spatial keyword (M k SK) queries over spatial text
                 data. State-of-the-art solutions for moving queries
                 employ safe zones that guarantee the validity of
                 reported results as long as the user remains within the
                 safe zone associated with a result. However, existing
                 safe-zone methods focus solely on spatial locations and
                 ignore text relevancy. We propose two algorithms for
                 computing safe zones that guarantee correct results at
                 any time and that aim to optimize the server-side
                 computation as well as the communication between the
                 server and the client. We exploit tight and
                 conservative approximations of safe zones and
                 aggressive computational space pruning. We present
                 techniques that aim to compute the next safe zone
                 efficiently, and we present two types of conservative
                 safe zones that aim to reduce the communication cost.
                 Empirical studies with real data suggest that the
                 proposals are efficient. To understand the
                 effectiveness of the proposed safe zones, we study
                 analytically the expected area of a safe zone, which
                 indicates on average for how long a safe zone remains
                 valid, and we study the expected number of influence
                 objects needed to define a safe zone, which gives an
                 estimate of the average communication cost. The
                 analytical modeling is validated through empirical
                 studies.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sadoghi:2013:AOB,
  author =       "Mohammad Sadoghi and Hans-Arno Jacobsen",
  title =        "Analysis and optimization for {Boolean} expression
                 indexing",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487260",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "BE-Tree is a novel dynamic data structure designed to
                 efficiently index Boolean expressions over a
                 high-dimensional discrete space. BE Tree-copes with
                 both high-dimensionality and expressiveness of Boolean
                 expressions by introducing an effective two-phase
                 space-cutting technique that specifically utilizes the
                 discrete and finite domain properties of the space.
                 Furthermore, BE-Tree employs self-adjustment policies
                 to dynamically adapt the tree as the workload changes.
                 Moreover, in BE-Tree, we develop two novel
                 cache-conscious predicate evaluation techniques,
                 namely, lazy and bitmap evaluations, that also exploit
                 the underlying discrete and finite space to
                 substantially reduce BE-Tree's matching time by up to
                 75\% BE-Tree is a general index structure for matching
                 Boolean expression which has a wide range of
                 applications including (complex) event processing,
                 publish/subscribe matching, emerging applications in
                 cospaces, profile matching for targeted web
                 advertising, and approximate string matching. Finally,
                 the superiority of BE-Tree is proven through a
                 comprehensive evaluation with state-of-the-art index
                 structures designed for matching Boolean expressions.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2013:PBM,
  author =       "Guoliang Li and Dong Deng and Jianhua Feng",
  title =        "A partition-based method for string similarity joins
                 with edit-distance constraints",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487261",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "As an essential operation in data cleaning, the
                 similarity join has attracted considerable attention
                 from the database community. In this article, we study
                 string similarity joins with edit-distance constraints,
                 which find similar string pairs from two large sets of
                 strings whose edit distance is within a given
                 threshold. Existing algorithms are efficient either for
                 short strings or for long strings, and there is no
                 algorithm that can efficiently and adaptively support
                 both short strings and long strings. To address this
                 problem, we propose a new filter, called the segment
                 filter. We partition a string into a set of segments
                 and use the segments as a filter to find similar string
                 pairs. We first create inverted indices for the
                 segments. Then for each string, we select some of its
                 substrings, identify the selected substrings from the
                 inverted indices, and take strings on the inverted
                 lists of the found substrings as candidates of this
                 string. Finally, we verify the candidates to generate
                 the final answer. We devise efficient techniques to
                 select substrings and prove that our method can
                 minimize the number of selected substrings. We develop
                 novel pruning techniques to efficiently verify the
                 candidates. We also extend our techniques to support
                 normalized edit distance. Experimental results show
                 that our algorithms are efficient for both short
                 strings and long strings, and outperform
                 state-of-the-art methods on real-world datasets.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Catallo:2013:TKD,
  author =       "Ilio Catallo and Eleonora Ciceri and Piero Fraternali
                 and Davide Martinenghi and Marco Tagliasacchi",
  title =        "Top-$k$ diversity queries over bounded regions",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487262",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Top-k diversity queries over objects embedded in a
                 low-dimensional vector space aim to retrieve the best k
                 objects that are both relevant to given user's criteria
                 and well distributed over a designated region. An
                 interesting case is provided by spatial Web objects,
                 which are produced in great quantity by location-based
                 services that let users attach content to places and
                 are found also in domains like trip planning, news
                 analysis, and real estate. In this article we present a
                 technique for addressing such queries that, unlike
                 existing methods for diversified top- k queries, does
                 not require accessing and scanning all relevant objects
                 in order to find the best k results. Our Space
                 Partitioning and Probing (SPP) algorithm works by
                 progressively exploring the vector space, while keeping
                 track of the already seen objects and of their
                 relevance and position. The goal is to provide a good
                 quality result set in terms of both relevance and
                 diversity. We assess quality by using as a baseline the
                 result set computed by MMR, one of the most popular
                 diversification algorithms, while minimizing the number
                 of accessed objects. In order to do so, SPP exploits
                 score-based and distance-based access methods, which
                 are available, for instance, in most geo-referenced Web
                 data sources. Experiments with both synthetic and real
                 data show that SPP produces results that are relevant
                 and spatially well distributed, while significantly
                 reducing the number of accessed objects and incurring a
                 very low computational overhead.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fazzinga:2013:RDC,
  author =       "Bettina Fazzinga and Sergio Flesca and Filippo Furfaro
                 and Elio Masciari",
  title =        "{RFID}-data compression for supporting aggregate
                 queries",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487263",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "RFID-based systems for object tracking and supply
                 chain management have been emerging since the RFID
                 technology proved effective in monitoring movements of
                 objects. The monitoring activity typically results in
                 huge numbers of readings, thus making the problem of
                 efficiently retrieving aggregate information from the
                 collected data a challenging issue. In fact, tackling
                 this problem is of crucial importance, as fast answers
                 to aggregate queries are often mandatory to support the
                 decision making process. In this regard, a compression
                 technique for RFID data is proposed, and used as the
                 core of a system supporting the efficient estimation of
                 aggregate queries. Specifically, this technique aims at
                 constructing a lossy synopsis of the data over which
                 aggregate queries can be estimated, without accessing
                 the original data. Owing to the lossy nature of the
                 compression, query estimates are approximate, and are
                 returned along with intervals that are guaranteed to
                 contain the exact query answers. The effectiveness of
                 the proposed approach has been experimentally
                 validated, showing a remarkable trade-off between the
                 efficiency and the accuracy of the query estimation.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pavlou:2013:GDF,
  author =       "Kyriacos E. Pavlou and Richard T. Snodgrass",
  title =        "Generalizing database forensics",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487264",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article we present refinements on previously
                 proposed approaches to forensic analysis of database
                 tampering. We significantly generalize the basic
                 structure of these algorithms to admit new
                 characterizations of the ``where'' axis of the
                 corruption diagram. Specifically, we introduce
                 page-based partitioning as well as attribute-based
                 partitioning along with their associated corruption
                 diagrams. We compare the structure of all the forensic
                 analysis algorithms and discuss the various design
                 choices available with respect to forensic analysis. We
                 characterize the forensic cost of the newly introduced
                 algorithms, compare their forensic cost, and give our
                 recommendations. We then introduce a comprehensive
                 taxonomy of the types of possible corruption events,
                 along with an associated forensic analysis protocol
                 that consolidates all extant forensic algorithms and
                 the corresponding type(s) of corruption events they
                 detect. The result is a generalization of these
                 algorithms and an overarching characterization of the
                 process of database forensic analysis, thus providing a
                 context within the overall operation of a DBMS for all
                 existing forensic analysis algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cheney:2013:RFN,
  author =       "James Cheney",
  title =        "Revisiting ``forward node-selecting queries over
                 trees''",
  journal =      j-TODS,
  volume =       "38",
  number =       "2",
  pages =        "13:1--13:??",
  month =        jun,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2487259.2487265",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 1 18:44:25 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In ``Forward Node-Selecting Queries over Trees,''
                 Olteanu [2007] gives three rewriting systems for
                 eliminating reverse XPath axis steps from
                 node-selecting queries over trees, together with
                 arguments for their correctness and termination for a
                 large class of input graphs, including cyclic ones.
                 These proofs are valid for tree or acyclic formulas,
                 but two of the rewrite systems ( TRS$_2$ and TRS$_3$ )
                 do not terminate on cyclic graphs; that is, there are
                 infinite rewrite sequences that never yield a normal
                 form. We investigate the reasons why the termination
                 arguments do not work for general cyclic formulas, and
                 develop alternative algorithms that can be used
                 instead. We prove that TRS$_2$ is weakly normalizing,
                 while TRS$_3$ is not weakly normalizing, but it can be
                 extended to a weakly normalizing system TRS$_3^\ocirc
                 $. The algorithms and proof techniques illustrate
                 unforeseen subtleties in the handling of cyclic
                 queries.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bonifati:2013:AIM,
  author =       "Angela Bonifati and Martin Goodfellow and Ioana
                 Manolescu and Domenica Sileo",
  title =        "Algebraic incremental maintenance of {XML} views",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "14:1--14:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2508020.2508021",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Materialized views can bring important performance
                 benefits when querying XML documents. In the presence
                 of XML document changes, materialized views need to be
                 updated to faithfully reflect the changed document. In
                 this work, we present an algebraic approach for
                 propagating source updates to XML materialized views
                 expressed in a powerful XML tree pattern formalism. Our
                 approach differs from the state-of-the-art in the area
                 in two important ways. First, it relies on
                 set-oriented, algebraic operations, to be contrasted
                 with node-based previous approaches. Second, it
                 exploits state-of-the-art features of XML stores and
                 XML query evaluation engines, notably XML structural
                 identifiers and associated structural join algorithms.
                 We present algorithms for determining how updates
                 should be propagated to views, and highlight the
                 benefits of our approach over existing algorithms
                 through a series of experiments.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Colazzo:2013:ALI,
  author =       "Dario Colazzo and Giorgio Ghelli and Luca Pardini and
                 Carlo Sartiani",
  title =        "Almost-linear inclusion for {XML} regular expression
                 types",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2508020.2508022",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Type inclusion is a fundamental operation in every
                 type-checking compiler, but it is quite expensive for
                 XML manipulation languages. A polynomial inclusion
                 checking algorithm for an expressive family of XML type
                 languages is known, but it runs in quadratic time both
                 in the best and in the worst cases. We present here an
                 algorithm that has a linear-time backbone, and resorts
                 to the quadratic approach for some specific parts of
                 the compared types. Our experiments show that the new
                 algorithm is much faster than the quadratic one, and
                 that it typically runs in linear time, hence it can be
                 used as a building block for a practical type-checking
                 compiler.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Qin:2013:ASS,
  author =       "Jianbin Qin and Wei Wang and Chuan Xiao and Yifei Lu
                 and Xuemin Lin and Haixun Wang",
  title =        "Asymmetric signature schemes for efficient exact edit
                 similarity query processing",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2508020.2508023",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Given a query string Q, an edit similarity search
                 finds all strings in a database whose edit distance
                 with {$Q$} is no more than a given threshold $ \tau $.
                 Most existing methods answering edit similarity queries
                 employ schemes to generate string subsequences as
                 signatures and generate candidates by set overlap
                 queries on query and data signatures. In this article,
                 we show that for any such signature scheme, the lower
                 bound of the minimum number of signatures is $ \tau + 1
                 $, which is lower than what is achieved by existing
                 methods. We then propose several asymmetric signature
                 schemes, that is, extracting different numbers of
                 signatures for the data and query strings, which
                 achieve this lower bound. A basic asymmetric scheme is
                 first established on the basis of matching $q$-chunks
                 and $q$-grams between two strings. Two efficient query
                 processing algorithms (IndexGram and IndexChunk) are
                 developed on top of this scheme. We also propose novel
                 candidate pruning methods to further improve the
                 efficiency. We then generalize the basic scheme by
                 incorporating novel ideas of floating $q$-chunks,
                 optimal selection of $q$-chunks, and reducing the
                 number of signatures using global ordering. As a
                 result, the Super and Turbo families of schemes are
                 developed together with their corresponding query
                 processing algorithms. We have conducted a
                 comprehensive experimental study using the six
                 asymmetric algorithms and nine previous
                 state-of-the-art algorithms. The experiment results
                 clearly showcase the efficiency of our methods and
                 demonstrate space and time characteristics of our
                 proposed algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Levandoski:2013:FEP,
  author =       "Justin J. Levandoski and Ahmed Eldawy and Mohamed F.
                 Mokbel and Mohamed E. Khalefa",
  title =        "Flexible and extensible preference evaluation in
                 database systems",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2493268",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Personalized database systems give users answers
                 tailored to their personal preferences. While numerous
                 preference evaluation methods for databases have been
                 proposed (e.g., skyline, top-k, k-dominance,
                 k-frequency), the implementation of these methods at
                 the core of a database system is a double-edged sword.
                 Core implementation provides efficient query processing
                 for arbitrary database queries, however, this approach
                 is not practical since each existing (and future)
                 preference method requires implementation within the
                 database engine. To solve this problem, this article
                 introduces FlexPref, a framework for extensible
                 preference evaluation in database systems. FlexPref,
                 implemented in the query processor, aims to support a
                 wide array of preference evaluation methods in a single
                 extensible code base. Integration with FlexPref is
                 simple, involving the registration of only three
                 functions that capture the essence of the preference
                 method. Once integrated, the preference method
                 ``lives'' at the core of the database, enabling the
                 efficient execution of preference queries involving
                 common database operations. This article also provides
                 a query optimization framework for FlexPref, as well as
                 a theoretical framework that defines the properties a
                 preference method must exhibit to be implemented in
                 FlexPref. To demonstrate the extensibility of FlexPref,
                 this article also provides case studies detailing the
                 implementation of seven state-of-the-art preference
                 evaluation methods within FlexPref. We also
                 experimentally study the strengths and weaknesses of an
                 implementation of FlexPref in PostgreSQL over a range
                 of single-table and multitable preference queries.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2013:IGP,
  author =       "Wenfei Fan and Xin Wang and Yinghui Wu",
  title =        "Incremental graph pattern matching",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2489791",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Graph pattern matching is commonly used in a variety
                 of emerging applications such as social network
                 analysis. These applications highlight the need for
                 studying the following two issues. First, graph pattern
                 matching is traditionally defined in terms of subgraph
                 isomorphism or graph simulation. These notions,
                 however, often impose too strong a topological
                 constraint on graphs to identify meaningful matches.
                 Second, in practice a graph is typically large, and is
                 frequently updated with small changes. It is often
                 prohibitively expensive to recompute matches starting
                 from scratch via batch algorithms when the graph is
                 updated. This article studies these two issues. (1) We
                 propose to define graph pattern matching based on a
                 notion of bounded simulation, which extends graph
                 simulation by specifying the connectivity of nodes in a
                 graph within a predefined number of hops. We show that
                 bounded simulation is able to find sensible matches
                 that the traditional matching notions fail to catch. We
                 also show that matching via bounded simulation is in
                 cubic time, by giving such an algorithm. (2) We provide
                 an account of results on incremental graph pattern
                 matching, for matching defined with graph simulation,
                 bounded simulation, and subgraph isomorphism. We show
                 that the incremental matching problem is unbounded,
                 that is, its cost is not determined alone by the size
                 of the changes in the input and output, for all these
                 matching notions. Nonetheless, when matching is defined
                 in terms of simulation or bounded simulation,
                 incremental matching is semibounded, that is, its
                 worst-time complexity is bounded by a polynomial in the
                 size of the changes in the input, output, and auxiliary
                 information that is necessarily maintained to reuse
                 previous computation, and the size of graph patterns.
                 We also develop incremental matching algorithms for
                 graph simulation and bounded simulation, by minimizing
                 unnecessary recomputation. In contrast, matching based
                 on subgraph isomorphism is neither bounded nor
                 semibounded. (3) We experimentally verify the
                 effectiveness and efficiency of these algorithms, and
                 show that: (a) the revised notion of graph pattern
                 matching allows us to identify communities commonly
                 found in real-life networks, and (b) the incremental
                 algorithms substantially outperform their batch
                 counterparts in response to small changes. These
                 suggest a promising framework for real-life graph
                 pattern matching.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Karvounarakis:2013:CDS,
  author =       "Grigoris Karvounarakis and Todd J. Green and Zachary
                 G. Ives and Val Tannen",
  title =        "Collaborative data sharing via update exchange and
                 provenance",
  journal =      j-TODS,
  volume =       "38",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2500127",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 30 16:33:21 MDT 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Recent work [Ives et al. 2005] proposed a new class of
                 systems for supporting data sharing among scientific
                 and other collaborations: this new collaborative data
                 sharing system connects heterogeneous logical peers
                 using a network of schema mappings. Each peer has a
                 locally controlled and edited database instance, but
                 wants to incorporate related data from other peers as
                 well. To achieve this, every peer's data and updates
                 propagate along the mappings to the other peers.
                 However, this operation, termed update exchange, is
                 filtered by trust conditions -expressing what data and
                 sources a peer judges to be authoritative-which may
                 cause a peer to reject another's updates. In order to
                 support such filtering, updates carry provenance
                 information. This article develops methods for
                 realizing such systems: we build upon techniques from
                 data integration, data exchange, incremental view
                 maintenance, and view update to propagate updates along
                 mappings, both to derived and optionally to source
                 instances. We incorporate a novel model for tracking
                 data provenance, such that curators may filter updates
                 based on trust conditions over this provenance. We
                 implement our techniques in a layer above an
                 off-the-shelf RDBMS, and we experimentally demonstrate
                 the viability of these techniques in the Orchestra
                 prototype system.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ozsoyoglu:2013:FIP,
  author =       "Z. Meral {\"O}zsoyoglu",
  title =        "Foreword to invited papers issue",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "20:1--20:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2539032.2539033",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Mozafari:2013:HPC,
  author =       "Barzan Mozafari and Kai Zeng and Loris D'antoni and
                 Carlo Zaniolo",
  title =        "High-performance complex event processing over
                 hierarchical data",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "21:1--21:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2536779",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "While Complex Event Processing (CEP) constitutes a
                 considerable portion of the so-called Big Data
                 analytics, current CEP systems can only process data
                 having a simple structure, and are otherwise limited in
                 their ability to efficiently support complex continuous
                 queries on structured or semistructured information.
                 However, XML-like streams represent a very popular form
                 of data exchange, comprising large portions of social
                 network and RSS feeds, financial feeds, configuration
                 files, and similar applications requiring advanced CEP
                 queries. In this article, we present the XSeq language
                 and system that support CEP on XML streams, via an
                 extension of XPath that is both powerful and amenable
                 to an efficient implementation. Specifically, the XSeq
                 language extends XPath with natural operators to
                 express sequential and Kleene-* patterns over XML
                 streams, while remaining highly amenable to efficient
                 execution. In fact, XSeq is designed to take full
                 advantage of the recently proposed Visibly Pushdown
                 Automata (VPA), where higher expressive power can be
                 achieved without compromising the computationally
                 attractive properties of finite state automata. Besides
                 the efficiency and expressivity benefits, the choice of
                 VPA as the underlying model also enables XSeq to go
                 beyond XML streams and be easily applicable to any data
                 with both sequential and hierarchical structures,
                 including JSON messages, RNA sequences, and software
                 traces. Therefore, we illustrate the XSeq's power for
                 CEP applications through examples from different
                 domains and provide formal results on its
                 expressiveness and complexity. Finally, we present
                 several optimization techniques for XSeq queries. Our
                 extensive experiments indicate that XSeq brings
                 outstanding performance to CEP applications: two orders
                 of magnitude improvement is obtained over the same
                 queries executed in general-purpose XML engines.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sarma:2013:CTL,
  author =       "Anish Das Sarma and Hongrae Lee and Hector Gonzalez
                 and Jayant Madhavan and Alon Halevy",
  title =        "Consistent thinning of large geographical data for map
                 visualization",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "22:1--22:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2539032.2539034",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Large-scale map visualization systems play an
                 increasingly important role in presenting geographic
                 datasets to end-users. Since these datasets can be
                 extremely large, a map rendering system often needs to
                 select a small fraction of the data to visualize them
                 in a limited space. This article addresses the
                 fundamental challenge of thinning: determining
                 appropriate samples of data to be shown on specific
                 geographical regions and zoom levels. Other than the
                 sheer scale of the data, the thinning problem is
                 challenging because of a number of other reasons: (1)
                 data can consist of complex geographical shapes, (2)
                 rendering of data needs to satisfy certain constraints,
                 such as data being preserved across zoom levels and
                 adjacent regions, and (3) after satisfying the
                 constraints, an optimal solution needs to be chosen
                 based on objectives such as maximality, fairness, and
                 importance of data. This article formally defines and
                 presents a complete solution to the thinning problem.
                 First, we express the problem as an integer programming
                 formulation that efficiently solves thinning for
                 desired objectives. Second, we present more efficient
                 solutions for maximality, based on DFS traversal of a
                 spatial tree. Third, we consider the common special
                 case of point datasets, and present an even more
                 efficient randomized algorithm. Fourth, we show that
                 contiguous regions are tractable for a general version
                 of maximality for which arbitrary regions are
                 intractable. Fifth, we examine the structure of our
                 integer programming formulation and show that for point
                 datasets, our program is integral. Finally, we have
                 implemented all techniques from this article in Google
                 Maps [Google 2005] visualizations of fusion tables
                 [Gonzalez et al. 2010], and we describe a set of
                 experiments that demonstrate the trade-offs among the
                 algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Teubner:2013:XFB,
  author =       "Jens Teubner and Louis Woods and Chongling Nie",
  title =        "{XLynx} --- An {FPGA}-based {XML} filter for hybrid
                 {XQuery} processing",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "23:1--23:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2536800",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "While offering unique performance and energy-saving
                 advantages, the use of Field-Programmable Gate Arrays
                 (FPGAs) for database acceleration has demanded major
                 concessions from system designers. Either the
                 programmable chips have been used for very basic
                 application tasks (such as implementing a rigid class
                 of selection predicates) or their circuit definition
                 had to be completely recompiled at runtime-a very
                 CPU-intensive and time-consuming effort. This work
                 eliminates the need for such concessions. As part of
                 our XLynx implementation-an FPGA-based XML filter-we
                 present skeleton automata, which is a design principle
                 for data-intensive hardware circuits that offers high
                 expressiveness and quick reconfiguration at the same
                 time. Skeleton automata provide a generic
                 implementation for a class of finite-state automata.
                 They can be parameterized to any particular automaton
                 instance in a matter of microseconds or less (as
                 opposed to minutes or hours for complete
                 recompilation). We showcase skeleton automata based on
                 XML projection [Marian and Sim{\'e}on 2003], a
                 filtering technique that illustrates the feasibility of
                 our strategy for a real-world and challenging task. By
                 performing XML projection in hardware and filtering
                 data in the network, we report on performance
                 improvements of several factors while remaining
                 nonintrusive to the back-end XML processor (we evaluate
                 XLynx using the Saxon engine).",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Losemann:2013:CRE,
  author =       "Katja Losemann and Wim Martens",
  title =        "The complexity of regular expressions and property
                 paths in {SPARQL}",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "24:1--24:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2494529",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The World Wide Web Consortium (W3C) recently
                 introduced property paths in SPARQL 1.1, a query
                 language for RDF data. Property paths allow SPARQL
                 queries to evaluate regular expressions over
                 graph-structured data. However, they differ from
                 standard regular expressions in several notable
                 aspects. For example, they have a limited form of
                 negation, they have numerical occurrence indicators as
                 syntactic sugar, and their semantics on graphs is
                 defined in a nonstandard manner. We formalize the W3C
                 semantics of property paths and investigate various
                 query evaluation problems on graphs. More specifically,
                 let $x$ and $y$ be two nodes in an edge-labeled graph
                 and $r$ be an expression. We study the complexities of:
                 (1) deciding whether there exists a path from $x$ to
                 $y$ that matches $r$ and (2) counting how many paths
                 from $x$ to $y$ match $r$. Our main results show that,
                 compared to an alternative semantics of regular
                 expressions on graphs, the complexity of (1) and (2)
                 under W3C semantics is significantly higher. Whereas
                 the alternative semantics remains in polynomial time
                 for large fragments of expressions, the W3C semantics
                 makes problems (1) and (2) intractable almost
                 immediately. As a side-result, we prove that the
                 membership problem for regular expressions with
                 numerical occurrence indicators and negation is in
                 polynomial time.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Letelier:2013:SAO,
  author =       "Andr{\'e}s Letelier and Jorge P{\'e}rez and Reinhard
                 Pichler and Sebastian Skritek",
  title =        "Static analysis and optimization of {Semantic Web}
                 queries",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "25:1--25:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2500130",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Static analysis is a fundamental task in query
                 optimization. In this article we study static analysis
                 and optimization techniques for SPARQL, which is the
                 standard language for querying Semantic Web data. Of
                 particular interest for us is the optionality feature
                 in SPARQL. It is crucial in Semantic Web data
                 management, where data sources are inherently
                 incomplete and the user is usually interested in
                 partial answers to queries. This feature is one of the
                 most complicated constructors in SPARQL and also the
                 one that makes this language depart from classical
                 query languages such as relational conjunctive queries.
                 We focus on the class of well-designed SPARQL queries,
                 which has been proposed in the literature as a fragment
                 of the language with good properties regarding query
                 evaluation. We first propose a tree representation for
                 SPARQL queries, called pattern trees, which captures
                 the class of well-designed SPARQL graph patterns. Among
                 other results, we propose several rules that can be
                 used to transform pattern trees into a simple normal
                 form, and study equivalence and containment. We also
                 study the evaluation and enumeration problems for this
                 class of queries.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Agarwal:2013:MS,
  author =       "Pankaj K. Agarwal and Graham Cormode and Zengfeng
                 Huang and Jeff M. Phillips and Zhewei Wei and Ke Yi",
  title =        "Mergeable summaries",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "26:1--26:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2500128",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the mergeability of data summaries.
                 Informally speaking, mergeability requires that, given
                 two summaries on two datasets, there is a way to merge
                 the two summaries into a single summary on the two
                 datasets combined together, while preserving the error
                 and size guarantees. This property means that the
                 summaries can be merged in a way akin to other
                 algebraic operators such as sum and max, which is
                 especially useful for computing summaries on massive
                 distributed data. Several data summaries are trivially
                 mergeable by construction, most notably all the
                 sketches that are linear functions of the datasets. But
                 some other fundamental ones, like those for heavy
                 hitters and quantiles, are not (known to be) mergeable.
                 In this article, we demonstrate that these summaries
                 are indeed mergeable or can be made mergeable after
                 appropriate modifications. Specifically, we show that
                 for $ \epsilon $-approximate heavy hitters, there is a
                 deterministic mergeable summary of size $ O (1 /
                 \epsilon) $; for $ \epsilon $-approximate quantiles,
                 there is a deterministic summary of size $ O((1 /
                 \epsilon) \log (\epsilon n)) $ that has a restricted
                 form of mergeability, and a randomized one of size $
                 O((1 / \epsilon) \log^{3 / 2} (1 / \epsilon)) $ with
                 full mergeability. We also extend our results to
                 geometric summaries such as \epsilon -approximations
                 which permit approximate multidimensional range
                 counting queries. While most of the results in this
                 article are theoretical in nature, some of the
                 algorithms are actually very simple and even perform
                 better than the previously best known algorithms, which
                 we demonstrate through experiments in a simulated
                 sensor network. We also achieve two results of
                 independent interest: (1) we provide the best known
                 randomized streaming bound for $ \epsilon $-approximate
                 quantiles that depends only on $ \epsilon $, of size $
                 O((1 / \epsilon) \log^{3 / 2} (1 / \epsilon)) $, and
                 (2) we demonstrate that the MG and the SpaceSaving
                 summaries for heavy hitters are isomorphic.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Konrad:2013:VXD,
  author =       "Christian Konrad and Fr{\'e}d{\'e}ric Magniez",
  title =        "Validating {XML} documents in the streaming model with
                 external memory",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "27:1--27:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2504590",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problem of validating XML documents of
                 size N against general DTDs in the context of streaming
                 algorithms. The starting point of this work is a
                 well-known space lower bound. There are XML documents
                 and DTDs for which p -pass streaming algorithms require
                 $ \Omega (N / p) $ space. We show that when allowing
                 access to external memory, there is a deterministic
                 streaming algorithm that solves this problem with
                 memory space $ O(\log^2 N) $, a constant number of
                 auxiliary read/write streams, and $ O(\log N) $ total
                 number of passes on the XML document and auxiliary
                 streams. An important intermediate step of this
                 algorithm is the computation of the
                 First-Child-Next-Sibling (FCNS) encoding of the initial
                 XML document in a streaming fashion. We study this
                 problem independently, and we also provide
                 memory-efficient streaming algorithms for decoding an
                 XML document given in its FCNS encoding. Furthermore,
                 validating XML documents encoding binary trees against
                 any DTD in the usual streaming model without external
                 memory can be done with sublinear memory. There is a
                 one-pass algorithm using $ O(\sqrt N \log N) $ space,
                 and a bidirectional two-pass algorithm using $ O(\log^2
                 N) $ space which perform this task.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cate:2013:LSM,
  author =       "Balder {Ten Cate} and V{\'\i}ctor Dalmau and Phokion
                 G. Kolaitis",
  title =        "Learning schema mappings",
  journal =      j-TODS,
  volume =       "38",
  number =       "4",
  pages =        "28:1--28:??",
  month =        nov,
  year =         "2013",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2539032.2539035",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Dec 9 11:35:10 MST 2013",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A schema mapping is a high-level specification of the
                 relationship between a source schema and a target
                 schema. Recently, a line of research has emerged that
                 aims at deriving schema mappings automatically or
                 semi-automatically with the help of data examples, that
                 is, pairs consisting of a source instance and a target
                 instance that depict, in some precise sense, the
                 intended behavior of the schema mapping. Several
                 different uses of data examples for deriving, refining,
                 or illustrating a schema mapping have already been
                 proposed and studied. In this article, we use the lens
                 of computational learning theory to systematically
                 investigate the problem of obtaining algorithmically a
                 schema mapping from data examples. Our aim is to
                 leverage the rich body of work on learning theory in
                 order to develop a framework for exploring the power
                 and the limitations of the various algorithmic methods
                 for obtaining schema mappings from data examples. We
                 focus on GAV schema mappings, that is, schema mappings
                 specified by GAV (Global-As-View) constraints. GAV
                 constraints are the most basic and the most widely
                 supported language for specifying schema mappings. We
                 present an efficient algorithm for learning GAV schema
                 mappings using Angluin's model of exact learning with
                 membership and equivalence queries. This is optimal,
                 since we show that neither membership queries nor
                 equivalence queries suffice, unless the source schema
                 consists of unary relations only. We also obtain
                 results concerning the learnability of schema mappings
                 in the context of Valiant's well-known PAC
                 (Probably-Approximately-Correct) learning model, and
                 concerning the learnability of restricted classes of
                 GAV schema mappings. Finally, as a byproduct of our
                 work, we show that there is no efficient algorithm for
                 approximating the shortest GAV schema mapping fitting a
                 given set of examples, unless the source schema
                 consists of unary relations only.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kostylev:2014:CAS,
  author =       "Egor V. Kostylev and Juan L. Reutter and Andr{\'a}s Z.
                 Salamon",
  title =        "Classification of annotation semirings over
                 containment of conjunctive queries",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "1:1--1:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2556524",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problem of query containment of
                 conjunctive queries over annotated databases.
                 Annotations are typically attached to tuples and
                 represent metadata, such as probability, multiplicity,
                 comments, or provenance. It is usually assumed that
                 annotations are drawn from a commutative semiring. Such
                 databases pose new challenges in query optimization,
                 since many related fundamental tasks, such as query
                 containment, have to be reconsidered in the presence of
                 propagation of annotations. We axiomatize several
                 classes of semirings for each of which containment of
                 conjunctive queries is equivalent to existence of a
                 particular type of homomorphism. For each of these
                 types, we also specify all semirings for which
                 existence of a corresponding homomorphism is a
                 sufficient (or necessary) condition for the
                 containment. We develop new decision procedures for
                 containment for some semirings which are not in any of
                 these classes. This generalizes and systematizes
                 previous approaches.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yi:2014:ISQ,
  author =       "Ke Yi and Lu Wang and Zhewei Wei",
  title =        "Indexing for summary queries: Theory and practice",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "2:1--2:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2508702",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Database queries can be broadly classified into two
                 categories: reporting queries and aggregation queries.
                 The former retrieves a collection of records from the
                 database that match the query's conditions, while the
                 latter returns an aggregate, such as count, sum,
                 average, or max (min), of a particular attribute of
                 these records. Aggregation queries are especially
                 useful in business intelligence and data analysis
                 applications where users are interested not in the
                 actual records, but some statistics of them. They can
                 also be executed much more efficiently than reporting
                 queries, by embedding properly precomputed aggregates
                 into an index. However, reporting and aggregation
                 queries provide only two extremes for exploring the
                 data. Data analysts often need more insight into the
                 data distribution than what those simple aggregates
                 provide, and yet certainly do not want the sheer volume
                 of data returned by reporting queries. In this article,
                 we design indexing techniques that allow for extracting
                 a statistical summary of all the records in the query.
                 The summaries we support include frequent items,
                 quantiles, and various sketches, all of which are of
                 central importance in massive data analysis. Our
                 indexes require linear space and extract a summary with
                 the optimal or near-optimal query cost. We illustrate
                 the efficiency and usefulness of our designs through
                 extensive experiments and a system demonstration.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kifer:2014:PFM,
  author =       "Daniel Kifer and Ashwin Machanavajjhala",
  title =        "{Pufferfish}: a framework for mathematical privacy
                 definitions",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "3:1--3:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2514689",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/cryptography2010.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we introduce a new and general
                 privacy framework called Pufferfish. The Pufferfish
                 framework can be used to create new privacy definitions
                 that are customized to the needs of a given
                 application. The goal of Pufferfish is to allow experts
                 in an application domain, who frequently do not have
                 expertise in privacy, to develop rigorous privacy
                 definitions for their data sharing needs. In addition
                 to this, the Pufferfish framework can also be used to
                 study existing privacy definitions. We illustrate the
                 benefits with several applications of this privacy
                 framework: we use it to analyze differential privacy
                 and formalize a connection to attackers who believe
                 that the data records are independent; we use it to
                 create a privacy definition called hedging privacy,
                 which can be used to rule out attackers whose prior
                 beliefs are inconsistent with the data; we use the
                 framework to define and study the notion of composition
                 in a broader context than before; we show how to apply
                 the framework to protect unbounded continuous
                 attributes and aggregate information; and we show how
                 to use the framework to rigorously account for prior
                 data releases.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ma:2014:SSC,
  author =       "Shuai Ma and Yang Cao and Wenfei Fan and Jinpeng Huai
                 and Tianyu Wo",
  title =        "Strong simulation: Capturing topology in graph pattern
                 matching",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "4:1--4:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2528937",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Graph pattern matching is finding all matches in a
                 data graph for a given pattern graph and is often
                 defined in terms of subgraph isomorphism, an
                 NP-complete problem. To lower its complexity, various
                 extensions of graph simulation have been considered
                 instead. These extensions allow graph pattern matching
                 to be conducted in cubic time. However, they fall short
                 of capturing the topology of data graphs, that is,
                 graphs may have a structure drastically different from
                 pattern graphs they match, and the matches found are
                 often too large to understand and analyze. To rectify
                 these problems, this article proposes a notion of
                 strong simulation, a revision of graph simulation for
                 graph pattern matching. (1) We identify a set of
                 criteria for preserving the topology of graphs matched.
                 We show that strong simulation preserves the topology
                 of data graphs and finds a bounded number of matches.
                 (2) We show that strong simulation retains the same
                 complexity as earlier extensions of graph simulation by
                 providing a cubic-time algorithm for computing strong
                 simulation. (3) We present the locality property of
                 strong simulation which allows us to develop an
                 effective distributed algorithm to conduct graph
                 pattern matching on distributed graphs. (4) We
                 experimentally verify the effectiveness and efficiency
                 of these algorithms using both real-life and synthetic
                 data.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gatterbauer:2014:OBP,
  author =       "Wolfgang Gatterbauer and Dan Suciu",
  title =        "Oblivious bounds on the probability of {Boolean}
                 functions",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "5:1--5:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2532641",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article develops upper and lower bounds for the
                 probability of Boolean functions by treating multiple
                 occurrences of variables as independent and assigning
                 them new individual probabilities. We call this
                 approach dissociation and give an exact
                 characterization of optimal oblivious bounds, that is,
                 when the new probabilities are chosen independently of
                 the probabilities of all other variables. Our
                 motivation comes from the weighted model counting
                 problem (or, equivalently, the problem of computing the
                 probability of a Boolean function), which is \#P-hard
                 in general. By performing several dissociations, one
                 can transform a Boolean formula whose probability is
                 difficult to compute into one whose probability is easy
                 to compute, and which is guaranteed to provide an upper
                 or lower bound on the probability of the original
                 formula by choosing appropriate probabilities for the
                 dissociated variables. Our new bounds shed light on the
                 connection between previous relaxation-based and
                 model-based approximations and unify them as concrete
                 choices in a larger design space. We also show how our
                 theory allows a standard relational database management
                 system (DBMS) to both upper and lower bound hard
                 probabilistic queries in guaranteed polynomial time.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fang:2014:MOP,
  author =       "Qiong Fang and Wilfred Ng and Jianlin Feng and Yuliang
                 Li",
  title =        "Mining order-preserving submatrices from probabilistic
                 matrices",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "6:1--6:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2533712",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Order-preserving submatrices (OPSMs) capture consensus
                 trends over columns shared by rows in a data matrix.
                 Mining OPSM patterns discovers important and
                 interesting local correlations in many real
                 applications, such as those involving biological data
                 or sensor data. The prevalence of uncertain data in
                 various applications, however, poses new challenges for
                 OPSM mining, since data uncertainty must be
                 incorporated into OPSM modeling and the algorithmic
                 aspects. In this article, we define new probabilistic
                 matrix representations to model uncertain data with
                 continuous distributions. A novel probabilistic
                 order-preserving submatrix (POPSM) model is formalized
                 in order to capture similar local correlations in
                 probabilistic matrices. The POPSM model adopts a new
                 probabilistic support measure that evaluates the extent
                 to which a row belongs to a POPSM pattern. Due to the
                 intrinsic high computational complexity of the POPSM
                 mining problem, we utilize the anti-monotonic property
                 of the probabilistic support measure and propose an
                 efficient Apriori-based mining framework called
                 ProbApri to mine POPSM patterns. The framework consists
                 of two mining methods, UniApri and NormApri, which are
                 developed for mining POPSM patterns, respectively, from
                 two representative types of probabilistic matrices, the
                 UniDist matrix (assuming uniform data distributions)
                 and the NormDist matrix (assuming normal data
                 distributions). We show that the NormApri method is
                 practical enough for mining POPSM patterns from
                 probabilistic matrices that model more general data
                 distributions. We demonstrate the superiority of our
                 approach by two applications. First, we use two
                 biological datasets to illustrate that the POPSM model
                 better captures the characteristics of the expression
                 levels of biologically correlated genes and greatly
                 promotes the discovery of patterns with high biological
                 significance. Our result is significantly better than
                 the counterpart OPSMRM (OPSM with repeated measurement)
                 model which adopts a set-valued matrix representation
                 to capture data uncertainty. Second, we run the
                 experiments on an RFID trace dataset and show that our
                 POPSM model is effective and efficient in capturing the
                 common visiting subroutes among users.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wang:2014:ESS,
  author =       "Jiannan Wang and Guoliang Li and Jianhua Feng",
  title =        "Extending string similarity join to tolerant fuzzy
                 token matching",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "7:1--7:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2535628",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "String similarity join that finds similar string pairs
                 between two string sets is an essential operation in
                 many applications and has attracted significant
                 attention recently in the database community. A
                 significant challenge in similarity join is to
                 implement an effective fuzzy match operation to find
                 all similar string pairs which may not match exactly.
                 In this article, we propose a new similarity function,
                 called fuzzy-token-matching-based similarity which
                 extends token-based similarity functions (e.g., jaccard
                 similarity and cosine similarity) by allowing fuzzy
                 match between two tokens. We study the problem of
                 similarity join using this new similarity function and
                 present a signature-based method to address this
                 problem. We propose new signature schemes and develop
                 effective pruning techniques to improve the
                 performance. We also extend our techniques to support
                 weighted tokens. Experimental results show that our
                 method achieves high efficiency and result quality and
                 significantly outperforms state-of-the-art
                 approaches.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Sen:2014:DRM,
  author =       "Siddhartha Sen and Robert E. Tarjan",
  title =        "Deletion without rebalancing in multiway search
                 trees",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "8:1--8:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2540068",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Some database systems that use a form of B-tree for
                 the underlying data structure do not do rebalancing on
                 deletion. This means that a bad sequence of deletions
                 can create a very unbalanced tree. Yet such databases
                 perform well in practice. Avoidance of rebalancing on
                 deletion has been justified empirically and by
                 average-case analysis, but to our knowledge, no
                 worst-case analysis has been done. We do such an
                 analysis. We show that the tree height remains
                 logarithmic in the number of insertions, independent of
                 the number of deletions. Furthermore, the amortized
                 time for an insertion or deletion, excluding the search
                 time, is $ O(1) $, and nodes are modified by insertions
                 and deletions with a frequency that is exponentially
                 small in their height. The latter results do not hold
                 for standard B-trees. By adding periodic rebuilding of
                 the tree, we obtain a data structure that is
                 theoretically superior to standard B-trees in many
                 ways. Our results suggest that rebalancing on deletion
                 not only is unnecessary but may be harmful.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nekrich:2014:ERS,
  author =       "Yakov Nekrich",
  title =        "Efficient range searching for categorical and plain
                 data",
  journal =      j-TODS,
  volume =       "39",
  number =       "1",
  pages =        "9:1--9:??",
  month =        jan,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2543924",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Feb 5 11:31:16 MST 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In the orthogonal range-searching problem, we store a
                 set of input points $S$ in a data structure; the answer
                 to a query $Q$ is a piece of information about points
                 in $ Q \cap S $, for example, the list of all points in
                 $ Q \cap S $ or the number of points in $Q$. In the
                 colored (or categorical) range-searching problem, the
                 set of input points is partitioned into categories; the
                 answer to a query is a piece of information about
                 categories of points in a query range. In this article,
                 we describe several new results for one- and
                 two-dimensional range-searching problems. We obtain an
                 optimal adaptive data structure for counting the number
                 of objects in a three-sided range and for counting
                 categories of objects in a one-dimensional range. We
                 also obtain new results on color range reporting in two
                 dimensions, approximate color counting in one
                 dimension, and some other related problems.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Xu:2014:OCP,
  author =       "Pan Xu and Srikanta Tirthapura",
  title =        "Optimality of Clustering Properties of Space-Filling
                 Curves",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "10:1--10:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2556686",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Space-filling curves have been used in the design of
                 data structures for multidimensional data for many
                 decades. A fundamental quality metric of a
                 space-filling curve is its ``clustering number'' with
                 respect to a class of queries, which is the average
                 number of contiguous segments on the space-filling
                 curve that a query region can be partitioned into. We
                 present a characterization of the clustering number of
                 a general class of space-filling curves, as well as the
                 first nontrivial lower bounds on the clustering number
                 for any space-filling curve. Our results answer
                 questions that have been open for more than 15 years.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Thomson:2014:FDT,
  author =       "Alexander Thomson and Thaddeus Diamond and Shu-Chun
                 Weng and Kun Ren and Philip Shao and Daniel J. Abadi",
  title =        "Fast Distributed Transactions and Strongly Consistent
                 Replication for {OLTP} Database Systems",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "11:1--11:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2556685",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "As more data management software is designed for
                 deployment in public and private clouds, or on a
                 cluster of commodity servers, new distributed storage
                 systems increasingly achieve high data access
                 throughput via partitioning and replication. In order
                 to achieve high scalability, however, today's systems
                 generally reduce transactional support, disallowing
                 single transactions from spanning multiple partitions.
                 This article describes Calvin, a practical transaction
                 scheduling and data replication layer that uses a
                 deterministic ordering guarantee to significantly
                 reduce the normally prohibitive contention costs
                 associated with distributed transactions. This allows
                 near-linear scalability on a cluster of commodity
                 machines, without eliminating traditional transactional
                 guarantees, introducing a single point of failure, or
                 requiring application developers to reason about data
                 partitioning. By replicating transaction inputs instead
                 of transactional actions, Calvin is able to support
                 multiple consistency levels-including Paxos-based
                 strong consistency across geographically distant
                 replicas-at no cost to transactional throughput.
                 Furthermore, Calvin introduces a set of tools that will
                 allow application developers to gain the full
                 performance benefit of Calvin's server-side transaction
                 scheduling mechanisms without introducing the
                 additional code complexity and inconvenience normally
                 associated with using DBMS stored procedures in place
                 of ad hoc client-side transactions.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nykiel:2014:SAM,
  author =       "Tomasz Nykiel and Michalis Potamias and Chaitanya
                 Mishra and George Kollios and Nick Koudas",
  title =        "Sharing across Multiple {MapReduce} Jobs",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "12:1--12:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2560796",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Large-scale data analysis lies in the core of modern
                 enterprises and scientific research. With the emergence
                 of cloud computing, the use of an analytical query
                 processing infrastructure can be directly associated
                 with monetary cost. MapReduce has been a popular
                 framework in the context of cloud computing, designed
                 to serve long-running queries (jobs) which can be
                 processed in batch mode. Taking into account that
                 different jobs often perform similar work, there are
                 many opportunities for sharing. In principle, sharing
                 similar work reduces the overall amount of work, which
                 can lead to reducing monetary charges for utilizing the
                 processing infrastructure. In this article we present a
                 sharing framework tailored to MapReduce, namely, {\tt
                 MRShare}. Our framework, {\tt MRShare}, transforms a
                 batch of queries into a new batch that will be executed
                 more efficiently, by merging jobs into groups and
                 evaluating each group as a single query. Based on our
                 cost model for MapReduce, we define an optimization
                 problem and we provide a solution that derives the
                 optimal grouping of queries. Given the query grouping,
                 we merge jobs appropriately and submit them to
                 MapReduce for processing. A key property of {\tt
                 MRShare} is that it is independent of the MapReduce
                 implementation. Experiments with our prototype, built
                 on top of Hadoop, demonstrate the overall effectiveness
                 of our approach. {\tt MRShare} is primarily designed
                 for handling I/O-intensive queries. However, with the
                 development of high-level languages operating on top of
                 MapReduce, user queries executed in this model become
                 more complex and CPU intensive. Commonly, executed
                 queries can be modeled as evaluating pipelines of
                 CPU-expensive filters over the input stream. Examples
                 of such filters include, but are not limited to, index
                 probes, or certain types of joins. In this article we
                 adapt some of the standard techniques for filter
                 ordering used in relational and stream databases,
                 propose their extensions, and implement them through
                 {\tt MRAdaptiveFilter}, an extension of {\tt MRShare}
                 for expensive filter ordering tailored to MapReduce,
                 which allows one to handle both single- and batch-query
                 execution modes. We present an experimental evaluation
                 that demonstrates additional benefits of {\tt
                 MRAdaptiveFilter}, when executing CPU-intensive queries
                 in {\tt MRShare}.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lu:2014:EAC,
  author =       "Ying Lu and Jiaheng Lu and Gao Cong and Wei Wu and
                 Cyrus Shahabi",
  title =        "Efficient Algorithms and Cost Models for Reverse
                 Spatial-Keyword $k$-Nearest Neighbor Search",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "13:1--13:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2576232",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Geographic objects associated with descriptive texts
                 are becoming prevalent, justifying the need for
                 spatial-keyword queries that consider both locations
                 and textual descriptions of the objects. Specifically,
                 the relevance of an object to a query is measured by
                 spatial-textual similarity that is based on both
                 spatial proximity and textual similarity. In this
                 article, we introduce the Reverse Spatial-Keyword
                 $k$-Nearest Neighbor (RSK $k$ NN) query, which finds
                 those objects that have the query as one of their $k$
                 nearest spatial-textual objects. The RSK $k$ NN queries
                 have numerous applications in online maps and GIS
                 decision support systems. To answer RSK $k$ NN queries
                 efficiently, we propose a hybrid index tree, called
                 IUR-tree (Intersection-Union R-tree) that effectively
                 combines location proximity with textual similarity.
                 Subsequently, we design a branch-and-bound search
                 algorithm based on the IUR-tree. To accelerate the
                 query processing, we improve IUR-tree by leveraging the
                 distribution of textual description, leading to some
                 variants of the IUR-tree called Clustered IUR-tree
                 (CIUR-tree) and combined clustered IUR-tree (C$^2$
                 IUR-tree), for each of which we develop optimized
                 algorithms. We also provide a theoretical cost model to
                 analyze the efficiency of our algorithms. Our empirical
                 studies show that the proposed algorithms are efficient
                 and scalable.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bartolini:2014:DPW,
  author =       "Ilaria Bartolini and Paolo Ciaccia and Marco Patella",
  title =        "Domination in the Probabilistic World: Computing
                 Skylines for Arbitrary Correlations and Ranking
                 Semantics",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "14:1--14:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2602135",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In a probabilistic database, deciding if a tuple $u$
                 is better than another tuple $v$ has not a univocal
                 solution, rather it depends on the specific
                 Probabilistic Ranking Semantics (PRS) one wants to
                 adopt so as to combine together tuples' scores and
                 probabilities. In deterministic databases it is known
                 that skyline queries are a remarkable alternative to
                 (top-$k$) ranking queries, because they remove from the
                 user the burden of specifying a scoring function that
                 combines values of different attributes into a single
                 score. The skyline of a deterministic relation $R$ is
                 the set of undominated tuples in $R$ --- tuple $u$
                 dominates tuple $v$ iff on all the attributes of
                 interest $u$ is better than or equal to $v$ and
                 strictly better on at least one attribute. Domination
                 is equivalent to having $ s(u) >= s(v)$ for all
                 monotone scoring functions $ s()$. The skyline of a
                 probabilistic relation $ R^p$ can be similarly defined
                 as the set of $P$-undominated tuples in $ R^p$, where
                 now $u$ $P$-dominates $v$ iff, whatever monotone
                 scoring function one would use to combine the skyline
                 attributes, $u$ is reputed better than $v$ by the PRS
                 at hand. This definition, which is applicable to
                 arbitrary ranking semantics and probabilistic
                 correlation models, is parametric in the adopted PRS,
                 thus it ensures that ranking and skyline queries will
                 always return consistent results. In this article we
                 provide an overall view of the problem of computing the
                 skyline of a probabilistic relation. We show how, under
                 mild conditions that indeed hold for all known PRSs,
                 checking $P$-domination can be cast into an
                 optimization problem, whose complexity we characterize
                 for a variety of combinations of ranking semantics and
                 correlation models. For each analyzed case we also
                 provide specific $P$-domination rules, which are
                 exploited by the algorithm we detail for the case where
                 the probabilistic model is known to the query
                 processor. We also consider the case in which the
                 probability of tuple events can only be obtained
                 through an oracle, and describe another skyline
                 algorithm for this loosely integrated scenario. Our
                 experimental evaluation of $P$-domination rules and
                 skyline algorithms confirms the theoretical analysis.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Deng:2014:CQR,
  author =       "Ting Deng and Wenfei Fan",
  title =        "On the Complexity of Query Result Diversification",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "15:1--15:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2602136",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Query result diversification is a bi-criteria
                 optimization problem for ranking query results. Given a
                 database $D$, a query $Q$, and a positive integer k, it
                 is to find a set of $k$ tuples from $ Q(D)$ such that
                 the tuples are as relevant as possible to the query,
                 and at the same time, as diverse as possible to each
                 other. Subsets of $ Q(D)$ are ranked by an objective
                 function defined in terms of relevance and diversity.
                 Query result diversification has found a variety of
                 applications in databases, information retrieval, and
                 operations research. This article investigates the
                 complexity of result diversification for relational
                 queries. (1) We identify three problems in connection
                 with query result diversification, to determine whether
                 there exists a set of $k$ tuples that is ranked above a
                 bound with respect to relevance and diversity, to
                 assess the rank of a given $k$-element set, and to
                 count how many $k$-element sets are ranked above a
                 given bound based on an objective function. (2) We
                 study these problems for a variety of query languages
                 and for the three objective functions proposed in
                 Gollapudi and Sharma [2009]. We establish the upper and
                 lower bounds of these problems, all matching, for both
                 combined complexity and data complexity. (3) We also
                 investigate several special settings of these problems,
                 identifying tractable cases. Moreover, (4) we
                 reinvestigate these problems in the presence of
                 compatibility constraints commonly found in practice,
                 and provide their complexity in all these settings.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Giatrakos:2014:DGQ,
  author =       "Nikos Giatrakos and Antonios Deligiannakis and Minos
                 Garofalakis and Izchak Sharfman and Assaf Schuster",
  title =        "Distributed Geometric Query Monitoring Using
                 Prediction Models",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "16:1--16:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2602137",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many modern streaming applications, such as online
                 analysis of financial, network, sensor, and other forms
                 of data, are inherently distributed in nature. An
                 important query type that is the focal point in such
                 application scenarios regards actuation queries, where
                 proper action is dictated based on a trigger condition
                 placed upon the current value that a monitored function
                 receives. Recent work [Sharfman et al. 2006, 2007b,
                 2008] studies the problem of (nonlinear) sophisticated
                 function tracking in a distributive manner. The main
                 concept behind the geometric monitoring approach
                 proposed there is for each distributed site to perform
                 the function monitoring over an appropriate subset of
                 the input domain. In the current work, we examine
                 whether the distributed monitoring mechanism can become
                 more efficient, in terms of the number of communicated
                 messages, by extending the geometric monitoring
                 framework to utilize prediction models. We initially
                 describe a number of local estimators (predictors) that
                 are useful for the applications that we consider and
                 which have already been shown particularly useful in
                 past work. We then demonstrate the feasibility of
                 incorporating predictors in the geometric monitoring
                 framework and show that prediction-based geometric
                 monitoring in fact generalizes the original geometric
                 monitoring framework. We propose a large variety of
                 different prediction-based monitoring models for the
                 distributed threshold monitoring of complex functions.
                 Our extensive experimentation with a variety of real
                 datasets, functions, and parameter settings indicates
                 that our approaches can provide significant
                 communication savings ranging between two times and up
                 to three orders of magnitude, compared to the
                 transmission cost of the original monitoring
                 framework.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lei:2014:RDQ,
  author =       "Chuan Lei and Elke A. Rundensteiner",
  title =        "Robust Distributed Query Processing for Streaming
                 Data",
  journal =      j-TODS,
  volume =       "39",
  number =       "2",
  pages =        "17:1--17:??",
  month =        may,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2602138",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed May 21 18:53:24 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Distributed stream processing systems must function
                 efficiently for data streams that fluctuate in their
                 arrival rates and data distributions. Yet repeated and
                 prohibitively expensive load reallocation across
                 machines may make these systems ineffective,
                 potentially resulting in data loss or even system
                 failure. To overcome this problem, we propose a
                 comprehensive solution, called the Robust Load
                 Distribution (RLD) strategy, that is resilient under
                 data fluctuations. RLD provides $ \epsilon $-optimal
                 query performance under an expected range of load
                 fluctuations without suffering from the performance
                 penalty caused by load migration. RLD is based on three
                 key strategies. First, we model robust distributed
                 stream processing as a parametric query optimization
                 problem in a parameter space that captures the stream
                 fluctuations. The notions of both robust logical and
                 robust physical plans that work together to proactively
                 handle all ranges of expected fluctuations in
                 parameters are abstracted as overlays of this parameter
                 space. Second, our Early-terminated Robust Partitioning
                 (ERP) finds a combination of robust logical plans
                 that together cover the parameter space, while
                 minimizing the number of prohibitively expensive
                 optimizer calls with a probabilistic bound on the space
                 coverage. Third, we design a family of algorithms for
                 physical plan generation. Our GreedyPhy exploits a
                 probabilistic model to efficiently find a robust
                 physical plan that sustains most frequently used robust
                 logical plans at runtime. Our CorPhy algorithm exploits
                 operator correlations for the robust physical plan
                 optimization. The resulting physical plan smooths the
                 workload on each node under all expected fluctuations.
                 Our OptPrune algorithm, using CorPhy as baseline, is
                 guaranteed to find the optimal physical plan that
                 maximizes the parameter space coverage with a practical
                 increase in optimization time. Lastly, we further
                 expand the capabilities of our proposed RLD framework
                 to also appropriately react under so-called ``space
                 drifts'', that is, a space drift is a change of the
                 parameter space where the observed runtime statistics
                 deviate from the expected optimization-time statistics.
                 Our RLD solution is capable of adjusting itself to the
                 unexpected yet significant data fluctuations beyond
                 those planned for via covering the parameter space. Our
                 experimental study using stock market and sensor
                 network streams demonstrates that our RLD methodology
                 consistently outperforms state-of-the-art solutions in
                 terms of efficiency and effectiveness in highly
                 fluctuating data stream environments.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2014:E,
  author =       "Christian S. Jensen",
  title =        "Editorial",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "18:1--18:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2662448",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhang:2014:TPI,
  author =       "Rui Zhang and Jianzhong Qi and Martin Stradling and
                 Jin Huang",
  title =        "Towards a Painless Index for Spatial Objects",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "19:1--19:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2629333",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Conventional spatial indexes, represented by the
                 R-tree, employ multidimensional tree structures that
                 are complicated and require enormous efforts to
                 implement in a full-fledged database management system
                 (DBMS). An alternative approach for supporting spatial
                 queries is mapping-based indexing, which maps both data
                 and queries into a one-dimensional space such that data
                 can be indexed and queries can be processed through a
                 one-dimensional indexing structure such as the B$^+$.
                 Mapping-based indexing requires implementing only a few
                 mapping functions, incurring much less effort in
                 implementation compared to conventional spatial index
                 structures. Yet, a major concern about using
                 mapping-based indexes is their lower efficiency than
                 conventional tree structures. In this article, we
                 propose a mapping-based spatial indexing scheme called
                 Size Separation Indexing (SSI). SSI is equipped with a
                 suite of techniques including size separation, data
                 distribution transformation, and more efficient mapping
                 algorithms. These techniques overcome the drawbacks of
                 existing mapping-based indexes and significantly
                 improve the efficiency of query processing. We show
                 through extensive experiments that, for window queries
                 on spatial objects with nonzero extents, SSI has two
                 orders of magnitude better performance than existing
                 mapping-based indexes and competitive performance to
                 the R-tree as a standalone implementation. We have also
                 implemented SSI on top of two off-the-shelf DBMSs,
                 PostgreSQL and a commercial platform, both having
                 R-tree implementation. In this case, SSI is up to two
                 orders of magnitude faster than their provided spatial
                 indexes. Therefore, we achieve a spatial index more
                 efficient than the R-tree in a DBMS implementation that
                 is at the same time easy to implement. This result may
                 upset a common perception that has existed for a long
                 time in this area that the R-tree is the best choice
                 for indexing spatial objects.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tatti:2014:FRI,
  author =       "Nikolaj Tatti and Fabian Moerchen and Toon Calders",
  title =        "Finding Robust Itemsets under Subsampling",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "20:1--20:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2656261",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Mining frequent patterns is plagued by the problem of
                 pattern explosion, making pattern reduction techniques
                 a key challenge in pattern mining. In this article we
                 propose a novel theoretical framework for pattern
                 reduction by measuring the robustness of a property of
                 an itemset such as closedness or nonderivability. The
                 robustness of a property is the probability that this
                 property holds on random subsets of the original data.
                 We study four properties, namely an itemset being
                 closed, free, non-derivable, or totally shattered, and
                 demonstrate how to compute the robustness analytically
                 without actually sampling the data. Our concept of
                 robustness has many advantages: Unlike statistical
                 approaches for reducing patterns, we do not assume a
                 null hypothesis or any noise model and, in contrast to
                 noise-tolerant or approximate patterns, the robust
                 patterns for a given property are always a subset of
                 the patterns with this property. If the underlying
                 property is monotonic then the measure is also
                 monotonic, allowing us to efficiently mine robust
                 itemsets. We further derive a parameter-free technique
                 for ranking itemsets that can be used for top- k
                 approaches. Our experiments demonstrate that we can
                 successfully use the robustness measure to reduce the
                 number of patterns and that ranking yields interesting
                 itemsets.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Choi:2014:MRS,
  author =       "Dong-Wan Choi and Chin-Wan Chung and Yufei Tao",
  title =        "Maximizing Range Sum in External Memory",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "21:1--21:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2629477",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article studies the MaxRS problem in spatial
                 databases. Given a set O of weighted points and a
                 rectangle r of a given size, the goal of the MaxRS
                 problem is to find a location of r such that the sum of
                 the weights of all the points covered by r is
                 maximized. This problem is useful in many
                 location-based services such as finding the best place
                 for a new franchise store with a limited delivery range
                 and finding the hotspot with the largest number of
                 nearby attractions for a tourist with a limited
                 reachable range. However, the problem has been studied
                 mainly in the theoretical perspective, particularly in
                 computational geometry. The existing algorithms from
                 the computational geometry community are in-memory
                 algorithms that do not guarantee the scalability. In
                 this article, we propose a scalable external-memory
                 algorithm ( ExactMaxRS ) for the MaxRS problem that is
                 optimal in terms of the I/O complexity. In addition, we
                 propose an approximation algorithm ( ApproxMaxCRS ) for
                 the MaxCRS problem that is a circle version of the
                 MaxRS problem. We prove the correctness and optimality
                 of the ExactMaxRS algorithm along with the
                 approximation bound of the ApproxMaxCRS algorithm.
                 Furthermore, motivated by the fact that all the
                 existing solutions simply assume that there is no tied
                 area for the best location, we extend the MaxRS problem
                 to a more fundamental problem, namely AllMaxRS, so that
                 all the locations with the same best score can be
                 retrieved. We first prove that the AllMaxRS problem
                 cannot be trivially solved by applying the techniques
                 for the MaxRS problem. Then we propose an
                 output-sensitive external-memory algorithm (
                 TwoPhaseMaxRS ) that gives the exact solution for the
                 AllMaxRS problem through two phases. Also, we prove
                 both the soundness and completeness of the result
                 returned from TwoPhaseMaxRS. From extensive
                 experimental results, we show that ExactMaxRS and
                 ApproxMaxCRS are several orders of magnitude faster
                 than methods adapted from existing algorithms, the
                 approximation bound in practice is much better than the
                 theoretical bound of ApproxMaxCRS, and TwoPhaseMaxRS is
                 not only much faster but also more robust than the
                 straightforward extension of ExactMaxRS.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Karwa:2014:PAG,
  author =       "Vishesh Karwa and Sofya Raskhodnikova and Adam Smith
                 and Grigory Yaroslavtsev",
  title =        "Private Analysis of Graph Structure",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "22:1--22:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2611523",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We present efficient algorithms for releasing useful
                 statistics about graph data while providing rigorous
                 privacy guarantees. Our algorithms work on datasets
                 that consist of relationships between individuals, such
                 as social ties or email communication. The algorithms
                 satisfy edge differential privacy, which essentially
                 requires that the presence or absence of any particular
                 relationship be hidden. Our algorithms output
                 approximate answers to subgraph counting queries. Given
                 a query graph H, for example, a triangle, $k$-star, or
                 $k$ triangle, the goal is to return the number of
                 edge-induced isomorphic copies of $H$ in the input
                 graph. The special case of triangles was considered by
                 Nissim et al. [2007] and a more general investigation
                 of arbitrary query graphs was initiated by Rastogi et
                 al. [2009]. We extend the approach of Nissim et al. to
                 a new class of statistics, namely $k$-star queries. We
                 also give algorithms for $k$-triangle queries using a
                 different approach based on the higher-order local
                 sensitivity. For the specific graph statistics we
                 consider (i.e., $k$-stars and $k$-triangles), we
                 significantly improve on the work of Rastogi et al.:
                 our algorithms satisfy a stronger notion of privacy
                 that does not rely on the adversary having a particular
                 prior distribution on the data, and add less noise to
                 the answers before releasing them. We evaluate the
                 accuracy of our algorithms both theoretically and
                 empirically, using a variety of real and synthetic
                 datasets. We give explicit, simple conditions under
                 which these algorithms add a small amount of noise. We
                 also provide the average-case analysis in the
                 Erd{\H{o}}s--R{\'e}nyi-Gilbert $ G(n, p)$ random graph
                 model. Finally, we give hardness results indicating
                 that the approach Nissim et al. used for triangles
                 cannot easily be extended to $k$-triangles (hence
                 justifying our development of a new algorithmic
                 approach).",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pang:2014:PPA,
  author =       "Hweehwa Pang and Xuhua Ding",
  title =        "Privacy-Preserving Ad-Hoc Equi-Join on Outsourced
                 Data",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "23:1--23:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2629501",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In IT outsourcing, a user may delegate the data
                 storage and query processing functions to a third-party
                 server that is not completely trusted. This gives rise
                 to the need to safeguard the privacy of the database as
                 well as the user queries over it. In this article, we
                 address the problem of running ad hoc equi-join queries
                 directly on encrypted data in such a setting. Our
                 contribution is the first solution that achieves
                 constant complexity per pair of records that are
                 evaluated for the join. After formalizing the privacy
                 requirements pertaining to the database and user
                 queries, we introduce a cryptographic construct for
                 securely joining records across relations. The
                 construct protects the database with a strong
                 encryption scheme. Moreover, information disclosure
                 after executing an equi-join is kept to the
                 minimum-that two input records combine to form an
                 output record if and only if they share common join
                 attribute values. There is no disclosure on records
                 that are not part of the join result. Building on this
                 construct, we then present join algorithms that
                 optimize the join execution by eliminating the need to
                 match every record pair from the input relations. We
                 provide a detailed analysis of the cost of the
                 algorithms and confirm the analysis through extensive
                 experiments with both synthetic and benchmark
                 workloads. Through this evaluation, we tease out useful
                 insights on how to configure the join algorithms to
                 deliver acceptable execution time in practice.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Malvestuto:2014:JLO,
  author =       "Francesco M. Malvestuto",
  title =        "A Join-Like Operator to Combine Data Cubes and Answer
                 Queries from Multiple Data Cubes",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "24:1--24:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2638545",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In order to answer a ``joint'' query from multiple
                 data cubes, Pourabass and Shoshani [2007] distinguish
                 the data cube on the measure of interest (called the
                 ``primary'' data cube) from the other data cubes
                 (called ``proxy'' data cubes) that are used to involve
                 the dimensions (in the query) not in the primary data
                 cube. They demonstrate in study cases that, if the
                 measures of the primary and proxy data cubes are
                 correlated, then the answer to a joint query is an
                 accurate estimate of its true value. Needless to say,
                 for two or more proxy data cubes, the result depends
                 upon the way the primary and proxy data cubes are
                 combined together; however, for certain combination
                 schemes Pourabass and Shoshani provide a sufficient
                 condition, that they call proxy noncommonality, for the
                 invariance of the result. In this article, we
                 introduce: (1) a merge operator combining the contents
                 of a primary data cube with the contents of a proxy
                 data cube, (2) merge expressions for general
                 combination schemes, and (3) an equivalence relation
                 between merge expressions having the same pattern.
                 Then, we prove that proxy noncommonality characterizes
                 patterns for which every two merge expressions are
                 equivalent. Moreover, we provide an efficient procedure
                 for answering joint queries in the special case of
                 perfect merge expressions. Finally, we show that our
                 results apply to data cubes in which measures are
                 obtained from unaggregated data using the aggregate
                 functions SUM, COUNT, MAX, and MIN, and a lot more.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gottlob:2014:QRO,
  author =       "Georg Gottlob and Giorgio Orsi and Andreas Pieris",
  title =        "Query Rewriting and Optimization for Ontological
                 Databases",
  journal =      j-TODS,
  volume =       "39",
  number =       "3",
  pages =        "25:1--25:??",
  month =        sep,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2638546",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 7 18:54:33 MDT 2014",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Ontological queries are evaluated against a knowledge
                 base consisting of an extensional database and an
                 ontology (i.e., a set of logical assertions and
                 constraints that derive new intensional knowledge from
                 the extensional database), rather than directly on the
                 extensional database. The evaluation and optimization
                 of such queries is an intriguing new problem for
                 database research. In this article, we discuss two
                 important aspects of this problem: query rewriting and
                 query optimization. Query rewriting consists of the
                 compilation of an ontological query into an equivalent
                 first-order query against the underlying extensional
                 database. We present a novel query rewriting algorithm
                 for rather general types of ontological constraints
                 that is well suited for practical implementations. In
                 particular, we show how a conjunctive query against a
                 knowledge base, expressed using linear and sticky
                 existential rules, that is, members of the recently
                 introduced Datalog\pm{} family of ontology languages,
                 can be compiled into a union of conjunctive queries
                 (UCQ) against the underlying database. Ontological
                 query optimization, in this context, attempts to
                 improve this rewriting process soas to produce possibly
                 small and cost-effective UCQ rewritings for an input
                 query.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2014:FIA,
  author =       "Christian S. Jensen",
  title =        "Foreword to Invited Articles Issue",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "26:1--26:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2697050",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hu:2014:EAT,
  author =       "Xiaocheng Hu and Yufei Tao and Chin-Wan Chung",
  title =        "{I/O}-Efficient Algorithms on Triangle Listing and
                 Counting",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "27:1--27:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2691190.2691193",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article studies I/O-efficient algorithms for the
                 triangle listing problem and the triangle counting
                 problem, whose solutions are basic operators in dealing
                 with many other graph problems. In the former problem,
                 given an undirected graph G, the objective is to find
                 all the cliques involving 3 vertices in G. In the
                 latter problem, the objective is to report just the
                 number of such cliques without having to enumerate
                 them. Both problems have been well studied in internal
                 memory, but still remain as difficult challenges when G
                 does not fit in memory, thus making it crucial to
                 minimize the number of disk I/Os performed. Although
                 previous research has attempted to tackle these
                 challenges, the state-of-the-art solutions rely on a
                 set of crippling assumptions to guarantee good
                 performance. Motivated by this, we develop a new
                 algorithm that is provably I/O and CPU efficient at the
                 same time, without making any assumption on the input G
                 at all. The algorithm uses ideas drastically different
                 from all the previous approaches, and outperforms the
                 existing competitors by a factor of over an order of
                 magnitude in our extensive experimentation.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2014:DXK,
  author =       "Marcelo Arenas and Jonny Daenen and Frank Neven and
                 Martin Ugarte and Jan {Van Den Bussche} and Stijn
                 Vansummeren",
  title =        "Discovering {XSD} Keys from {XML} Data",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "28:1--28:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2638547",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A great deal of research into the learning of schemas
                 from XML data has been conducted in recent years to
                 enable the automatic discovery of XML schemas from XML
                 documents when no schema or only a low-quality one is
                 available. Unfortunately, and in strong contrast to,
                 for instance, the relational model, the automatic
                 discovery of even the simplest of XML constraints,
                 namely XML keys, has been left largely unexplored in
                 this context. A major obstacle here is the
                 unavailability of a theory on reasoning about XML keys
                 in the presence of XML schemas, which is needed to
                 validate the quality of candidate keys. The present
                 article embarks on a fundamental study of such a theory
                 and classifies the complexity of several crucial
                 properties concerning XML keys in the presence of an
                 XSD, like, for instance, testing for consistency,
                 boundedness, satisfiability, universality, and
                 equivalence. Of independent interest, novel results are
                 obtained related to cardinality estimation of XPath
                 result sets. A mining algorithm is then developed
                 within the framework of levelwise search. The algorithm
                 leverages known discovery algorithms for functional
                 dependencies in the relational model, but incorporates
                 the properties mentioned before to assess and refine
                 the quality of derived keys. An experimental study on
                 an extensive body of real-world XML data evaluating the
                 effectiveness of the proposed algorithm is provided.",
  acknowledgement = ack-nhfb,
  articleno =    "28",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jung:2014:SLM,
  author =       "Hyungsoo Jung and Hyuck Han and Alan Fekete and Gernot
                 Heiser and Heon Y. Yeom",
  title =        "A Scalable Lock Manager for Multicores",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "29:1--29:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2691190.2691192",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Modern implementations of DBMS software are intended
                 to take advantage of high core counts that are becoming
                 common in high-end servers. However, we have observed
                 that several database platforms, including MySQL,
                 Shore-MT, and a commercial system, exhibit throughput
                 collapse as load increases into oversaturation (where
                 there are more request threads than cores), even for a
                 workload with little or no logical contention for
                 locks, such as a read-only workload. Our analysis of
                 MySQL identifies latch contention within the lock
                 manager as the bottleneck responsible for this
                 collapse. We design a lock manager with reduced
                 latching, implement it in MySQL, and show that it
                 avoids the collapse and generally improves performance.
                 Our efficient implementation of a lock manager is
                 enabled by a staged allocation and deallocation of
                 locks. Locks are preallocated in bulk, so that the lock
                 manager only has to perform simple list manipulation
                 operations during the acquire and release phases of a
                 transaction. Deallocation of the lock data structures
                 is also performed in bulk, which enables the use of
                 fast implementations of lock acquisition and release as
                 well as concurrent deadlock checking.",
  acknowledgement = ack-nhfb,
  articleno =    "29",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papadopoulos:2014:LQA,
  author =       "Stavros Papadopoulos and Graham Cormode and Antonios
                 Deligiannakis and Minos Garofalakis",
  title =        "Lightweight Query Authentication on Streams",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "30:1--30:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2656336",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider a stream outsourcing setting, where a data
                 owner delegates the management of a set of disjoint
                 data streams to an untrusted server. The owner
                 authenticates his streams via signatures. The server
                 processes continuous queries on the union of the
                 streams for clients trusted by the owner. Along with
                 the results, the server sends proofs of result
                 correctness derived from the owner's signatures, which
                 are verifiable by the clients. We design novel
                 constructions for a collection of fundamental problems
                 over streams represented as linear algebraic queries.
                 In particular, our basic schemes authenticate dynamic
                 vector sums, matrix products, and dot products. These
                 techniques can be adapted for authenticating a wide
                 range of important operations in streaming
                 environments, including group-by queries, joins,
                 in-network aggregation, similarity matching, and event
                 processing. We also present extensions to address the
                 case of sliding window queries, and when multiple
                 clients are interested in different subsets of the
                 data. These methods take advantage of a novel nonce
                 chaining technique that we introduce, which is used to
                 reduce the verification cost without affecting any
                 other costs. All our schemes are lightweight and offer
                 strong cryptographic guarantees derived from formal
                 definitions and proofs. We experimentally confirm the
                 practicality of our schemes in the
                 performance-sensitive streaming setting.",
  acknowledgement = ack-nhfb,
  articleno =    "30",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gheerbrant:2014:NEQ,
  author =       "Am{\'e}lie Gheerbrant and Leonid Libkin and Cristina
                 Sirangelo",
  title =        "Na{\"\i}ve Evaluation of Queries over Incomplete
                 Databases",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "31:1--31:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2691190.2691194",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The term na{\"\i}ve evaluation refers to evaluating
                 queries over incomplete databases as if nulls were
                 usual data values, that is, to using the standard
                 database query evaluation engine. Since the semantics
                 of query answering over incomplete databases is that of
                 certain answers, we would like to know when na{\"\i}ve
                 evaluation computes them, that is, when certain answers
                 can be found without inventing new specialized
                 algorithms. For relational databases it is well known
                 that unions of conjunctive queries possess this
                 desirable property, and results on preservation of
                 formulae under homomorphisms tell us that, within
                 relational calculus, this class cannot be extended
                 under the open-world assumption. Our goal here is
                 twofold. First, we develop a general framework that
                 allows us to determine, for a given semantics of
                 incompleteness, classes of queries for which na{\"\i}ve
                 evaluation computes certain answers. Second, we apply
                 this approach to a variety of semantics, showing that
                 for many classes of queries beyond unions of
                 conjunctive queries, na{\"\i}ve evaluation makes
                 perfect sense under assumptions different from open
                 world. Our key observations are: (1) na{\"\i}ve
                 evaluation is equivalent to monotonicity of queries
                 with respect to a semantics-induced ordering, and (2)
                 for most reasonable semantics of incompleteness, such
                 monotonicity is captured by preservation under various
                 types of homomorphisms. Using these results we find
                 classes of queries for which na{\"\i}ve evaluation
                 works, for example, positive first-order formulae for
                 the closed-world semantics. Even more, we introduce a
                 general relation-based framework for defining semantics
                 of incompleteness, show how it can be used to capture
                 many known semantics and to introduce new ones, and
                 describe classes of first-order queries for which
                 na{\"\i}ve evaluation works under such semantics.",
  acknowledgement = ack-nhfb,
  articleno =    "31",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kimelfeld:2014:CMM,
  author =       "Benny Kimelfeld and Phokion G. Kolaitis",
  title =        "The Complexity of Mining Maximal Frequent Subgraphs",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "32:1--32:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2629550",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A frequent subgraph of a given collection of graphs is
                 a graph that is isomorphic to a subgraph of at least as
                 many graphs in the collection as a given threshold.
                 Frequent subgraphs generalize frequent itemsets and
                 arise in various contexts, from bioinformatics to the
                 Web. Since the space of frequent subgraphs is typically
                 extremely large, research in graph mining has focused
                 on special types of frequent subgraphs that can be
                 orders of magnitude smaller in number, yet encapsulate
                 the space of all frequent subgraphs. Maximal frequent
                 subgraphs (i.e., the ones not properly contained in any
                 frequent subgraph) constitute the most useful such
                 type. In this article, we embark on a comprehensive
                 investigation of the computational complexity of mining
                 maximal frequent subgraphs. Our study is carried out by
                 considering the effect of three different parameters:
                 possible restrictions on the class of graphs; a fixed
                 bound on the threshold; and a fixed bound on the number
                 of desired answers. We focus on specific classes of
                 connected graphs: general graphs, planar graphs, graphs
                 of bounded degree, and graphs of bounded treewidth
                 (trees being a special case). Moreover, each class has
                 two variants: that in which the nodes are unlabeled,
                 and that in which they are uniquely labeled. We
                 delineate the complexity of the enumeration problem for
                 each of these variants by determining when it is
                 solvable in (total or incremental) polynomial time and
                 when it is NP-hard. Specifically, for the labeled
                 classes, we show that bounding the threshold yields
                 tractability but, in most cases, bounding the number of
                 answers does not, unless P=NP; an exception is the case
                 of labeled trees, where bounding either of these two
                 parameters yields tractability. The state of affairs
                 turns out to be quite different for the unlabeled
                 classes. The main (and most challenging to prove)
                 result concerns unlabeled trees: we show NP-hardness,
                 even if the input consists of two trees and both the
                 threshold and the number of desired answers are equal
                 to just two. In other words, we establish that the
                 following problem is NP-complete: given two unlabeled
                 trees, do they have more than one maximal subtree in
                 common?",
  acknowledgement = ack-nhfb,
  articleno =    "32",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bienvenu:2014:OBD,
  author =       "Meghyn Bienvenu and Balder {Ten Cate} and Carsten Lutz
                 and Frank Wolter",
  title =        "Ontology-Based Data Access: a Study through
                 Disjunctive {Datalog}, {CSP}, and {MMSNP}",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "33:1--33:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2661643",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Ontology-based data access is concerned with querying
                 incomplete data sources in the presence of
                 domain-specific knowledge provided by an ontology. A
                 central notion in this setting is that of an
                 ontology-mediated query, which is a database query
                 coupled with an ontology. In this article, we study
                 several classes of ontology-mediated queries, where the
                 database queries are given as some form of conjunctive
                 query and the ontologies are formulated in description
                 logics or other relevant fragments of first-order
                 logic, such as the guarded fragment and the unary
                 negation fragment. The contributions of the article are
                 threefold. First, we show that popular
                 ontology-mediated query languages have the same
                 expressive power as natural fragments of disjunctive
                 datalog, and we study the relative succinctness of
                 ontology-mediated queries and disjunctive datalog
                 queries. Second, we establish intimate connections
                 between ontology-mediated queries and constraint
                 satisfaction problems (CSPs) and their logical
                 generalization, MMSNP formulas. Third, we exploit these
                 connections to obtain new results regarding: (i)
                 first-order rewritability and datalog rewritability of
                 ontology-mediated queries; (ii) P/NP dichotomies for
                 ontology-mediated queries; and (iii) the query
                 containment problem for ontology-mediated queries.",
  acknowledgement = ack-nhfb,
  articleno =    "33",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2014:TPP,
  author =       "Chao Li and Daniel Yang Li and Gerome Miklau and Dan
                 Suciu",
  title =        "A Theory of Pricing Private Data",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "34:1--34:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2691190.2691191",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Personal data has value to both its owner and to
                 institutions who would like to analyze it. Privacy
                 mechanisms protect the owner's data while releasing to
                 analysts noisy versions of aggregate query results. But
                 such strict protections of the individual's data have
                 not yet found wide use in practice. Instead, Internet
                 companies, for example, commonly provide free services
                 in return for valuable sensitive information from
                 users, which they exploit and sometimes sell to third
                 parties. As awareness of the value of personal data
                 increases, so has the drive to compensate the end-user
                 for her private information. The idea of monetizing
                 private data can improve over the narrower view of
                 hiding private data, since it empowers individuals to
                 control their data through financial means. In this
                 article we propose a theoretical framework for
                 assigning prices to noisy query answers as a function
                 of their accuracy, and for dividing the price amongst
                 data owners who deserve compensation for their loss of
                 privacy. Our framework adopts and extends key
                 principles from both differential privacy and query
                 pricing in data markets. We identify essential
                 properties of the pricing function and micropayments,
                 and characterize valid solutions.",
  acknowledgement = ack-nhfb,
  articleno =    "34",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Davidson:2014:TCN,
  author =       "Susan Davidson and Sanjeev Khanna and Tova Milo and
                 Sudeepa Roy",
  title =        "Top-$k$ and Clustering with Noisy Comparisons",
  journal =      j-TODS,
  volume =       "39",
  number =       "4",
  pages =        "35:1--35:??",
  month =        dec,
  year =         "2014",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2684066",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Jan 7 15:35:46 MST 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problems of max/top-$k$ and clustering
                 when the comparison operations may be performed by
                 oracles whose answer may be erroneous. Comparisons may
                 either be of type or of value: given two data elements,
                 the answer to a type comparison is ``yes'' if the
                 elements have the same type and therefore belong to the
                 same group (cluster); the answer to a value comparison
                 orders the two data elements. We give efficient
                 algorithms that are guaranteed to achieve correct
                 results with high probability, analyze the cost of
                 these algorithms in terms of the total number of
                 comparisons (i.e., using a fixed-cost model), and show
                 that they are essentially the best possible. We also
                 show that fewer comparisons are needed when values and
                 types are correlated, or when the error model is one in
                 which the error decreases as the distance between the
                 two elements in the sorted order increases. Finally, we
                 examine another important class of cost functions,
                 concave functions, which balances the number of rounds
                 of interaction with the oracle with the number of
                 questions asked of the oracle. Results of this article
                 form an important first step in providing a formal
                 basis for max/top-$k$ and clustering queries in
                 crowdsourcing applications, that is, when the oracle is
                 implemented using the crowd. We explain what
                 simplifying assumptions are made in the analysis, what
                 results carry to a generalized crowdsourcing setting,
                 and what extensions are required to support a
                 full-fledged model.",
  acknowledgement = ack-nhfb,
  articleno =    "35",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2015:EUE,
  author =       "Christian S. Jensen",
  title =        "Editorial: Updates to the Editorial Board",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2747020",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "1e",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pripuzic:2015:TSE,
  author =       "Kresimir Pripuzi{\'c} and Ivana Podnar Zarko and Karl
                 Aberer",
  title =        "Time- and Space-Efficient Sliding Window Top-$k$ Query
                 Processing",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2736701",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A sliding window top-k ( top-k/w ) query monitors
                 incoming data stream objects within a sliding window of
                 size w to identify the k highest-ranked objects with
                 respect to a given scoring function over time.
                 Processing of such queries is challenging because, even
                 when an object is not a top-k/w object at the time when
                 it enters the processing system, it might become one in
                 the future. Thus a set of potential top-k/w objects has
                 to be stored in memory while its size should be
                 minimized to efficiently cope with high data streaming
                 rates. Existing approaches typically store top-k/w and
                 candidate sliding window objects in a k-skyband over a
                 two-dimensional score-time space. However, due to
                 continuous changes of the k-skyband, its maintenance is
                 quite costly. Probabilistic k-skyband is a novel data
                 structure storing data stream objects from a sliding
                 window with significant probability to become top-k/w
                 objects in future. Continuous probabilistic k-skyband
                 maintenance offers considerably improved runtime
                 performance compared to k-skyband maintenance,
                 especially for large values of k, at the expense of a
                 small and controllable error rate. We propose two
                 possible probabilistic k-skyband usages: ( i ) When it
                 is used to process all sliding window objects, the
                 resulting top-k/w algorithm is approximate and adequate
                 for processing random-order data streams. ( ii ) When
                 probabilistic k-skyband is used to process only a
                 subset of most recent sliding window objects, it can
                 improve the runtime performance of continuous k-skyband
                 maintenance, resulting in a novel exact top-k/w
                 algorithm. Our experimental evaluation systematically
                 compares different top-k/w processing algorithms and
                 shows that while competing algorithms offer either time
                 efficiency at the expanse of space efficiency or
                 vice-versa, our algorithms based on the probabilistic
                 k-skyband are both time and space efficient.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Olteanu:2015:SBF,
  author =       "Dan Olteanu and Jakub Z{\'a}vodn{\'y}",
  title =        "Size Bounds for Factorised Representations of Query
                 Results",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2656335",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study two succinct representation systems for
                 relational data based on relational algebra expressions
                 with unions, Cartesian products, and singleton
                 relations: f-representations, which employ algebraic
                 factorisation using distributivity of product over
                 union, and d-representations, which are
                 f-representations where further succinctness is brought
                 by explicit sharing of repeated subexpressions. In
                 particular we study such representations for results of
                 conjunctive queries. We derive tight asymptotic bounds
                 for representation sizes and present algorithms to
                 compute representations within these bounds. We compare
                 the succinctness of f-representations and
                 d-representations for results of equi-join queries, and
                 relate them to fractional edge covers and fractional
                 hypertree decompositions of the query hypergraph.
                 Recent work showed that f-representations can
                 significantly boost the performance of query evaluation
                 in centralised and distributed settings and of machine
                 learning tasks.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pawlik:2015:ECT,
  author =       "Mateusz Pawlik and Nikolaus Augsten",
  title =        "Efficient Computation of the Tree Edit Distance",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699485",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider the classical tree edit distance between
                 ordered labelled trees, which is defined as the
                 minimum-cost sequence of node edit operations that
                 transform one tree into another. The state-of-the-art
                 solutions for the tree edit distance are not
                 satisfactory. The main competitors in the field either
                 have optimal worst-case complexity but the worst case
                 happens frequently, or they are very efficient for some
                 tree shapes but degenerate for others. This leads to
                 unpredictable and often infeasible runtimes. There is
                 no obvious way to choose between the algorithms. In
                 this article we present RTED, a robust tree edit
                 distance algorithm. The asymptotic complexity of our
                 algorithm is smaller than or equal to the complexity of
                 the best competitors for any input instance, that is,
                 our algorithm is both efficient and worst-case optimal.
                 This is achieved by computing a dynamic decomposition
                 strategy that depends on the input trees. RTED is shown
                 optimal among all algorithms that use LRH (
                 left-right-heavy ) strategies, which include RTED and
                 the fastest tree edit distance algorithms presented in
                 literature. In our experiments on synthetic and
                 real-world data we empirically evaluate our solution
                 and compare it to the state-of-the-art.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Drosou:2015:MRD,
  author =       "Marina Drosou and Evaggelia Pitoura",
  title =        "Multiple Radii {DisC} Diversity: Result
                 Diversification Based on Dissimilarity and Coverage",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "4:1--4:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699499",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Recently, result diversification has attracted a lot
                 of attention as a means to improve the quality of
                 results retrieved by user queries. In this article, we
                 introduce a novel definition of diversity called DisC
                 diversity. Given a tuning parameter $r$, which we call
                 radius, we consider two items to be similar if their
                 distance is smaller than or equal to $r$. A DisC
                 diverse subset of a result contains items such that
                 each item in the result is represented by a similar
                 item in the diverse subset and the items in the diverse
                 subset are dissimilar to each other. We show that
                 locating a minimum DisC diverse subset is an NP-hard
                 problem and provide algorithms for its approximation.
                 We extend our definition to the multiple radii case,
                 where each item is associated with a different radius
                 based on its importance, relevance, or other factors.
                 We also propose adapting DisC diverse subsets to a
                 different degree of diversification by adjusting $r$,
                 that is, increasing the radius (or zooming-out) and
                 decreasing the radius (or zooming-in). We present
                 efficient implementations of our algorithms based on
                 the $M$-tree, a spatial index structure, and
                 experimentally evaluate their performance.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Benedikt:2015:ASA,
  author =       "Michael Benedikt and Pierre Bourhis and Clemens Ley",
  title =        "Analysis of Schemas with Access Restrictions",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "5:1--5:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699500",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study verification of systems whose transitions
                 consist of accesses to a Web-based data source. An
                 access is a lookup on a relation within a relational
                 database, fixing values for a set of positions in the
                 relation. For example, a transition can represent
                 access to a Web form, where the user is restricted to
                 filling in values for a particular set of fields. We
                 look at verifying properties of a schema describing the
                 possible accesses of such a system. We present a
                 language where one can describe the properties of an
                 access path and also specify additional restrictions on
                 accesses that are enforced by the schema. Our main
                 property language, AccessLTL, is based on a first-order
                 extension of linear-time temporal logic, interpreting
                 access paths as sequences of relational structures. We
                 also present a lower-level automaton model, A-automata,
                 into which AccessLTL specifications can compile. We
                 show that AccessLTL and A-automata can express static
                 analysis problems related to ``querying with limited
                 access patterns'' that have been studied in the
                 database literature in the past, such as whether an
                 access is relevant to answering a query and whether two
                 queries are equivalent in the accessible data they can
                 return. We prove decidability and complexity results
                 for several restrictions and variants of AccessLTL and
                 explain which properties of paths can be expressed in
                 each restriction.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Athanassoulis:2015:OUD,
  author =       "Manos Athanassoulis and Shimin Chen and Anastasia
                 Ailamaki and Philip B. Gibbons and Radu Stoica",
  title =        "Online Updates on Data Warehouses via Judicious Use of
                 Solid-State Storage",
  journal =      j-TODS,
  volume =       "40",
  number =       "1",
  pages =        "6:1--6:??",
  month =        mar,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699484",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 26 05:54:21 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data warehouses have been traditionally optimized for
                 read-only query performance, allowing only offline
                 updates at night, essentially trading off data
                 freshness for performance. The need for $ 24 \times 7 $
                 operations in global markets and the rise of online and
                 other quickly reacting businesses make concurrent
                 online updates increasingly desirable. Unfortunately,
                 state-of-the-art approaches fall short of supporting
                 fast analysis queries over fresh data. The conventional
                 approach of performing updates in place can
                 dramatically slow down query performance, while prior
                 proposals using differential updates either require
                 large in-memory buffers or may incur significant update
                 migration cost. This article presents a novel approach
                 for supporting online updates in data warehouses that
                 overcomes the limitations of prior approaches by making
                 judicious use of available SSDs to cache incoming
                 updates. We model the problem of query processing with
                 differential updates as a type of outer join between
                 the data residing on disks and the updates residing on
                 SSDs. We present MaSM algorithms for performing such
                 joins and periodic migrations, with small memory
                 footprints, low query overhead, low SSD writes,
                 efficient in-place migration of updates, and correct
                 ACID support. We present detailed modeling of the
                 proposed approach, and provide proofs regarding the
                 fundamental properties of the MaSM algorithms. Our
                 experimentation shows that MaSM incurs only up to 7\%
                 overhead both on synthetic range scans (varying range
                 size from 4KB to 100GB) and in a TPC-H query replay
                 study, while also increasing the update throughput by
                 orders of magnitude.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2015:EBT,
  author =       "Christian S. Jensen",
  title =        "Editorial: The Best of Two Worlds --- Present Your
                 {TODS} Paper at {SIGMOD}",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "7:1--7:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2770931",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Beedkar:2015:CGS,
  author =       "Kaustubh Beedkar and Klaus Berberich and Rainer
                 Gemulla and Iris Miliaraki",
  title =        "Closing the Gap: Sequence Mining at Scale",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2757217",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Frequent sequence mining is one of the fundamental
                 building blocks in data mining. While the problem has
                 been extensively studied, few of the available
                 techniques are sufficiently scalable to handle datasets
                 with billions of sequences; such large-scale datasets
                 arise, for instance, in text mining and session
                 analysis. In this article, we propose MG-FSM, a
                 scalable algorithm for frequent sequence mining on
                 MapReduce. MG-FSM can handle so-called ``gap
                 constraints'', which can be used to limit the output to
                 a controlled set of frequent sequences. Both positional
                 and temporal gap constraints, as well as appropriate
                 maximality and closedness constraints, are supported.
                 At its heart, MG-FSM partitions the input database in a
                 way that allows us to mine each partition independently
                 using any existing frequent sequence mining algorithm.
                 We introduce the notion of $ \omega $-equivalency,
                 which is a generalization of the notion of a
                 ``projected database'' used by many frequent pattern
                 mining algorithms. We also present a number of
                 optimization techniques that minimize partition size,
                 and therefore computational and communication costs,
                 while still maintaining correctness. Our experimental
                 study in the contexts of text mining and session
                 analysis suggests that MG-FSM is significantly more
                 efficient and scalable than alternative approaches.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ameloot:2015:DDF,
  author =       "Tom J. Ameloot",
  title =        "Deciding Determinism with Fairness for Simple
                 Transducer Networks",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2757215",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A distributed database system often operates in an
                 asynchronous communication model where messages can be
                 arbitrarily delayed. This communication model causes
                 nondeterministic effects like unpredictable arrival
                 orders of messages. Nonetheless, in general we want the
                 distributed system to be deterministic; the system
                 should produce the same output despite the
                 nondeterministic effects on messages. Previously, two
                 interpretations of determinism have been proposed. The
                 first says that all infinite fair computation traces
                 produce the same output. The second interpretation is a
                 confluence notion, saying that all finite computation
                 traces can still be extended to produce the same
                 output. A decidability result for the confluence notion
                 was previously obtained for so-called simple transducer
                 networks, a model from the field of declarative
                 networking. In the current article, we also present a
                 decidability result for simple transducer networks, but
                 this time for the first interpretation of determinism,
                 with infinite fair computation traces. We also compare
                 the expressivity of simple transducer networks under
                 both interpretations.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Nagendra:2015:EPS,
  author =       "Mithila Nagendra and K. Sel{\c{c}}uk Candan",
  title =        "Efficient Processing of Skyline-Join Queries over
                 Multiple Data Sources",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699483",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Efficient processing of skyline queries has been an
                 area of growing interest. Many of the earlier skyline
                 techniques assumed that the skyline query is applied to
                 a single data table. Naturally, these algorithms were
                 not suitable for many applications in which the skyline
                 query may involve attributes belonging to multiple data
                 sources. In other words, if the data used in the
                 skyline query are stored in multiple tables, then join
                 operations would be required before the skyline can be
                 searched. The task of computing skylines on multiple
                 data sources has been coined as the skyline-join
                 problem and various skyline-join algorithms have been
                 proposed. However, the current proposals suffer several
                 drawbacks: they often need to scan the input tables
                 exhaustively in order to obtain the set of skyline-join
                 results; moreover, the pruning techniques employed to
                 eliminate the tuples are largely based on expensive
                 pairwise tuple-to-tuple comparisons. In this article,
                 we aim to address these shortcomings by proposing two
                 novel skyline-join algorithms, namely skyline-sensitive
                 join (S$^2$ J) and symmetric skyline-sensitive join
                 (S$^3$ J), to process skyline queries over two data
                 sources. Our approaches compute the results using a
                 novel layer/region pruning technique ( LR-pruning )
                 that prunes the join space in blocks as opposed to
                 individual data points, thereby avoiding excessive
                 pairwise point-to-point dominance checks. Furthermore,
                 the S$^3$ J algorithm utilizes an early stopping
                 condition in order to successfully compute the skyline
                 results by accessing only a subset of the input tables.
                 In addition to S$^2$ J and S$^3$ J, we also propose the
                 S$^2$ J-M and S$^3$ J-M algorithms. These algorithms
                 extend S$^2$ J's and S$^3$ J's two-way skyline-join
                 ability to efficiently process skyline-join queries
                 over more than two data sources. S$^2$ J-M and S$^3$
                 J-M leverage the extended concept of LR-pruning, called
                 M -way LR-pruning, to compute multi-way skyline-joins
                 in which more than two data sources are integrated
                 during skyline processing. We report extensive
                 experimental results that confirm the advantages of the
                 proposed algorithms over state-of-the-art skyline-join
                 techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yuan:2015:OBL,
  author =       "Ganzhao Yuan and Zhenjie Zhang and Marianne Winslett
                 and Xiaokui Xiao and Yin Yang and Zhifeng Hao",
  title =        "Optimizing Batch Linear Queries under Exact and
                 Approximate Differential Privacy",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2699501",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Differential privacy is a promising privacy-preserving
                 paradigm for statistical query processing over
                 sensitive data. It works by injecting random noise into
                 each query result such that it is provably hard for the
                 adversary to infer the presence or absence of any
                 individual record from the published noisy results. The
                 main objective in differentially private query
                 processing is to maximize the accuracy of the query
                 results while satisfying the privacy guarantees.
                 Previous work, notably Li et al. [2010], has suggested
                 that, with an appropriate strategy, processing a batch
                 of correlated queries as a whole achieves considerably
                 higher accuracy than answering them individually.
                 However, to our knowledge there is currently no
                 practical solution to find such a strategy for an
                 arbitrary query batch; existing methods either return
                 strategies of poor quality (often worse than naive
                 methods) or require prohibitively expensive
                 computations for even moderately large domains.
                 Motivated by this, we propose a low-rank mechanism
                 (LRM), the first practical differentially private
                 technique for answering batch linear queries with high
                 accuracy. LRM works for both exact (i.e., $ \epsilon
                 $-) and approximate (i.e., ($ \epsilon $, $ \delta $)-)
                 differential privacy definitions. We derive the utility
                 guarantees of LRM and provide guidance on how to set
                 the privacy parameters, given the user's utility
                 expectation. Extensive experiments using real data
                 demonstrate that our proposed method consistently
                 outperforms state-of-the-art query processing solutions
                 under differential privacy, by large margins.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Termehchy:2015:CEC,
  author =       "Arash Termehchy and Ali Vakilian and Yodsawalai
                 Chodpathumwan and Marianne Winslett",
  title =        "Cost-Effective Conceptual Design for Information
                 Extraction",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2716321",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "It is well established that extracting and annotating
                 occurrences of entities in a collection of unstructured
                 text documents with their concepts improves the
                 effectiveness of answering queries over the collection.
                 However, it is very resource intensive to create and
                 maintain large annotated collections. Since the
                 available resources of an enterprise are limited and/or
                 its users may have urgent information needs, it may
                 have to select only a subset of relevant concepts for
                 extraction and annotation. We call this subset a
                 conceptual design for the annotated collection. In this
                 article, we introduce and formally define the problem
                 of cost-effective conceptual design where, given a
                 collection, a set of relevant concepts, and a fixed
                 budget, one likes to find a conceptual design that most
                 improves the effectiveness of answering queries over
                 the collection. We provide efficient algorithms for
                 special cases of the problem and prove it is generally
                 NP-hard in the number of relevant concepts. We propose
                 three efficient approximations to solve the problem: a
                 greedy algorithm, an approximate popularity
                 maximization (APM for short), and approximate
                 annotation-benefit maximization (AAM for short). We
                 show that, if there are no constraints regrading the
                 overlap of concepts, APM is a fully polynomial time
                 approximation scheme. We also prove that if the
                 relevant concepts are mutually exclusive, the greedy
                 algorithm delivers a constant approximation ratio if
                 the concepts are equally costly, APM has a constant
                 approximation ratio, and AAM is a fully polynomial-time
                 approximation scheme. Our empirical results using a
                 Wikipedia collection and a search engine query log
                 validate the proposed formalization of the problem and
                 show that APM and AAM efficiently compute conceptual
                 designs. They also indicate that, in general, APM
                 delivers the optimal conceptual designs if the relevant
                 concepts are not mutually exclusive. Also, if the
                 relevant concepts are mutually exclusive, the
                 conceptual designs delivered by AAM improve the
                 effectiveness of answering queries over the collection
                 more than the solutions provided by APM.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cao:2015:EPS,
  author =       "Xin Cao and Gao Cong and Tao Guo and Christian S.
                 Jensen and Beng Chin Ooi",
  title =        "Efficient Processing of Spatial Group Keyword
                 Queries",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "13:1--13:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2772600",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "With the proliferation of geo-positioning and
                 geo-tagging techniques, spatio-textual objects that
                 possess both a geographical location and a textual
                 description are gaining in prevalence, and spatial
                 keyword queries that exploit both location and textual
                 description are gaining in prominence. However, the
                 queries studied so far generally focus on finding
                 individual objects that each satisfy a query rather
                 than finding groups of objects where the objects in a
                 group together satisfy a query. We define the problem
                 of retrieving a group of spatio-textual objects such
                 that the group's keywords cover the query's keywords
                 and such that the objects are nearest to the query
                 location and have the smallest inter-object distances.
                 Specifically, we study three instantiations of this
                 problem, all of which are NP-hard. We devise exact
                 solutions as well as approximate solutions with
                 provable approximation bounds to the problems. In
                 addition, we solve the problems of retrieving top- k
                 groups of three instantiations, and study a weighted
                 version of the problem that incorporates object
                 weights. We present empirical studies that offer
                 insight into the efficiency of the solutions, as well
                 as the accuracy of the approximate solutions.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Vincent:2015:TCD,
  author =       "Millist Vincent and Jixue Liu and Hong-Cheu Liu and
                 Sebastian Link",
  title =        "Technical Correspondence: {``Differential
                 Dependencies: Reasoning and Discovery''} Revisited",
  journal =      j-TODS,
  volume =       "40",
  number =       "2",
  pages =        "14:1--14:??",
  month =        jun,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2757214",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 7 09:22:19 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Song:2011:DDR} and response
                 \cite{Song:2017:RDD}.",
  abstract =     "To address the frequently occurring situation where
                 data is inexact or imprecise, a number of extensions to
                 the classical notion of a functional dependency (FD)
                 integrity constraint have been proposed in recent
                 years. One of these extensions is the notion of a
                 differential dependency (DD), introduced in the recent
                 article ``Differential Dependencies: Reasoning and
                 Discovery'' by Song and Chen in the March 2011 edition
                 of this journal. A DD generalises the notion of an FD
                 by requiring only that the values of the attribute from
                 the RHS of the DD satisfy a distance constraint
                 whenever the values of attributes from the LHS of the
                 DD satisfy a distance constraint. In contrast, an FD
                 requires that the values from the attributes in the RHS
                 of an FD be equal whenever the values of the attributes
                 from the LHS of the FD are equal. The article
                 ``Differential Dependencies: Reasoning and Discovery''
                 investigated a number of aspects of DDs, the most
                 important of which, since they form the basis for the
                 other topics investigated, were the consistency problem
                 (determining whether there exists a relation instance
                 that satisfies a set of DDs) and the implication
                 problem (determining whether a set of DDs logically
                 implies another DD). Concerning these problems, a
                 number of results were claimed in ``Differential
                 Dependencies: Reasoning and Discovery''. In this
                 article we conduct a detailed analysis of the
                 correctness of these results. The outcomes of our
                 analysis are that, for almost every claimed result, we
                 show there are either fundamental errors in the proof
                 or the result is false. For some of the claimed results
                 we are able to provide corrected proofs, but for other
                 results their correctness remains open.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lu:2015:BQA,
  author =       "Jiaheng Lu and Chunbin Lin and Wei Wang and Chen Li
                 and Xiaokui Xiao",
  title =        "Boosting the Quality of Approximate String Matching by
                 Synonyms",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "15:1--15:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818177",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A string-similarity measure quantifies the similarity
                 between two text strings for approximate string
                 matching or comparison. For example, the strings
                 ``Sam'' and ``Samuel'' can be considered to be similar.
                 Most existing work that computes the similarity of two
                 strings only considers syntactic similarities, for
                 example, number of common words or $q$-grams. While
                 this is indeed an indicator of similarity, there are
                 many important cases where syntactically-different
                 strings can represent the same real-world object. For
                 example, ``Bill'' is a short form of ``William,'' and
                 ``Database Management Systems'' can be abbreviated as
                 ``DBMS.'' Given a collection of predefined synonyms,
                 the purpose of this article is to explore such existing
                 knowledge to effectively evaluate the similarity
                 between two strings and efficiently perform similarity
                 searches and joins, thereby boosting the quality of
                 approximate string matching. In particular, we first
                 present an expansion-based framework to measure string
                 similarities efficiently while considering synonyms. We
                 then study efficient algorithms for similarity searches
                 and joins by proposing two novel indexes, called
                 SI-trees and QP-trees, which combine
                 signature-filtering and length-filtering strategies. In
                 order to improve the efficiency of our algorithms, we
                 develop an estimator to estimate the size of candidates
                 to enable an online selection of signature filters.
                 This estimator provides strong low-error,
                 high-confidence guarantees while requiring only
                 logarithmic space and time costs, thus making our
                 method attractive both in theory and in practice.
                 Finally, the experimental results from a comprehensive
                 study of the algorithms with three real datasets verify
                 the effectiveness and efficiency of our approaches.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Rusu:2015:WDA,
  author =       "Florin Rusu and Zixuan Zhuang and Mingxi Wu and Chris
                 Jermaine",
  title =        "Workload-Driven Antijoin Cardinality Estimation",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "16:1--16:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818178",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Antijoin cardinality estimation is among a handful of
                 problems that has eluded accurate efficient solutions
                 amenable to implementation in relational query
                 optimizers. Given the widespread use of antijoin and
                 subset-based queries in analytical workloads and the
                 extensive research targeted at join cardinality
                 estimation---a seemingly related problem---the lack of
                 adequate solutions for antijoin cardinality estimation
                 is intriguing. In this article, we introduce a novel
                 sampling-based estimator for antijoin cardinality that
                 (unlike existent estimators) provides sufficient
                 accuracy and efficiency to be implemented in a query
                 optimizer. The proposed estimator incorporates three
                 novel ideas. First, we use prior workload information
                 when learning a mixture superpopulation model of the
                 data offline. Second, we design a Bayesian statistics
                 framework that updates the superpopulation model
                 according to the live queries, thus allowing the
                 estimator to adapt dynamically to the online workload.
                 Third, we develop an efficient algorithm for sampling
                 from a hypergeometric distribution in order to generate
                 Monte Carlo trials, without explicitly instantiating
                 either the population or the sample. When put together,
                 these ideas form the basis of an efficient antijoin
                 cardinality estimator satisfying the strict
                 requirements of a query optimizer, as shown by the
                 extensive experimental results over
                 synthetically-generated as well as massive TPC-H
                 data.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chen:2015:OLQ,
  author =       "Zitong Chen and Yubao Liu and Raymond Chi-Wing Wong
                 and Jiamin Xiong and Ganglin Mai and Cheng Long",
  title =        "Optimal Location Queries in Road Networks",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "17:1--17:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818179",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we study an optimal location query
                 based on a road network. Specifically, given a road
                 network containing clients and servers, an optimal
                 location query finds a location on the road network
                 such that when a new server is set up at this location,
                 a certain cost function computed based on the clients
                 and servers (including the new server) is optimized.
                 Two types of cost functions, namely, MinMax and MaxSum,
                 have been used for this query. The optimal location
                 query problem with MinMax as the cost function is
                 called the MinMax query, which finds a location for
                 setting up a new server such that the maximum cost of a
                 client being served by his/her closest server is
                 minimized. The optimal location query problem with
                 MaxSum as the cost function is called the MaxSum query,
                 which finds a location for setting up a new server such
                 that the sum of the weights of clients attracted by the
                 new server is maximized. The MinMax query and the
                 MaxSum query correspond to two types of optimal
                 location query with the objectives defined from the
                 clients' perspective and from the new server's
                 perspective, respectively. Unfortunately, the existing
                 solutions for the optimal query problem are not
                 efficient. In this article, we propose an efficient
                 algorithm, namely, MinMax-Alg ( MaxSum-Alg ), for the
                 MinMax (MaxSum) query, which is based on a novel idea
                 of nearest location component. We also discuss two
                 extensions of the optimal location query, namely, the
                 optimal multiple-location query and the optimal
                 location query on a 3D road network. Extensive
                 experiments were conducted, showing that our algorithms
                 are faster than the state of the art by at least an
                 order of magnitude on large real benchmark datasets.
                 For example, in our largest real datasets, the state of
                 the art ran for more than 10 (12) hours while our
                 algorithm ran within 3 (2) minutes only for the MinMax
                 (MaxSum) query, that is, our algorithm ran at least 200
                 (600) times faster than the state of the art.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Rietveld:2015:RLD,
  author =       "Kristian F. D. Rietveld and Harry A. G. Wijshoff",
  title =        "Reducing Layered Database Applications to their
                 Essence through Vertical Integration",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "18:1--18:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818180",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In the last decade, improvements on single-core
                 performance of CPUs has stagnated. Consequently,
                 methods for the development and optimization of
                 software for these platforms have to be reconsidered.
                 Software must be optimized such that the available
                 single-core performance is exploited more effectively.
                 This can be achieved by reducing the number of
                 instructions that need to be executed. In this article,
                 we show that layered database applications execute many
                 redundant, nonessential, instructions that can be
                 eliminated without affecting the course of execution
                 and the output of the application. This elimination is
                 performed using a vertical integration process which
                 breaks down the different layers of layered database
                 applications. By doing so, applications are being
                 reduced to their essence, and as a consequence,
                 transformations can be carried out that affect both the
                 application code and the data access code which were
                 not possible before. We show that this vertical
                 integration process can be fully automated and, as
                 such, be integrated in an operational workflow.
                 Experimental evaluation of this approach shows that up
                 to 95\% of the instructions can be eliminated. The
                 reduction of instructions leads to a more efficient use
                 of the available hardware resources. This results in
                 greatly improved performance of the application and a
                 significant reduction in energy consumption.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cheng:2015:SDM,
  author =       "Yu Cheng and Florin Rusu",
  title =        "{SCANRAW}: a Database Meta-Operator for Parallel
                 In-Situ Processing and Loading",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "19:1--19:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818181",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Traditional databases incur a significant
                 data-to-query delay due to the requirement to load data
                 inside the system before querying. Since this is not
                 acceptable in many domains generating massive amounts
                 of raw data (e.g., genomics), databases are entirely
                 discarded. External tables, on the other hand, provide
                 instant SQL querying over raw files. Their performance
                 across a query workload is limited though by the speed
                 of repeated full scans, tokenizing, and parsing of the
                 entire file. In this article, we propose SCANRAW, a
                 novel database meta-operator for in-situ processing
                 over raw files that integrates data loading and
                 external tables seamlessly, while preserving their
                 advantages: optimal performance across a query workload
                 and zero time-to-query. We decompose loading and
                 external table processing into atomic stages in order
                 to identify common functionality. We analyze
                 alternative implementations and discuss possible
                 optimizations for each stage. Our major contribution is
                 a parallel superscalar pipeline implementation that
                 allows SCANRAW to take advantage of the current many-
                 and multicore processors by overlapping the execution
                 of independent stages. Moreover, SCANRAW overlaps query
                 processing with loading by speculatively using the
                 additional I/O bandwidth arising during the conversion
                 process for storing data into the database, such that
                 subsequent queries execute faster. As a result, SCANRAW
                 makes intelligent use of the available system
                 resources---CPU cycles and I/O bandwidth---by switching
                 dynamically between tasks to ensure that optimal
                 performance is achieved. We implement SCANRAW in a
                 state-of-the-art database system and evaluate its
                 performance across a variety of synthetic and
                 real-world datasets. Our results show that SCANRAW with
                 speculative loading achieves the best-possible
                 performance for a query sequence at any point in the
                 processing. Moreover, SCANRAW maximizes resource
                 utilization for the entire workload execution while
                 speculatively loading data and without interfering with
                 normal query processing.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Parchas:2015:UGP,
  author =       "Panos Parchas and Francesco Gullo and Dimitris
                 Papadias and Francesco Bonchi",
  title =        "Uncertain Graph Processing through Representative
                 Instances",
  journal =      j-TODS,
  volume =       "40",
  number =       "3",
  pages =        "20:1--20:??",
  month =        oct,
  year =         "2015",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818182",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Oct 24 11:43:27 MDT 2015",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Data in several applications can be represented as an
                 uncertain graph whose edges are labeled with a
                 probability of existence. Exact query processing on
                 uncertain graphs is prohibitive for most applications,
                 as it involves evaluation over an exponential number of
                 instantiations. Thus, typical approaches employ
                 Monte-Carlo sampling, which (i) draws a number of
                 possible graphs (samples), (ii) evaluates the query on
                 each of them, and (iii) aggregates the individual
                 answers to generate the final result. However, this
                 approach can also be extremely time consuming for large
                 uncertain graphs commonly found in practice. To
                 facilitate efficiency, we study the problem of
                 extracting a single representative instance from an
                 uncertain graph. Conventional processing techniques can
                 then be applied on this representative to closely
                 approximate the result on the original graph. In order
                 to maintain data utility, the representative instance
                 should preserve structural characteristics of the
                 uncertain graph. We start with representatives that
                 capture the expected vertex degrees, as this is a
                 fundamental property of the graph topology. We then
                 generalize the notion of vertex degree to the concept
                 of n -clique cardinality, that is, the number of
                 cliques of size n that contain a vertex. For the first
                 problem, we propose two methods: Average Degree
                 Rewiring (ADR), which is based on random edge rewiring,
                 and Approximate B-Matching (ABM), which applies graph
                 matching techniques. For the second problem, we develop
                 a greedy approach and a game-theoretic framework. We
                 experimentally demonstrate, with real uncertain graphs,
                 that indeed the representative instances can be used to
                 answer, efficiently and accurately, queries based on
                 several metrics such as shortest path distance,
                 clustering coefficient, and betweenness centrality.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Ameloot:2016:WFM,
  author =       "Tom J. Ameloot and Bas Ketsman and Frank Neven and
                 Daniel Zinn",
  title =        "Weaker Forms of Monotonicity for Declarative
                 Networking: a More Fine-Grained Answer to the
                 {CALM}-Conjecture",
  journal =      j-TODS,
  volume =       "40",
  number =       "4",
  pages =        "21:1--21:??",
  month =        jan,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2809784",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jan 21 12:35:53 MST 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The CALM-conjecture, first stated by Hellerstein
                 [2010] and proved in its revised form by Ameloot et al.
                 [2013] within the framework of relational transducer
                 networks, asserts that a query has a coordination-free
                 execution strategy if and only if the query is
                 monotone. Zinn et al. [2012] extended the framework of
                 relational transducer networks to allow for specific
                 data distribution strategies and showed that the
                 nonmonotone win-move query is coordination-free for
                 domain-guided data distributions. In this article, we
                 extend the story by equating increasingly larger
                 classes of coordination-free computations with
                 increasingly weaker forms of monotonicity and present
                 explicit Datalog variants that capture each of these
                 classes. One such fragment is based on stratified
                 Datalog where rules are required to be connected with
                 the exception of the last stratum. In addition, we
                 characterize coordination-freeness as those
                 computations that do not require knowledge about all
                 other nodes in the network, and therefore, can not
                 globally coordinate. The results in this article can be
                 interpreted as a more fine-grained answer to the
                 CALM-conjecture.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bonifati:2016:LJQ,
  author =       "Angela Bonifati and Radu Ciucanu and S{\l}awek
                 Staworko",
  title =        "Learning Join Queries from User Examples",
  journal =      j-TODS,
  volume =       "40",
  number =       "4",
  pages =        "24:1--24:??",
  month =        jan,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2818637",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jan 21 12:35:53 MST 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We investigate the problem of learning join queries
                 from user examples. The user is presented with a set of
                 candidate tuples and is asked to label them as positive
                 or negative examples, depending on whether or not she
                 would like the tuples as part of the join result. The
                 goal is to quickly infer an arbitrary n -ary join
                 predicate across an arbitrary number m of relations
                 while keeping the number of user interactions as
                 minimal as possible. We assume no prior knowledge of
                 the integrity constraints across the involved
                 relations. Inferring the join predicate across multiple
                 relations when the referential constraints are unknown
                 may occur in several applications, such as data
                 integration, reverse engineering of database queries,
                 and schema inference. In such scenarios, the number of
                 tuples involved in the join is typically large. We
                 introduce a set of strategies that let us inspect the
                 search space and aggressively prune what we call
                 uninformative tuples, and we directly present to the
                 user the informative ones---that is, those that allow
                 the user to quickly find the goal query she has in
                 mind. In this article, we focus on the inference of
                 joins with equality predicates and also allow
                 disjunctive join predicates and projection in the
                 queries. We precisely characterize the frontier between
                 tractability and intractability for the following
                 problems of interest in these settings: consistency
                 checking, learnability, and deciding the
                 informativeness of a tuple. Next, we propose several
                 strategies for presenting tuples to the user in a given
                 order that allows minimization of the number of
                 interactions. We show the efficiency of our approach
                 through an experimental study on both benchmark and
                 synthetic datasets.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yang:2016:NFI,
  author =       "Xiaochun Yang and Tao Qiu and Bin Wang and Baihua
                 Zheng and Yaoshu Wang and Chen Li",
  title =        "Negative Factor: Improving Regular-Expression Matching
                 in Strings",
  journal =      j-TODS,
  volume =       "40",
  number =       "4",
  pages =        "25:1--25:46",
  month =        jan,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2847525",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Jan 21 12:35:53 MST 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The problem of finding matches of a regular expression
                 (RE) on a string exists in many applications, such as
                 text editing, biosequence search, and shell commands.
                 Existing techniques first identify candidates using
                 substrings in the RE, then verify each of them using an
                 automaton. These techniques become inefficient when
                 there are many candidate occurrences that need to be
                 verified. In this article, we propose a novel technique
                 that prunes false negatives by utilizing negative
                 factors, which are substrings that cannot appear in an
                 answer. A main advantage of the technique is that it
                 can be integrated with many existing algorithms to
                 improve their efficiency significantly. We present a
                 detailed description of this technique. We develop an
                 efficient algorithm that utilizes negative factors to
                 prune candidates, then improve it by using bit
                 operations to process negative factors in parallel. We
                 show that negative factors, when used with necessary
                 factors (substrings that must appear in each answer),
                 can achieve much better pruning power. We analyze the
                 large number of negative factors, and develop an
                 algorithm for finding a small number of high-quality
                 negative factors. We conducted a thorough experimental
                 study of this technique on real datasets, including DNA
                 sequences, proteins, and text documents, and show
                 significant performance improvement of the
                 state-of-the-art tools by an order of magnitude.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhou:2016:FRW,
  author =       "Zhuojie Zhou and Nan Zhang and Zhiguo Gong and Gautam
                 Das",
  title =        "Faster Random Walks by Rewiring Online Social Networks
                 On-the-Fly",
  journal =      j-TODS,
  volume =       "40",
  number =       "4",
  pages =        "26:1--26:??",
  month =        feb,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2847526",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many online social networks feature restrictive web
                 interfaces that only allow the query of a user's local
                 neighborhood. To enable analytics over such an online
                 social network through its web interface, many recent
                 efforts use Markov Chain Monte Carlo (MCMC) methods
                 such as random walks to sample users in the social
                 network and thereby support analytics based on the
                 samples. The problem with such an approach, however, is
                 the large amount of queries often required for a random
                 walk to converge to a desired (stationary) sampling
                 distribution. In this article, we consider a novel
                 problem of enabling a faster random walk over online
                 social networks by ``rewiring'' the social network
                 on-the-fly. Specifically, we develop a Modified
                 TOpology Sampling (MTO-Sampling) scheme that, by using
                 only information exposed by the restrictive web
                 interface, constructs a ``virtual''
                 random-walk-friendly overlay topology of the social
                 network while performing a random walk and ensures that
                 the random walk follows the modified overlay topology
                 rather than the original one. We describe in this
                 article instantiations of MTO-Sampling for various
                 types of random walks, such as Simple Random Walk
                 (MTO-SRW), Metropolis--Hastings Random Walk (MTO-MHRW),
                 and General Random Walk (MTO-GRW). We not only rigidly
                 prove that MTO-Sampling improves the efficiency of
                 sampling, but we also demonstrate the significance of
                 such improvement through experiments on real-world
                 online social networks such as Google Plus, Epinion,
                 Facebook, etc.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2016:EUE,
  author =       "Christian S. Jensen",
  title =        "Editorial: Updates to the Editorial Board",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "1:1--1:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2893581",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "1e",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Libkin:2016:STV,
  author =       "Leonid Libkin",
  title =        "{SQL}'s Three-Valued Logic and Certain Answers",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "1:1--1:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877206",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The goal of the article is to bridge the difference
                 between theoretical and practical approaches to
                 answering queries over databases with nulls.
                 Theoretical research has long ago identified the notion
                 of correctness of query answering over incomplete data:
                 one needs to find certain answers, which are true
                 regardless of how incomplete information is
                 interpreted. This serves as the notion of correctness
                 of query answering, but carries a huge complexity tag.
                 In practice, on the other hand, query answering must be
                 very efficient, and to achieve this, SQL uses
                 three-valued logic for evaluating queries on databases
                 with nulls. Due to the complexity mismatch, the two
                 approaches cannot coincide, but perhaps they are
                 related in some way. For instance, does SQL always
                 produce answers we can be certain about? This is not
                 so: SQL's and certain answers semantics could be
                 totally unrelated. We show, however, that a slight
                 modification of the three-valued semantics for
                 relational calculus queries can provide the required
                 certainty guarantees. The key point of the new scheme
                 is to fully utilize the three-valued semantics, and
                 classify answers not into certain or noncertain, as was
                 done before, but rather into certainly true, certainly
                 false, or unknown. This yields relatively small changes
                 to the evaluation procedure, which we consider at the
                 level of both declarative (relational calculus) and
                 procedural (relational algebra) queries. These new
                 evaluation procedures give us certainty guarantees even
                 for queries returning tuples with null values.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhang:2016:MOF,
  author =       "Ce Zhang and Arun Kumar and Christopher R{\'e}",
  title =        "Materialization Optimizations for Feature Selection
                 Workloads",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "2:1--2:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877204",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "There is an arms race in the data management industry
                 to support statistical analytics. Feature selection,
                 the process of selecting a feature set that will be
                 used to build a statistical model, is widely regarded
                 as the most critical step of statistical analytics.
                 Thus, we argue that managing the feature selection
                 process is a pressing data management challenge. We
                 study this challenge by describing a feature selection
                 language and a supporting prototype system that builds
                 on top of current industrial R-integration layers. From
                 our interactions with analysts, we learned that feature
                 selection is an interactive human-in-the-loop process,
                 which means that feature selection workloads are rife
                 with reuse opportunities. Thus, we study how to
                 materialize portions of this computation using not only
                 classical database materialization optimizations but
                 also methods that have not previously been used in
                 database optimization, including structural
                 decomposition methods (like QR factorization) and
                 warmstart. These new methods have no analogue in
                 traditional SQL systems, but they may be interesting
                 for array and scientific database applications. On a
                 diverse set of datasets and programs, we find that
                 traditional database-style approaches that ignore these
                 new opportunities are more than two orders of magnitude
                 slower than an optimal plan in this new trade-off space
                 across multiple R backends. Furthermore, we show that
                 it is possible to build a simple cost-based optimizer
                 to automatically select a near-optimal execution plan
                 for feature selection.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Olteanu:2016:EFP,
  author =       "Dan Olteanu and Sebastiaan J. {Van Schaik}",
  title =        "{ENFrame}: a Framework for Processing Probabilistic
                 Data",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "3:1--3:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877205",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article introduces ENFrame, a framework for
                 processing probabilistic data. Using ENFrame, users can
                 write programs in a fragment of Python with constructs
                 such as loops, list comprehension, aggregate operations
                 on lists, and calls to external database engines.
                 Programs are then interpreted probabilistically by
                 ENFrame. We exemplify ENFrame on three clustering
                 algorithms ( k -means, k -medoids, and Markov
                 clustering) and one classification algorithm ( k
                 -nearest-neighbour). A key component of ENFrame is an
                 event language to succinctly encode correlations, trace
                 the computation of user programs, and allow for
                 computation of discrete probability distributions for
                 program variables. We propose a family of sequential
                 and concurrent, exact, and approximate algorithms for
                 computing the probability of interconnected events.
                 Experiments with k -medoids clustering and k
                 -nearest-neighbour show orders-of-magnitude
                 improvements of exact processing using ENFrame over
                 na{\"\i}ve processing in each possible world, of
                 approximate over exact, and of concurrent over
                 sequential processing.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fink:2016:DQN,
  author =       "Robert Fink and Dan Olteanu",
  title =        "Dichotomies for Queries with Negation in Probabilistic
                 Databases",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "4:1--4:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877203",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article charts the tractability frontier of two
                 classes of relational algebra queries in
                 tuple-independent probabilistic databases. The first
                 class consists of queries with join, projection,
                 selection, and negation but without repeating relation
                 symbols and union. The second class consists of
                 quantified queries that express the following binary
                 relationships among sets of entities: set division, set
                 inclusion, set equivalence, and set incomparability.
                 Quantified queries are expressible in relational
                 algebra using join, projection, nested negation, and
                 repeating relation symbols. Each query in the two
                 classes has either polynomial-time or \#P-hard data
                 complexity and the tractable queries can be recognised
                 efficiently. Our result for the first query class
                 extends a known dichotomy for conjunctive queries
                 without self-joins to such queries with negation. For
                 quantified queries, their tractability is sensitive to
                 their outermost projection operator: They are tractable
                 if no attribute representing set identifiers is
                 projected away and \#P-hard otherwise.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhou:2016:BEQ,
  author =       "Xiaoling Zhou and Jianbin Qin and Chuan Xiao and Wei
                 Wang and Xuemin Lin and Yoshiharu Ishikawa",
  title =        "{BEVA}: an Efficient Query Processing Algorithm for
                 Error-Tolerant Autocompletion",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "5:1--5:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877201",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Query autocompletion has become a standard feature in
                 many search applications, especially for search
                 engines. A recent trend is to support the
                 error-tolerant autocompletion, which increases the
                 usability significantly by matching prefixes of
                 database strings and allowing a small number of errors.
                 In this article, we systematically study the query
                 processing problem for error-tolerant autocompletion
                 with a given edit distance threshold. We propose a
                 general framework that encompasses existing methods and
                 characterizes different classes of algorithms and the
                 minimum amount of information they need to maintain
                 under different constraints. We then propose a novel
                 evaluation strategy that achieves the minimum active
                 node size by eliminating ancestor-descendant
                 relationships among active nodes entirely. In addition,
                 we characterize the essence of edit distance
                 computation by a novel data structure named edit vector
                 automaton (EVA). It enables us to compute new active
                 nodes and their associated states efficiently by table
                 lookups. In order to support large distance thresholds,
                 we devise a partitioning scheme to reduce the size and
                 construction cost of the automaton, which results in
                 the universal partitioned EVA (UPEVA) to handle
                 arbitrarily large thresholds. Our extensive evaluation
                 demonstrates that our proposed method outperforms
                 existing approaches in both space and time
                 efficiencies.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fagin:2016:DCI,
  author =       "Ronald Fagin and Benny Kimelfeld and Frederick Reiss
                 and Stijn Vansummeren",
  title =        "Declarative Cleaning of Inconsistencies in Information
                 Extraction",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "6:1--6:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877202",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The population of a predefined relational schema from
                 textual content, commonly known as Information
                 Extraction (IE), is a pervasive task in contemporary
                 computational challenges associated with Big Data.
                 Since the textual content varies widely in nature and
                 structure (from machine logs to informal natural
                 language), it is notoriously difficult to write IE
                 programs that unambiguously extract the sought
                 information. For example, during extraction, an IE
                 program could annotate a substring as both an address
                 and a person name. When this happens, the extracted
                 information is said to be inconsistent, and some way of
                 removing inconsistencies is crucial to compute the
                 final output. Industrial-strength IE systems like GATE
                 and IBM SystemT therefore provide a built-in collection
                 of cleaning operations to remove inconsistencies from
                 extracted relations. These operations, however, are
                 collected in an ad hoc fashion through use cases.
                 Ideally, we would like to allow IE developers to
                 declare their own policies. But existing cleaning
                 operations are defined in an algorithmic way, and hence
                 it is not clear how to extend the built-in operations
                 without requiring low-level coding of internal or
                 external functions. We embark on the establishment of a
                 framework for declarative cleaning of inconsistencies
                 in IE through principles of database theory.
                 Specifically, building upon the formalism of document
                 spanners for IE, we adopt the concept of prioritized
                 repairs, which has been recently proposed as an
                 extension of the traditional database repairs to
                 incorporate priorities among conflicting facts. We show
                 that our framework captures the popular cleaning
                 policies, as well as the POSIX semantics for extraction
                 through regular expressions. We explore the problem of
                 determining whether a cleaning declaration is
                 unambiguous (i.e., always results in a single repair)
                 and whether it increases the expressive power of the
                 extraction language. We give both positive and negative
                 results, some of which are general and some of which
                 apply to policies used in practice.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Pham:2016:ISS,
  author =       "Huy Pham and Cyrus Shahabi and Yan Liu",
  title =        "Inferring Social Strength from Spatiotemporal Data",
  journal =      j-TODS,
  volume =       "41",
  number =       "1",
  pages =        "7:1--7:??",
  month =        apr,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2877200",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 20 11:19:20 MDT 2016",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The advent of geolocation technologies has generated
                 unprecedented rich datasets of people's location
                 information at a very high fidelity. These location
                 datasets can be used to study human behavior; for
                 example, social studies have shown that people who are
                 seen together frequently at the same place and same
                 time are most probably socially related. In this
                 article, we are interested in inferring these social
                 connections by analyzing people's location information;
                 this is useful in a variety of application domains,
                 from sales and marketing to intelligence analysis. In
                 particular, we propose an entropy-based model (EBM)
                 that not only infers social connections but also
                 estimates the strength of social connections by
                 analyzing people's co-occurrences in space and time. We
                 examine two independent methods: diversity and weighted
                 frequency, through which co-occurrences contribute to
                 the strength of a social connection. In addition, we
                 take the characteristics of each location into
                 consideration in order to compensate for cases where
                 only limited location information is available. We also
                 study the role of location semantics in improving our
                 computation of social strength. We develop a parallel
                 implementation of our algorithm using MapReduce to
                 create a scalable and efficient solution for online
                 applications. We conducted extensive sets of
                 experiments with real-world datasets including both
                 people's location data and their social connections,
                 where we used the latter as the ground truth to verify
                 the results of applying our approach to the former. We
                 show that our approach is valid across different
                 networks and outperforms the competitors.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cheung:2016:SBL,
  author =       "Alvin Cheung and Samuel Madden and Armando
                 Solar-Lezama",
  title =        "{Sloth}: Being Lazy Is a Virtue (When Issuing Database
                 Queries)",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2894749",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many web applications store persistent data in
                 databases. During execution, such applications spend a
                 significant amount of time communicating with the
                 database for retrieval and storing of persistent data
                 over the network. These network round-trips represent a
                 significant fraction of the overall execution time for
                 many applications (especially those that issue a lot of
                 database queries) and, as a result, increase their
                 latency. While there has been prior work that aims to
                 eliminate round-trips by batching queries, they are
                 limited by (1) a requirement that developers manually
                 identify batching opportunities, or (2) the fact that
                 they employ static program analysis techniques that
                 cannot exploit many opportunities for batching, as many
                 of these opportunities require knowing precise
                 information about the state of the running program. In
                 this article, we present S loth, a new system that
                 extends traditional lazy evaluation to expose query
                 batching opportunities during application execution,
                 even across loops, branches, and method boundaries.
                 Many such opportunities often require expensive and
                 sophisticated static analysis to recognize from the
                 application source code. Rather than doing so, Sloth
                 instead makes use of dynamic analysis to capture
                 information about the program state and, based on that
                 information, decides how to batch queries and when to
                 issue them to the database. We formalize extended lazy
                 evaluation and prove that it preserves program
                 semantics when executed under standard semantics.
                 Furthermore, we describe our implementation of Sloth
                 and our experience in evaluating Sloth using over 100
                 benchmarks from two large-scale open-source
                 applications, in which Sloth achieved up to a $ 3
                 \times $ reduction in page load time by delaying
                 computation using extended lazy evaluation.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2016:EDC,
  author =       "Christian S. Jensen",
  title =        "Editorial: The Dark Citations of {TODS} Papers and
                 What to Do about It --- or: Cite the Journal Paper",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2946798",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "8e",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Caruccio:2016:SQV,
  author =       "Loredana Caruccio and Giuseppe Polese and Genoveffa
                 Tortora",
  title =        "Synchronization of Queries and Views Upon Schema
                 Evolutions: a Survey",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2903726",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "One of the problems arising upon the evolution of a
                 database schema is that some queries and views defined
                 on the previous schema version might no longer work
                 properly. Thus, evolving a database schema entails the
                 redefinition of queries and views to adapt them to the
                 new schema. Although this problem has been mainly
                 raised in the context of traditional information
                 systems, solutions to it are also advocated in other
                 database-related areas, such as Data Integration, Web
                 Data Integration, and Data Warehouses. The problem is a
                 critical one, since industrial organizations often need
                 to adapt their databases and data warehouses to
                 frequent changes in the real world. In this article, we
                 provide a survey of existing approaches and tools to
                 the problem of adapting queries and views upon a
                 database schema evolution; we also propose a
                 classification framework to enable a uniform comparison
                 method among many heterogeneous approaches and tools.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Deng:2016:CMT,
  author =       "Ting Deng and Wenfei Fan and Floris Geerts",
  title =        "Capturing Missing Tuples and Missing Values",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2901737",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Databases in real life are often neither entirely
                 closed-world nor entirely open-world. Databases in an
                 enterprise are typically partially closed, in which a
                 part of the data is constrained by master data that
                 contains complete information about the enterprise in
                 certain aspects. It has been shown that, despite
                 missing tuples, such a database may turn out to have
                 complete information for answering a query. This
                 article studies partially closed databases from which
                 both tuples and attribute values may be missing. We
                 specify such a database in terms of conditional tables
                 constrained by master data, referred to as c
                 -instances. We first propose three models to
                 characterize whether a c -instance T is complete for a
                 query Q relative to master data. That is, depending on
                 how missing values in T are instantiated, the answer to
                 Q in T remains unchanged when new tuples are added. We
                 then investigate three problems, to determine (a)
                 whether a given c -instance is complete for a query Q,
                 (b) whether there exists a c -instance that is complete
                 for Q relative to master data available, and (c)
                 whether a c -instance is a minimal-size database that
                 is complete for Q. We establish matching lower and
                 upper bounds on these problems for queries expressed in
                 a variety of languages in each of the three models for
                 specifying relative completeness.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dutt:2016:PBF,
  author =       "Anshuman Dutt and Jayant R. Haritsa",
  title =        "Plan Bouquets: a Fragrant Approach to Robust Query
                 Processing",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2901738",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Identifying efficient execution plans for declarative
                 OLAP queries typically entails estimation of several
                 predicate selectivities. In practice, these estimates
                 often differ significantly from the values actually
                 encountered during query execution, leading to poor
                 plan choices and grossly inflated response times. We
                 propose here a conceptually new approach to address
                 this classical problem, wherein the compile-time
                 estimation process is completely eschewed for
                 error-prone selectivities. Instead, from the set of
                 optimal plans in the query's selectivity error space, a
                 limited subset, called the ``plan bouquet,'' is
                 selected such that at least one of the bouquet plans is
                 2-optimal at each location in the space. Then, at run
                 time, a sequence of cost-budgeted executions from the
                 plan bouquet is carried out, eventually finding a plan
                 that executes to completion within its assigned budget.
                 The duration and switching of these executions is
                 controlled by a graded progression of isosurfaces
                 projected onto the optimal performance profile. We
                 prove that this construction results, for the first
                 time, in guarantees on worst-case performance
                 sub-optimality. Moreover, it ensures repeatable
                 execution strategies across different invocations of a
                 query. We then present a suite of enhancements to the
                 basic plan bouquet algorithm, including randomized
                 variants, that result in significantly stronger
                 performance guarantees. An efficient isosurface
                 identification algorithm is also introduced to curtail
                 the bouquet construction overheads. The plan bouquet
                 approach has been empirically evaluated on both
                 PostgreSQL and a commercial DBMS, over the TPC-H and
                 TPC-DS benchmark environments. Our experimental results
                 indicate that it delivers substantial improvements in
                 the worst-case behavior, without impairing the
                 average-case performance, as compared to the native
                 optimizers of these systems. Moreover, it can be
                 implemented using existing optimizer infrastructure,
                 making it relatively easy to incorporate in current
                 database engines. Overall, the plan bouquet approach
                 provides novel performance guarantees that open up new
                 possibilities for robust query processing.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jung:2016:RWR,
  author =       "Jinhong Jung and Kijung Shin and Lee Sael and U.
                 Kang",
  title =        "Random Walk with Restart on Large Graphs Using Block
                 Elimination",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2901736",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Given a large graph, how can we calculate the
                 relevance between nodes fast and accurately? Random
                 walk with restart (RWR) provides a good measure for
                 this purpose and has been applied to diverse data
                 mining applications including ranking, community
                 detection, link prediction, and anomaly detection.
                 Since calculating RWR from scratch takes a long time,
                 various preprocessing methods, most of which are
                 related to inverting adjacency matrices, have been
                 proposed to speed up the calculation. However, these
                 methods do not scale to large graphs because they
                 usually produce large dense matrices that do not fit
                 into memory. In addition, the existing methods are
                 inappropriate when graphs dynamically change because
                 the expensive preprocessing task needs to be computed
                 repeatedly. In this article, we propose B ear, a fast,
                 scalable, and accurate method for computing RWR on
                 large graphs. Bear has two versions: a preprocessing
                 method BearS for static graphs and an incremental
                 update method BearD for dynamic graphs. BearS consists
                 of the preprocessing step and the query step. In the
                 preprocessing step, BearS reorders the adjacency matrix
                 of a given graph so that it contains a large and
                 easy-to-invert submatrix, and precomputes several
                 matrices including the Schur complement of the
                 submatrix. In the query step, BearS quickly computes
                 the RWR scores for a given query node using a block
                 elimination approach with the matrices computed in the
                 preprocessing step. For dynamic graphs, BearD
                 efficiently updates the changed parts in the
                 preprocessed matrices of BearS based on the observation
                 that only small parts of the preprocessed matrices
                 change when few edges are inserted or deleted. Through
                 extensive experiments, we show that BearS significantly
                 outperforms other state-of-the-art methods in terms of
                 preprocessing and query speed, space efficiency, and
                 accuracy. We also show that BearD quickly updates the
                 preprocessed matrices and immediately computes queries
                 when the graph changes.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2016:ECF,
  author =       "Marcelo Arenas and Gonzalo I. Diaz",
  title =        "The Exact Complexity of the First-Order Logic
                 Definability Problem",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "13:1--13:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2886095",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the definability problem for first-order
                 logic, denoted by FO-D ef. The input of FO-Def is a
                 relational database instance I and a relation R; the
                 question to answer is whether there exists a
                 first-order query Q (or, equivalently, a relational
                 algebra expression Q ) such that Q evaluated on I gives
                 R as an answer. Although the study of FO-D ef dates
                 back to 1978, when the decidability of this problem was
                 shown, the exact complexity of FO-Def remains as a
                 fundamental open problem. In this article, we provide a
                 polynomial-time algorithm for solving FO-Def that uses
                 calls to a graph-isomorphism subroutine (or oracle). As
                 a consequence, the first-order definability problem is
                 found to be complete for the class GI of all problems
                 that are polynomial-time Turing reducible to the graph
                 isomorphism problem, thus closing the open question
                 about the exact complexity of this problem. The
                 technique used is also applied to a generalized version
                 of the problem that accepts a finite set of relation
                 pairs, and whose exact complexity was also open; this
                 version is also found to be GI -complete.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cohen:2016:CLT,
  author =       "Sara Cohen and Yaacov Y. Weiss",
  title =        "The Complexity of Learning Tree Patterns from Example
                 Graphs",
  journal =      j-TODS,
  volume =       "41",
  number =       "2",
  pages =        "14:1--14:??",
  month =        jun,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2890492",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:57 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article investigates the problem of learning tree
                 patterns that return nodes with a given set of labels,
                 from example graphs provided by the user. Example
                 graphs are annotated by the user as being either
                 positive or negative. The goal is then to determine
                 whether there exists a tree pattern returning tuples of
                 nodes with the given labels in each of the positive
                 examples, but in none of the negative examples, and
                 furthermore, to find one such pattern if it exists.
                 These are called the satisfiability and learning
                 problems, respectively. This article thoroughly
                 investigates the satisfiability and learning problems
                 in a variety of settings. In particular, we consider
                 example sets that (1) may contain only positive
                 examples, or both positive and negative examples, (2)
                 may contain directed or undirected graphs, and (3) may
                 have multiple occurrences of labels or be uniquely
                 labeled (to some degree). In addition, we consider tree
                 patterns of different types that can allow, or
                 prohibit, wildcard labeled nodes and descendant edges.
                 We also consider two different semantics for mapping
                 tree patterns to graphs. The complexity of
                 satisfiability is determined for the different
                 combinations of settings. For cases in which
                 satisfiability is polynomial, it is also shown that
                 learning is polynomial. (This is nontrivial as
                 satisfying patterns may be exponential in size.)
                 Finally, the minimal learning problem, that is, that of
                 finding a minimal-sized satisfying pattern, is studied
                 for cases in which satisfiability is polynomial.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bailis:2016:SAV,
  author =       "Peter Bailis and Alan Fekete and Ali Ghodsi and Joseph
                 M. Hellerstein and Ion Stoica",
  title =        "Scalable Atomic Visibility with {RAMP} Transactions",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2909870",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Databases can provide scalability by partitioning data
                 across several servers. However, multipartition,
                 multioperation transactional access is often expensive,
                 employing coordination-intensive locking, validation,
                 or scheduling mechanisms. Accordingly, many real-world
                 systems avoid mechanisms that provide useful semantics
                 for multipartition operations. This leads to incorrect
                 behavior for a large class of applications including
                 secondary indexing, foreign key enforcement, and
                 materialized view maintenance. In this work, we
                 identify a new isolation model-Read Atomic (RA)
                 isolation-that matches the requirements of these use
                 cases by ensuring atomic visibility: either all or none
                 of each transaction's updates are observed by other
                 transactions. We present algorithms for Read Atomic
                 Multipartition (RAMP) transactions that enforce atomic
                 visibility while offering excellent scalability,
                 guaranteed commit despite partial failures (via
                 coordination-free execution ), and minimized
                 communication between servers (via partition
                 independence ). These RAMP transactions correctly
                 mediate atomic visibility of updates and provide
                 readers with snapshot access to database state by using
                 limited multiversioning and by allowing clients to
                 independently resolve nonatomic reads. We demonstrate
                 that, in contrast with existing algorithms, RAMP
                 transactions incur limited overhead-even under high
                 contention-and scale linearly to 100 servers.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{To:2016:PSE,
  author =       "Quoc-Cuong To and Benjamin Nguyen and Philippe
                 Pucheral",
  title =        "Private and Scalable Execution of {SQL} Aggregates on
                 a Secure Decentralized Architecture",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2894750",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Current applications, from complex sensor systems
                 (e.g., quantified self) to online e-markets, acquire
                 vast quantities of personal information that usually
                 end up on central servers where they are exposed to
                 prying eyes. Conversely, decentralized architectures
                 that help individuals keep full control of their data
                 complexify global treatments and queries, impeding the
                 development of innovative services. This article aims
                 precisely at reconciling individual's privacy on one
                 side and global benefits for the community and business
                 perspectives on the other. It promotes the idea of
                 pushing the security to secure hardware devices
                 controlling the data at the place of their acquisition.
                 Thanks to these tangible physical elements of trust,
                 secure distributed querying protocols can reestablish
                 the capacity to perform global computations, such as
                 Structured Query Language (SQL) aggregates, without
                 revealing any sensitive information to central servers.
                 This article studies how to secure the execution of
                 such queries in the presence of honest-but-curious and
                 malicious attackers. It also discusses how the
                 resulting querying protocols can be integrated in a
                 concrete decentralized architecture. Cost models and
                 experiments on SQL/Asymmetric Architecture (AA), our
                 distributed prototype running on real tamper-resistant
                 hardware, demonstrate that this approach can scale to
                 nationwide applications.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Burdick:2016:DFL,
  author =       "Douglas Burdick and Ronald Fagin and Phokion G.
                 Kolaitis and Lucian Popa and Wang-Chiew Tan",
  title =        "A Declarative Framework for Linking Entities",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2894748",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We introduce and develop a declarative framework for
                 entity linking and, in particular, for entity
                 resolution. As in some earlier approaches, our
                 framework is based on a systematic use of constraints.
                 However, the constraints we adopt are link-to-source
                 constraints, unlike in earlier approaches where
                 source-to-link constraints were used to dictate how to
                 generate links. Our approach makes it possible to focus
                 entirely on the intended properties of the outcome of
                 entity linking, thus separating the constraints from
                 any procedure of how to achieve that outcome. The core
                 language consists of link-to-source constraints that
                 specify the desired properties of a link relation in
                 terms of source relations and built-in predicates such
                 as similarity measures. A key feature of the
                 link-to-source constraints is that they employ
                 disjunction, which enables the declarative listing of
                 all the reasons two entities should be linked. We also
                 consider extensions of the core language that capture
                 collective entity resolution by allowing
                 interdependencies among the link relations. We identify
                 a class of ``good'' solutions for entity-linking
                 specifications, which we call maximum-value solutions
                 and which capture the strength of a link by counting
                 the reasons that justify it. We study natural
                 algorithmic problems associated with these solutions,
                 including the problem of enumerating the ``good''
                 solutions and the problem of finding the certain links,
                 which are the links that appear in every ``good''
                 solution. We show that these problems are tractable for
                 the core language but may become intractable once we
                 allow interdependencies among the link relations. We
                 also make some surprising connections between our
                 declarative framework, which is deterministic, and
                 probabilistic approaches such as ones based on Markov
                 Logic Networks.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bourhis:2016:BRR,
  author =       "Pierre Bourhis and Gabriele Puppis and Cristian
                 Riveros and Slawek Staworko",
  title =        "Bounded Repairability for Regular Tree Languages",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2898995",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study the problem of bounded repairability of a
                 given restriction tree language R into a target tree
                 language T. More precisely, we say that R is bounded
                 repairable with respect to T if there exists a bound on
                 the number of standard tree editing operations
                 necessary to apply to any tree in R to obtain a tree in
                 T. We consider a number of possible specifications for
                 tree languages: bottom-up tree automata (on curry
                 encoding of unranked trees) that capture the class of
                 XML schemas and document type definitions (DTDs). We
                 also consider a special case when the restriction
                 language R is universal (i.e., contains all trees over
                 a given alphabet). We give an effective
                 characterization of bounded repairability between pairs
                 of tree languages represented with automata. This
                 characterization introduces two tools-synopsis trees
                 and a coverage relation between them-allowing one to
                 reason about tree languages that undergo a bounded
                 number of editing operations. We then employ this
                 characterization to provide upper bounds to the
                 complexity of deciding bounded repairability and show
                 that these bounds are tight. In particular, when the
                 input tree languages are specified with arbitrary
                 bottom-up automata, the problem is coNExp-complete. The
                 problem remains coNExp-complete even if we use
                 deterministic nonrecursive DTDs to specify the input
                 languages. The complexity of the problem can be reduced
                 if we assume that the alphabet, the set of node labels,
                 is fixed: the problem becomes PS pace-complete for
                 nonrecursive DTDs and coNP-complete for deterministic
                 nonrecursive DTDs. Finally, when the restriction tree
                 language R is universal, we show that the bounded
                 repairability problem becomes Exp-complete if the
                 target language is specified by an arbitrary bottom-up
                 tree automaton and becomes tractable (P-complete, in
                 fact) when a deterministic bottom-up automaton is
                 used.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bender:2016:BTC,
  author =       "Michael A. Bender and Roozbeh Ebrahimi and Haodong Hu
                 and Bradley C. Kuszmaul",
  title =        "{B}-Trees and Cache-Oblivious {B}-Trees with
                 Different-Sized Atomic Keys",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2907945",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Most B-tree articles assume that all $N$ keys have the
                 same size $K$, that $ f = B / K$ keys fit in a disk
                 block, and therefore that the search cost is $
                 O(\log_{f + 1} N)$ block transfers. When keys have
                 variable size, B-tree operations have no nontrivial
                 performance guarantees, however. This article provides
                 B-tree-like performance guarantees on dictionaries that
                 contain keys of different sizes in a model in which
                 keys must be stored and compared as opaque objects. The
                 resulting atomic-key dictionaries exhibit performance
                 bounds in terms of the average key size and match the
                 bounds when all keys are the same size. Atomic-key
                 dictionaries can be built with minimal modification to
                 the B-tree structure, simply by choosing the pivot keys
                 properly. This article describes both static and
                 dynamic atomic-key dictionaries. In the static case, if
                 there are N keys with average size K, the search cost
                 is $ O(\lceil K / B \rceil \log_{1 + \lceil B / K
                 \rceil } N)$ expected transfers. It is not possible to
                 transform these expected bounds into worst-case bounds.
                 The cost to build the tree is $ O(N K)$ operations and
                 $ O(N K / B)$ transfers if all keys are presented in
                 sorted order. If not, the cost is the sorting cost. For
                 the dynamic dictionaries, the amortized cost to insert
                 a key $ \kappa $ of arbitrary length at an arbitrary
                 rank is dominated by the cost to search for $ \kappa $.
                 Specifically, the amortized cost to insert a key $
                 \kappa $ of arbitrary length and random rank is $
                 O(\lceil K / B \rceil \log_{1 + \lceil B / K \rceil } N
                 + | \kappa | / B)$ transfers. A dynamic-programming
                 algorithm is shown for constructing a search tree with
                 minimal expected cost. This article also gives a
                 cache-oblivious static atomic-key B-tree, which
                 achieves the same asymptotic performance as the static
                 B-tree dictionary, mentioned previously. A
                 cache-oblivious data structure or algorithm is not
                 parameterized by the block size $B$ or memory size $M$
                 in the memory hierarchy; rather, it is universal,
                 working simultaneously for all possible values of $B$
                 or $M$. On a machine with block size $B$, if there are
                 $N$ keys with average size $K$, search operations costs
                 $ O(\lceil K / B \rceil \log_{1 + \lceil B / K \rceil
                 }N)$ block transfers in expectation. This
                 cache-oblivious layout can be built in $ O(N \log (N
                 K))$ processor operations.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Mazowiecki:2016:MDR,
  author =       "Filip Mazowiecki and Filip Murlak and Adam Witkowski",
  title =        "Monadic {Datalog} and Regular Tree Pattern Queries",
  journal =      j-TODS,
  volume =       "41",
  number =       "3",
  pages =        "20:1--20:??",
  month =        aug,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2925986",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Containment of monadic datalog programs over trees is
                 decidable. The situation is more complex when tree
                 nodes carry labels from an infinite alphabet that can
                 be tested for equality. It then matters whether the
                 descendant relation is allowed or not: the descendant
                 relation can be eliminated easily from monadic programs
                 only when label equalities are not used. With
                 descendant, even containment of linear monadic programs
                 in unions of conjunctive queries is undecidable, and
                 positive results are known only for bounded-depth
                 trees. We show that without descendant, containment of
                 connected monadic programs is decidable over ranked
                 trees, but over unranked trees it is so only for linear
                 programs. With descendant, it becomes decidable over
                 unranked trees under restriction to downward programs:
                 each rule only moves down from the node in the head.
                 This restriction is motivated by regular tree pattern
                 queries, a recent formalism in the area of ActiveXML,
                 which we show to be equivalent to linear downward
                 programs.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tian:2016:BHW,
  author =       "Yuanyuan Tian and Fatma {\"O}zcan and Tao Zou and
                 Romulo Goncalves and Hamid Pirahesh",
  title =        "Building a Hybrid Warehouse: Efficient Joins between
                 Data Stored in {HDFS} and Enterprise Warehouse",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "21:1--21:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2972950",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The Hadoop Distributed File System (HDFS) has become
                 an important data repository in the enterprise as the
                 center for all business analytics, from SQL queries and
                 machine learning to reporting. At the same time,
                 enterprise data warehouses (EDWs) continue to support
                 critical business analytics. This has created the need
                 for a new generation of a special federation between
                 Hadoop-like big data platforms and EDWs, which we call
                 the hybrid warehouse. There are many applications that
                 require correlating data stored in HDFS with EDW data,
                 such as the analysis that associates click logs stored
                 in HDFS with the sales data stored in the database. All
                 existing solutions reach out to HDFS and read the data
                 into the EDW to perform the joins, assuming that the
                 Hadoop side does not have efficient SQL support. In
                 this article, we show that it is actually better to do
                 most data processing on the HDFS side, provided that we
                 can leverage a sophisticated execution engine for joins
                 on the Hadoop side. We identify the best hybrid
                 warehouse architecture by studying various algorithms
                 to join database and HDFS tables. We utilize Bloom
                 filters to minimize the data movement and exploit the
                 massive parallelism in both systems to the fullest
                 extent possible. We describe a new zigzag join
                 algorithm and show that it is a robust join algorithm
                 for hybrid warehouses that performs well in almost all
                 cases. We further develop a sophisticated cost model
                 for the various join algorithms and show that it can
                 facilitate query optimization in the hybrid warehouse
                 to correctly choose the right algorithm under different
                 predicate and join selectivities.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Khamis:2016:JGR,
  author =       "Mahmoud Abo Khamis and Hung Q. Ngo and Christopher
                 R{\'e} and Atri Rudra",
  title =        "Joins via Geometric Resolutions: Worst Case and
                 Beyond",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "22:1--22:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2967101",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We present a simple geometric framework for the
                 relational join. Using this framework, we design an
                 algorithm that achieves the fractional hypertree-width
                 bound, which generalizes classical and recent
                 worst-case algorithmic results on computing joins. In
                 addition, we use our framework and the same algorithm
                 to show a series of what are colloquially known as
                 beyond worst-case results. The framework allows us to
                 prove results for data stored in BTrees,
                 multidimensional data structures, and even multiple
                 indices per table. A key idea in our framework is
                 formalizing the inference one does with an index as a
                 type of geometric resolution, transforming the
                 algorithmic problem of computing joins to a geometric
                 problem. Our notion of geometric resolution can be
                 viewed as a geometric analog of logical resolution. In
                 addition to the geometry and logic connections, our
                 algorithm can also be thought of as backtracking search
                 with memoization.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Papadimitriou:2016:GBA,
  author =       "Dimitra Papadimitriou and Georgia Koutrika and John
                 Mylopoulos and Yannis Velegrakis",
  title =        "The Goal Behind the Action: Toward Goal-Aware Systems
                 and Applications",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "23:1--23:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2934666",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Human activity is almost always intentional, be it in
                 a physical context or as part of an interaction with a
                 computer system. By understanding why user-generated
                 events are happening and what purposes they serve, a
                 system can offer a significantly improved and more
                 engaging experience. However, goals cannot be easily
                 captured. Analyzing user actions such as clicks and
                 purchases can reveal patterns and behaviors, but
                 understanding the goals behind these actions is a
                 different and challenging issue. Our work presents a
                 unified, multidisciplinary viewpoint for goal
                 management that covers many different cases where goals
                 can be used and techniques with which they can be
                 exploited. Our purpose is to provide a common reference
                 point to the concepts and challenging tasks that need
                 to be formally defined when someone wants to approach a
                 data analysis problem from a goal-oriented point of
                 view. This work also serves as a springboard to discuss
                 several open challenges and opportunities for
                 goal-oriented approaches in data management, analysis,
                 and sharing systems and applications.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fazzinga:2016:EIC,
  author =       "Bettina Fazzinga and Sergio Flesca and Filippo Furfaro
                 and Francesco Parisi",
  title =        "Exploiting Integrity Constraints for Cleaning
                 Trajectories of {RFID}-Monitored Objects",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "24:1--24:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2939368",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A probabilistic framework for cleaning the data
                 collected by Radio-Frequency IDentification (RFID)
                 tracking systems is introduced. What has to be cleaned
                 is the set of trajectories that are the possible
                 interpretations of the readings: a trajectory in this
                 set is a sequence whose generic element is a location
                 covered by the reader(s) that made the detection at the
                 corresponding time point. The cleaning is guided by
                 integrity constraints and consists of discarding the
                 inconsistent trajectories and assigning to the others a
                 suitable probability of being the actual one. The
                 probabilities are evaluated by adopting probabilistic
                 conditioning that logically consists of the following
                 steps. First, the trajectories are assigned a priori
                 probabilities that rely on the independence assumption
                 between the time points. Then, these probabilities are
                 revised according to the spatio-temporal correlations
                 encoded by the constraints. This is done by
                 conditioning the a priori probability of each
                 trajectory to the event that the constraints are
                 satisfied: this means taking the ratio of this a priori
                 probability to the sum of the a priori probabilities of
                 all the consistent trajectories. Instead of performing
                 these steps by materializing all the trajectories and
                 their a priori probabilities (which is infeasible,
                 owing to the typically huge number of trajectories),
                 our approach exploits a data structure called
                 conditioned trajectory graph (ct-graph) that compactly
                 represents the trajectories and their conditioned
                 probabilities, and an algorithm for efficiently
                 constructing the ct-graph, which progressively builds
                 it while avoiding the construction of components
                 encoding inconsistent trajectories.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Maabout:2016:SMU,
  author =       "Sofian Maabout and Carlos Ordonez and Patrick Kamnang
                 Wanko and Nicolas Hanusse",
  title =        "Skycube Materialization Using the Topmost Skyline or
                 Functional Dependencies",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "25:1--25:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2955092",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Given a table $ T(I d, D_1, \ldots {}, D_d) $, the
                 skycube of $T$ is the set of skylines with respect to
                 to all nonempty subsets (subspaces) of the set of all
                 dimensions $ \{ D_1, \ldots {}, D_d \} $. To optimize
                 the evaluation of any skyline query, the solutions
                 proposed so far in the literature either (i) precompute
                 all of the skylines or (ii) use compression techniques
                 so that the derivation of any skyline can be done with
                 little effort. Even though solutions (i) are appealing
                 because skyline queries have optimal execution time,
                 they suffer from time and space scalability because the
                 number of skylines to be materialized is exponential
                 with respect to d. On the other hand, solutions (ii)
                 are attractive in terms of memory consumption, but as
                 we show, they also have a high time complexity. In this
                 article, we make contributions to both kinds of
                 solutions. We first observe that skyline patterns are
                 monotonic. This property leads to a simple yet
                 efficient solution for full and partial skycube
                 materialization when the skyline with respect to all
                 dimensions, the topmost skyline, is small. On the other
                 hand, when the topmost skyline is large relative to the
                 size of the input table, it turns out that functional
                 dependencies, a fundamental concept in databases,
                 uncover a monotonic property between skylines. Equipped
                 with this information, we show that closed attributes
                 sets are fundamental for partial and full skycube
                 materialization. Extensive experiments with real and
                 synthetic datasets show that our solutions generally
                 outperform state-of-the-art algorithms.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dignos:2016:EKR,
  author =       "Anton Dign{\"o}s and Michael H. B{\"o}hlen and Johann
                 Gamper and Christian S. Jensen",
  title =        "Extending the Kernel of a Relational {DBMS} with
                 Comprehensive Support for Sequenced Temporal Queries",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "26:1--26:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2967608",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many databases contain temporal, or time-referenced,
                 data and use intervals to capture the temporal aspect.
                 While SQL-based database management systems (DBMSs) are
                 capable of supporting the management of interval data,
                 the support they offer can be improved considerably. A
                 range of proposed temporal data models and query
                 languages offer ample evidence to this effect. Natural
                 queries that are very difficult to formulate in SQL are
                 easy to formulate in these temporal query languages.
                 The increased focus on analytics over historical data
                 where queries are generally more complex exacerbates
                 the difficulties and thus the potential benefits of a
                 temporal query language. Commercial DBMSs have recently
                 started to offer limited temporal functionality in a
                 step-by-step manner, focusing on the representation of
                 intervals and neglecting the implementation of the
                 query evaluation engine. This article demonstrates how
                 it is possible to extend the relational database engine
                 to achieve a full-fledged, industrial-strength
                 implementation of sequenced temporal queries, which
                 intuitively are queries that are evaluated at each time
                 point. Our approach reduces temporal queries to
                 nontemporal queries over data with adjusted intervals,
                 and it leaves the processing of nontemporal queries
                 unaffected. Specifically, the approach hinges on three
                 concepts: interval adjustment, timestamp propagation,
                 and attribute scaling. Interval adjustment is enabled
                 by introducing two new relational operators, a temporal
                 normalizer and a temporal aligner, and the latter two
                 concepts are enabled by the replication of timestamp
                 attributes and the use of so-called scaling functions.
                 By providing a set of reduction rules, we can transform
                 any temporal query, expressed in terms of temporal
                 relational operators, to a query expressed in terms of
                 relational operators and the two new operators. We
                 prove that the size of a transformed query is linear in
                 the number of temporal operators in the original query.
                 An integration of the new operators and the
                 transformation rules, along with query optimization
                 rules, into the kernel of PostgreSQL is reported.
                 Empirical studies with the resulting temporal DBMS are
                 covered that offer insights into pertinent design
                 properties of the article's proposal. The new system is
                 available as open-source software.",
  acknowledgement = ack-nhfb,
  articleno =    "26",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bourhis:2016:GBD,
  author =       "Pierre Bourhis and Marco Manna and Michael Morak and
                 Andreas Pieris",
  title =        "Guarded-Based Disjunctive Tuple-Generating
                 Dependencies",
  journal =      j-TODS,
  volume =       "41",
  number =       "4",
  pages =        "27:1--27:??",
  month =        dec,
  year =         "2016",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2976736",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:58 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We perform an in-depth complexity analysis of query
                 answering under guarded-based classes of disjunctive
                 tuple-generating dependencies (DTGDs), focusing on
                 (unions of) conjunctive queries ((U)CQs). We show that
                 the problem under investigation is very hard, namely
                 2ExpTime-complete, even for fixed sets of dependencies
                 of a very restricted form. This is a surprising lower
                 bound that demonstrates the enormous impact of
                 disjunction on query answering under guarded-based
                 tuple-generating dependencies, and also reveals the
                 source of complexity for expressive logics such as the
                 guarded fragment of first-order logic. We then proceed
                 to investigate whether prominent subclasses of (U)CQs
                 (i.e., queries of bounded treewidth and
                 hypertree-width, and acyclic queries) have a positive
                 impact on the complexity of the problem under
                 consideration. We show that queries of bounded
                 treewidth and bounded hypertree-width do not reduce the
                 complexity of our problem, even if we focus on
                 predicates of bounded arity or on fixed sets of DTGDs.
                 Regarding acyclic queries, although the problem remains
                 2ExpTime-complete in general, in some relevant settings
                 the complexity reduces to ExpTime-complete. Finally,
                 with the aim of identifying tractable cases, we focus
                 our attention on atomic queries. We show that atomic
                 queries do not make the query answering problem easier
                 under classes of guarded-based DTGDs that allow more
                 than one atom to occur in the body of the dependencies.
                 However, the complexity significantly decreases in the
                 case of dependencies that can have only one atom in the
                 body. In particular, we obtain a Ptime-completeness if
                 we focus on predicates of bounded arity, and
                 AC$_0$-membership when the set of dependencies and the
                 query are fixed. Interestingly, our results can be used
                 as a generic tool for establishing complexity results
                 for query answering under various description logics.",
  acknowledgement = ack-nhfb,
  articleno =    "27",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Beame:2017:EMC,
  author =       "Paul Beame and Jerry Li and Sudeepa Roy and Dan
                 Suciu",
  title =        "Exact Model Counting of Query Expressions: Limitations
                 of Propositional Methods",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2984632",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We prove exponential lower bounds on the running time
                 of the state-of-the-art exact model counting
                 algorithms-algorithms for exactly computing the number
                 of satisfying assignments, or the satisfying
                 probability, of Boolean formulas. These algorithms can
                 be seen, either directly or indirectly, as building
                 Decision-Decomposable Negation Normal Form
                 (decision-DNNF) representations of the input Boolean
                 formulas. Decision-DNNFs are a special case of d -DNNFs
                 where d stands for deterministic. We show that any
                 knowledge compilation representations from a class
                 (called DLDDs in this article) that contain
                 decision-DNNFs can be converted into equivalent Free
                 Binary Decision Diagrams (FBDDs), also known as
                 Read-Once Branching Programs, with only a
                 quasi-polynomial increase in representation size.
                 Leveraging known exponential lower bounds for FBDDs, we
                 then obtain similar exponential lower bounds for
                 decision-DNNFs, which imply exponential lower bounds
                 for model-counting algorithms. We also separate the
                 power of decision-DNNFs from d -DNNFs and a
                 generalization of decision-DNNFs known as AND-FBDDs. We
                 then prove new lower bounds for FBDDs that yield
                 exponential lower bounds on the running time of these
                 exact model counters when applied to the problem of
                 query evaluation in tuple-independent probabilistic
                 databases-computing the probability of an answer to a
                 query given independent probabilities of the individual
                 tuples in a database instance. This approach to the
                 query evaluation problem, in which one first obtains
                 the lineage for the query and database instance as a
                 Boolean formula and then performs weighted model
                 counting on the lineage, is known as grounded
                 inference. A second approach, known as lifted inference
                 or extensional query evaluation, exploits the
                 high-level structure of the query as a first-order
                 formula. Although it has been widely believed that
                 lifted inference is strictly more powerful than
                 grounded inference on the lineage alone, no formal
                 separation has previously been shown for query
                 evaluation. In this article, we show such a formal
                 separation for the first time. In particular, we
                 exhibit a family of database queries for which
                 polynomial-time extensional query evaluation techniques
                 were previously known but for which query evaluation
                 via grounded inference using the state-of-the-art exact
                 model counters requires exponential time.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2017:EUE,
  author =       "Christian S. Jensen",
  title =        "Editorial: Updates to the Editorial Board",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3041040",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "1e",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Liu:2017:SMD,
  author =       "Xiufeng Liu and Lukasz Golab and Wojciech Golab and
                 Ihab F. Ilyas and Shichao Jin",
  title =        "Smart Meter Data Analytics: Systems, Algorithms, and
                 Benchmarking",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3004295",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Smart electricity meters have been replacing
                 conventional meters worldwide, enabling automated
                 collection of fine-grained (e.g., every 15 minutes or
                 hourly) consumption data. A variety of smart meter
                 analytics algorithms and applications have been
                 proposed, mainly in the smart grid literature. However,
                 the focus has been on what can be done with the data
                 rather than how to do it efficiently. In this article,
                 we examine smart meter analytics from a software
                 performance perspective. First, we design a performance
                 benchmark that includes common smart meter analytics
                 tasks. These include offline feature extraction and
                 model building as well as a framework for online
                 anomaly detection that we propose. Second, since
                 obtaining real smart meter data is difficult due to
                 privacy issues, we present an algorithm for generating
                 large realistic datasets from a small seed of real
                 data. Third, we implement the proposed benchmark using
                 five representative platforms: a traditional numeric
                 computing platform (Matlab), a relational DBMS with a
                 built-in machine learning toolkit (PostgreSQL/MADlib),
                 a main-memory column store (``System C''), and two
                 distributed data processing platforms (Hive and
                 Spark/Spark Streaming). We compare the five platforms
                 in terms of application development effort and
                 performance on a multicore machine as well as a cluster
                 of 16 commodity servers.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Currim:2017:DMM,
  author =       "Sabah Currim and Richard T. Snodgrass and Young-Kyoon
                 Suh and Rui Zhang",
  title =        "{DBMS} Metrology: Measuring Query Time",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2996454",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "It is surprisingly hard to obtain accurate and precise
                 measurements of the time spent executing a query
                 because there are many sources of variance. To
                 understand these sources, we review relevant
                 per-process and overall measures obtainable from the
                 Linux kernel and introduce a structural causal model
                 relating these measures. A thorough correlational
                 analysis provides strong support for this model. We
                 attempted to determine why a particular measurement
                 wasn't repeatable and then to devise ways to eliminate
                 or reduce that variance. This enabled us to articulate
                 a timing protocol that applies to proprietary DBMSes,
                 that ensures the repeatability of a query, and that
                 obtains a quite accurate query execution time while
                 dropping very few outliers. This resulting query time
                 measurement procedure, termed the Tucson Timing
                 Protocol Version 2 (TTPv2), consists of the following
                 steps: (i) perform sanity checks to ensure data
                 validity; (ii) drop some query executions via clearly
                 motivated predicates; (iii) drop some entire queries at
                 a cardinality, again via clearly motivated predicates;
                 (iv) for those that remain, compute a single measured
                 time by a carefully justified formula over the
                 underlying measures of the remaining query executions;
                 and (v) perform post-analysis sanity checks. The result
                 is a mature, general, robust, self-checking protocol
                 that provides a more precise and more accurate timing
                 of the query. The protocol is also applicable to other
                 operating domains in which measurements of multiple
                 processes each doing computation and I/O is needed.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wu:2017:CFC,
  author =       "You Wu and Pankaj K. Agarwal and Chengkai Li and Jun
                 Yang and Cong Yu",
  title =        "Computational Fact Checking through Query
                 Perturbations",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "4:1--4:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2996453",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Our media is saturated with claims of ``facts'' made
                 from data. Database research has in the past focused on
                 how to answer queries, but has not devoted much
                 attention to discerning more subtle qualities of the
                 resulting claims, for example, is a claim
                 ``cherry-picking''? This article proposes a framework
                 that models claims based on structured data as
                 parameterized queries. Intuitively, with its choice of
                 the parameter setting, a claim presents a particular
                 (and potentially biased) view of the underlying data. A
                 key insight is that we can learn a lot about a claim by
                 ``perturbing'' its parameters and seeing how its
                 conclusion changes. For example, a claim is not robust
                 if small perturbations to its parameters can change its
                 conclusions significantly. This framework allows us to
                 formulate practical fact-checking
                 tasks-reverse-engineering vague claims, and countering
                 questionable claims-as computational problems. Along
                 with the modeling framework, we develop an algorithmic
                 framework that enables efficient instantiations of
                 ``meta'' algorithms by supplying appropriate
                 algorithmic building blocks. We present real-world
                 examples and experiments that demonstrate the power of
                 our model, efficiency of our algorithms, and usefulness
                 of their results.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Teflioudi:2017:EAM,
  author =       "Christina Teflioudi and Rainer Gemulla",
  title =        "Exact and Approximate Maximum Inner Product Search
                 with {LEMP}",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "5:1--5:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2996452",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We study exact and approximate methods for maximum
                 inner product search, a fundamental problem in a number
                 of data mining and information retrieval tasks. We
                 propose the LEMP framework, which supports both exact
                 and approximate search with quality guarantees. At its
                 heart, LEMP transforms a maximum inner product search
                 problem over a large database of vectors into a number
                 of smaller cosine similarity search problems. This
                 transformation allows LEMP to prune large parts of the
                 search space immediately and to select suitable search
                 algorithms for each of the remaining problems
                 individually. LEMP is able to leverage existing methods
                 for cosine similarity search, but we also provide a
                 number of novel search algorithms tailored to our
                 setting. We conducted an extensive experimental study
                 that provides insight into the performance of many
                 state-of-the-art techniques-including LEMP-on multiple
                 real-world datasets. We found that LEMP often was
                 significantly faster or more accurate than alternative
                 methods.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Shi:2017:UUA,
  author =       "Xiaogang Shi and Bin Cui and Gillian Dobbie and Beng
                 Chin Ooi",
  title =        "{UniAD}: a Unified Ad Hoc Data Processing System",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "6:1--6:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3009957",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Instead of constructing complex declarative queries,
                 many users prefer to write their programs using
                 procedural code embedded with simple queries. Since
                 many users are not expert programmers or the programs
                 are written in a rush, these programs usually exhibit
                 poor performance in practice and it is a challenge to
                 automatically and efficiently optimize these programs.
                 In this article, we present UniAD, which stands for Uni
                 fied execution for Ad hoc Data processing, a system
                 designed to simplify the programming of data processing
                 tasks and provide efficient execution for user
                 programs. We provide the background of program
                 semantics and propose a novel intermediate
                 representation, called Unified Intermediate
                 Representation (UniIR), which utilizes a simple and
                 expressive mechanism HOQ to describe the operations
                 performed in programs. By combining both procedural and
                 declarative logics with the proposed intermediate
                 representation, we can perform various optimizations
                 across the boundary between procedural and declarative
                 code. We propose a transformation-based optimizer to
                 automatically optimize programs and implement the UniAD
                 system. The extensive experimental results on various
                 benchmarks demonstrate that our techniques can
                 significantly improve the performance of a wide range
                 of data processing programs.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Song:2017:RDD,
  author =       "Shaoxu Song and Lei Chen",
  title =        "Response to {``Differential Dependencies
                 Revisited''}",
  journal =      j-TODS,
  volume =       "42",
  number =       "1",
  pages =        "7:1--7:??",
  month =        mar,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/2983602",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Mar 2 18:02:59 MST 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "See \cite{Song:2011:DDR,Vincent:2015:TCD}.",
  abstract =     "A recent article [Vincent et al. 2015] concerns the
                 correctness of several results in reasoning about
                 differential dependencies ( dds), originally reported
                 in Song and Chen [2011]. The major concern by Vincent
                 et al. [2015] roots from assuming a type of infeasible
                 differential functions in the given dds for consistency
                 and implication analysis, which are not allowed in Song
                 and Chen [2011]. A differential function is said to be
                 infeasible if there is no tuple pair with values that
                 can satisfy the specified distance constraints. For
                 example, [price($ < 2 $, $ > 4$)] requires the
                 difference of two price values to be $ < 2$ and $ > 4$
                 at the same time, which is clearly impossible. Although
                 dds involving infeasible differential functions may be
                 syntactically interesting, they are semantically
                 meaningless and would neither be specified by domain
                 experts nor discovered from data. For these reasons,
                 infeasible differential functions are not considered
                 [Song and Chen 2011] and the results in Song and Chen
                 [2011] are correct, in contrast to what is claimed in
                 Vincent et al. [2015].",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Paparrizos:2017:FAT,
  author =       "John Paparrizos and Luis Gravano",
  title =        "Fast and Accurate Time-Series Clustering",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "8:1--8:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3044711",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The proliferation and ubiquity of temporal data across
                 many disciplines has generated substantial interest in
                 the analysis and mining of time series. Clustering is
                 one of the most popular data-mining methods, not only
                 due to its exploratory power but also because it is
                 often a preprocessing step or subroutine for other
                 techniques. In this article, we present $k$-Shape and
                 $k$-MultiShapes ($k$-MS), two novel algorithms for
                 time-series clustering. $k$-Shape and $k$-MS rely on a
                 scalable iterative refinement procedure. As their
                 distance measure, $k$-Shape and $k$-MS use shape-based
                 distance (SBD), a normalized version of the
                 cross-correlation measure, to consider the shapes of
                 time series while comparing them. Based on the
                 properties of SBD, we develop two new methods, namely
                 ShapeExtraction (SE) and MultiShapesExtraction (MSE),
                 to compute cluster centroids that are used in every
                 iteration to update the assignment of time series to
                 clusters. $k$-Shape relies on SE to compute a single
                 centroid per cluster based on all time series in each
                 cluster. In contrast, $k$-MS relies on MSE to compute
                 multiple centroids per cluster to account for the
                 proximity and spatial distribution of time series in
                 each cluster. To demonstrate the robustness of SBD,
                 $k$-Shape, and $k$-MS, we perform an extensive
                 experimental evaluation on 85 datasets against
                 state-of-the-art distance measures and clustering
                 methods for time series using rigorous statistical
                 analysis. SBD, our efficient and parameter-free
                 distance measure, achieves similar accuracy to Dynamic
                 Time Warping (DTW), a highly accurate but
                 computationally expensive distance measure that
                 requires parameter tuning. For clustering, we compare
                 $k$-Shape and $k$-MS against scalable and non-scalable
                 partitional, hierarchical, spectral, density-based, and
                 shapelet-based methods, with combinations of the most
                 competitive distance measures. $k$-Shape outperforms
                 all scalable methods in terms of accuracy. Furthermore,
                 $k$-Shape also outperforms all non-scalable approaches,
                 with one exception, namely $k$-medoids with DTW, which
                 achieves similar accuracy. However, unlike $k$-Shape,
                 this approach requires tuning of its distance measure
                 and is significantly slower than $k$-Shape. $k$-MS
                 performs similarly to $k$-Shape in comparison to rival
                 methods, but $k$-MS is significantly more accurate than
                 $k$-Shape. Beyond clustering, we demonstrate the
                 effectiveness of $k$-Shape to reduce the search space
                 of one-nearest-neighbor classifiers for time series.
                 Overall, SBD, $k$-Shape, and $k$-MS emerge as
                 domain-independent, highly accurate, and efficient
                 methods for time-series comparison and clustering with
                 broad applications.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Koutris:2017:CQA,
  author =       "Paraschos Koutris and Jef Wijsen",
  title =        "Consistent Query Answering for Self-Join-Free
                 Conjunctive Queries Under Primary Key Constraints",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3068334",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A relational database is said to be uncertain if
                 primary key constraints can possibly be violated. A
                 repair (or possible world) of an uncertain database is
                 obtained by selecting a maximal number of tuples
                 without ever selecting two distinct tuples with the
                 same primary key value. For any Boolean query $q$,
                 CERTAINTY($q$) is the problem that takes an uncertain
                 database db as input and asks whether $q$ is true in
                 every repair of db. The complexity of this problem has
                 been particularly studied for $q$ ranging over the
                 class of self-join-free Boolean conjunctive queries. A
                 research challenge is to determine, given $q$, whether
                 CERTAINTY($q$) belongs to complexity classes FO, $P$,
                 or coNP-complete. In this article, we combine existing
                 techniques for studying this complexity classification
                 task. We show that, for any self-join-free Boolean
                 conjunctive query $q$, it can be decided whether or not
                 CERTAINTY($q$) is in FO. We additionally show how to
                 construct a single SQL query for solving CERTAINTY($q$)
                 if it is in FO. Further, for any self-join-free Boolean
                 conjunctive query $q$, CERTAINTY($q$) is either in $P$
                 or coNP-complete and the complexity dichotomy is
                 effective. This settles a research question that has
                 been open for 10 years.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Yu:2017:ODM,
  author =       "Yanwei Yu and Lei Cao and Elke A. Rundensteiner and
                 Qin Wang",
  title =        "Outlier Detection over Massive-Scale Trajectory
                 Streams",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3013527",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The detection of abnormal moving objects over
                 high-volume trajectory streams is critical for
                 real-time applications ranging from military
                 surveillance to transportation management. Yet this
                 outlier detection problem, especially along both the
                 spatial and temporal dimensions, remains largely
                 unexplored. In this work, we propose a rich taxonomy of
                 novel classes of neighbor-based trajectory outlier
                 definitions that model the anomalous behavior of moving
                 objects for a large range of real-time applications.
                 Our theoretical analysis and empirical study on two
                 real-world datasets-the Beijing Taxi trajectory data
                 and the Ground Moving Target Indicator data stream-and
                 one generated Moving Objects dataset demonstrate the
                 effectiveness of our taxonomy in effectively capturing
                 different types of abnormal moving objects.
                 Furthermore, we propose a general strategy for
                 efficiently detecting these new outlier classes called
                 the {\em minimal examination\/} (MEX) framework. The
                 MEX framework features three core optimization
                 principles, which leverage spatiotemporal as well as
                 the predictability properties of the neighbor evidence
                 to minimize the detection costs. Based on this
                 foundation, we design algorithms that detect the
                 outliers based on these classes of new outlier
                 semantics that successfully leverage our optimization
                 principles. Our comprehensive experimental study
                 demonstrates that our proposed MEX strategy drives the
                 detection costs 100-fold down into the practical realm
                 for applications that analyze high-volume trajectory
                 streams in near real time.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Han:2017:CCF,
  author =       "Yunheng Han and Weiwei Sun and Baihua Zheng",
  title =        "{COMPRESS}: a Comprehensive Framework of Trajectory
                 Compression in Road Networks",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3015457",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/datacompression.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "More and more advanced technologies have become
                 available to collect and integrate an unprecedented
                 amount of data from multiple sources, including GPS
                 trajectories about the traces of moving objects. Given
                 the fact that GPS trajectories are vast in size while
                 the information carried by the trajectories could be
                 redundant, we focus on trajectory compression in this
                 article. As a systematic solution, we propose a
                 comprehensive framework, namely, COMPRESS
                 (Comprehensive Paralleled Road-Network-Based Trajectory
                 Compression), to compress GPS trajectory data in an
                 urban road network. In the preprocessing step, COMPRESS
                 decomposes trajectories into spatial paths and temporal
                 sequences, with a thorough justification for trajectory
                 decomposition. In the compression step, COMPRESS
                 performs spatial compression on spatial paths, and
                 temporal compression on temporal sequences in parallel.
                 It introduces two alternative algorithms with different
                 strengths for lossless spatial compression and designs
                 lossy but error-bounded algorithms for temporal
                 compression. It also presents query processing
                 algorithms to support error-bounded location-based
                 queries on compressed trajectories without full
                 decompression. All algorithms under COMPRESS are
                 efficient and have the time complexity of $ O (| T |)
                 $, where $ | T | $ is the size of the input trajectory
                 $T$. We have also conducted a comprehensive
                 experimental study to demonstrate the effectiveness of
                 COMPRESS, whose compression ratio is significantly
                 better than related approaches.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{TenCate:2017:AAS,
  author =       "Balder {Ten Cate} and Phokion G. Kolaitis and Kun Qian
                 and Wang-Chiew Tan",
  title =        "Approximation Algorithms for Schema-Mapping Discovery
                 from Data Examples",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3044712",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In recent years, data examples have been at the core
                 of several different approaches to schema-mapping
                 design. In particular, Gottlob and Senellart introduced
                 a framework for schema-mapping discovery from a single
                 data example, in which the derivation of a schema
                 mapping is cast as an optimization problem. Our goal is
                 to refine and study this framework in more depth. Among
                 other results, we design a polynomial-time
                 $\log(n)$-approximation algorithm for computing optimal
                 schema mappings from a given set of data examples
                 (where $n$ is the combined size of the given data
                 examples) for a restricted class of schema mappings;
                 moreover, we show that this approximation ratio cannot
                 be improved. In addition to the complexity-theoretic
                 results, we implemented the aforementioned
                 $\log(n)$-approximation algorithm and carried out an
                 experimental evaluation in a real-world mapping
                 scenario.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Maniu:2017:IFQ,
  author =       "Silviu Maniu and Reynold Cheng and Pierre Senellart",
  title =        "An Indexing Framework for Queries on Probabilistic
                 Graphs",
  journal =      j-TODS,
  volume =       "42",
  number =       "2",
  pages =        "13:1--13:??",
  month =        jun,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3044713",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Jun 23 16:29:07 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Information in many applications, such as mobile
                 wireless systems, social networks, and road networks,
                 is captured by graphs. In many cases, such information
                 is uncertain. We study the problem of querying a
                 probabilistic graph, in which vertices are connected to
                 each other probabilistically. In particular, we examine
                 ``source-to-target'' queries (ST-queries), such as
                 computing the shortest path between two vertices. The
                 major difference with the deterministic setting is that
                 query answers are enriched with probabilistic
                 annotations. Evaluating ST-queries over probabilistic
                 graphs is \#P-hard, as it requires examining an
                 exponential number of ``possible worlds''-database
                 instances generated from the probabilistic graph.
                 Existing solutions to the ST-query problem, which
                 sample possible worlds, have two downsides: (i) a
                 possible world can be very large and (ii) many samples
                 are needed for reasonable accuracy. To tackle these
                 issues, we study the ProbTree, a data structure that
                 stores a succinct, or indexed, version of the possible
                 worlds of the graph. Existing ST-query solutions are
                 executed on top of this structure, with the number of
                 samples and sizes of the possible worlds reduced. We
                 examine lossless and lossy methods for generating the
                 ProbTree, which reflect the tradeoff between the
                 accuracy and efficiency of query evaluation. We analyze
                 the correctness and complexity of these approaches. Our
                 extensive experiments on real datasets show that the
                 ProbTree is fast to generate and small in size. It also
                 enhances the accuracy and efficiency of existing
                 ST-query algorithms significantly.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gan:2017:HAE,
  author =       "Junhao Gan and Yufei Tao",
  title =        "On the Hardness and Approximation of {Euclidean}
                 {DBSCAN}",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "14:1--14:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3083897",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "DBSCAN is a method proposed in 1996 for clustering
                 multi-dimensional points, and has received extensive
                 applications. Its computational hardness is still
                 unsolved to this date. The original KDD'96 paper
                 claimed an algorithm of $ O(n \log n) $ ``average
                 runtime complexity'' (where n is the number of data
                 points) without a rigorous proof. In 2013, a genuine $
                 O(n \log n) $-time algorithm was found in 2D space
                 under Euclidean distance. The hardness of
                 dimensionality $ d \geq 3$ has remained open ever
                 since. This article considers the problem of computing
                 DBSCAN clusters from scratch (assuming no existing
                 indexes) under Euclidean distance. We prove that, for $
                 d \geq 3$, the problem requires $ \omega (n^{4 / 3}) $
                 time to solve, unless very significant
                 breakthroughs-ones widely believed to be
                 impossible-could be made in theoretical computer
                 science. Motivated by this, we propose a relaxed
                 version of the problem called $ \rho $- approximate
                 DBSCAN, which returns the same clusters as DBSCAN,
                 unless the clusters are ``unstable'' (i.e., they change
                 once the input parameters are slightly perturbed). The
                 $ \rho $-approximate problem can be settled in $ O(n) $
                 expected time regardless of the constant dimensionality
                 d. The article also enhances the previous result on the
                 exact DBSCAN problem in 2D space. We show that, if the
                 n data points have been pre-sorted on each dimension
                 (i.e., one sorted list per dimension), the problem can
                 be settled in $ O(n) $ worst-case time. As a corollary,
                 when all the coordinates are integers, the 2D DBSCAN
                 problem can be solved in $ O(n \log \log n) $ time
                 deterministically, improving the existing $ O(n \log n)
                 $ bound.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Martens:2017:BCS,
  author =       "Wim Martens and Frank Neven and Matthias Niewerth and
                 Thomas Schwentick",
  title =        "{BonXai}: Combining the Simplicity of {DTD} with the
                 Expressiveness of {XML} Schema",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "15:1--15:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3105960",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "While the migration from DTD to XML Schema was driven
                 by a need for increased expressivity and flexibility,
                 the latter was also significantly more complex to use
                 and understand. Whereas DTDs are characterized by their
                 simplicity, XML Schema Documents are notoriously
                 difficult. In this article, we introduce the XML
                 specification language BonXai, which incorporates many
                 features of XML Schema but is arguably almost as easy
                 to use as DTDs. In brief, the latter is achieved by
                 sacrificing the explicit use of types in favor of
                 simple patterns expressing contexts for elements. The
                 goal of BonXai is not to replace XML Schema but rather
                 to provide a simpler alternative for users who want to
                 go beyond the expressiveness and features of DTD but do
                 not need the explicit use of types. Furthermore, XML
                 Schema processing tools can be used as a back-end for
                 BonXai, since BonXai can be automatically converted
                 into XML Schema. A particularly strong point of BonXai
                 is its solid foundation rooted in a decade of
                 theoretical work around pattern-based schemas. We
                 present a formal model for a core fragment of BonXai
                 and the translation algorithms to and from a core
                 fragment of XML Schema. We prove that BonXai and XML
                 Schema can be converted back-and-forth on the level of
                 tree languages and we formally study the size
                 trade-offs between the two languages.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zheng:2017:ESB,
  author =       "Weiguo Zheng and Lei Zou and Lei Chen and Dongyan
                 Zhao",
  title =        "Efficient {SimRank}-Based Similarity Join",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "16:1--16:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3083899",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Graphs have been widely used to model complex data in
                 many real-world applications. Answering vertex join
                 queries over large graphs is meaningful and
                 interesting, which can benefit friend recommendation in
                 social networks and link prediction, and so on. In this
                 article, we adopt ``SimRank'' [13] to evaluate the
                 similarity between two vertices in a large graph
                 because of its generality. Note that ``SimRank'' is
                 purely structure dependent, and it does not rely on the
                 domain knowledge. Specifically, we define a
                 SimRank-based join (SRJ) query to find all vertex pairs
                 satisfying the threshold from two sets of vertices $U$
                 and $V$. To reduce the search space, we propose a
                 shortest-path-distance-based upper bound for SimRank
                 scores to prune unpromising vertex pairs. In the
                 verification, we propose a novel index, called h-go
                 cover$^+$, to efficiently compute the SimRank score of
                 any single vertex pair. Given a graph $G$, we only
                 materialize the SimRank scores of a small proportion of
                 vertex pairs (i.e., the h-go cover$^+$ vertex pairs),
                 based on which the SimRank score of any vertex pair can
                 be computed easily. To find the h-go cover$^+$ vertex
                 pairs, we propose an efficient method without building
                 the vertex-pair graph. Hence, large graphs can be dealt
                 with easily. Extensive experiments over both real and
                 synthetic datasets confirm the efficiency of our
                 solution.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kaminski:2017:QNA,
  author =       "Mark Kaminski and Egor V. Kostylev and Bernardo Cuenca
                 Grau",
  title =        "Query Nesting, Assignment, and Aggregation in {SPARQL
                 1.1}",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "17:1--17:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3083898",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Answering aggregate queries is a key requirement of
                 emerging applications of Semantic Technologies, such as
                 data warehousing, business intelligence, and sensor
                 networks. To fulfil the requirements of such
                 applications, the standardization of SPARQL 1.1 led to
                 the introduction of a wide range of constructs that
                 enable value computation, aggregation, and query
                 nesting. In this article, we provide an in-depth formal
                 analysis of the semantics and expressive power of these
                 new constructs as defined in the SPARQL 1.1
                 specification, and hence lay the necessary foundations
                 for the development of robust, scalable, and extensible
                 query engines supporting complex numerical and
                 analytics tasks.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Tschirschnitz:2017:DID,
  author =       "Fabian Tschirschnitz and Thorsten Papenbrock and Felix
                 Naumann",
  title =        "Detecting Inclusion Dependencies on Very Many Tables",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "18:1--18:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3105959",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Detecting inclusion dependencies, the prerequisite of
                 foreign keys, in relational data is a challenging task.
                 Detecting them among the hundreds of thousands or even
                 millions of tables on the web is daunting. Still, such
                 inclusion dependencies can help connect disparate
                 pieces of information on the Web and reveal unknown
                 relationships among tables. With the algorithm M any,
                 we present a novel inclusion dependency detection
                 algorithm, specialized for the very many --- but
                 typically small --- tables found on the Web. We make
                 use of Bloom filters and indexed bit-vectors to show
                 the feasibility of our approach. Our evaluation on two
                 corpora of Web tables shows a superior runtime over
                 known approaches and its usefulness to reveal hidden
                 structures on the Web.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schubert:2017:DRR,
  author =       "Erich Schubert and J{\"o}rg Sander and Martin Ester
                 and Hans Peter Kriegel and Xiaowei Xu",
  title =        "{DBSCAN} Revisited, Revisited: Why and How You Should
                 (Still) Use {DBSCAN}",
  journal =      j-TODS,
  volume =       "42",
  number =       "3",
  pages =        "19:1--19:??",
  month =        aug,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3068335",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Aug 24 16:30:15 MDT 2017",
  bibsource =    "http://www.acm.org/pubs/contents/journals/tods/;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "At SIGMOD 2015, an article was presented with the
                 title ``DBSCAN Revisited: Mis-Claim, Un-Fixability, and
                 Approximation'' that won the conference's best paper
                 award. In this technical correspondence, we want to
                 point out some inaccuracies in the way DBSCAN was
                 represented, and why the criticism should have been
                 directed at the assumption about the performance of
                 spatial index structures such as R-trees and not at an
                 algorithm that can use such indexes. We will also
                 discuss the relationship of DBSCAN performance and the
                 indexability of the dataset, and discuss some
                 heuristics for choosing appropriate DBSCAN parameters.
                 Some indicators of bad parameters will be proposed to
                 help guide future users of this algorithm in choosing
                 parameters such as to obtain both meaningful results
                 and good performance. In new experiments, we show that
                 the new SIGMOD 2015 methods do not appear to offer
                 practical benefits if the DBSCAN parameters are well
                 chosen and thus they are primarily of theoretical
                 interest. In conclusion, the original DBSCAN algorithm
                 with effective indexes and reasonably chosen parameter
                 values performs competitively compared to the method
                 proposed by Gan and Tao.",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Aberger:2017:ERE,
  author =       "Christopher R. Aberger and Andrew Lamb and Susan Tu
                 and Andres N{\"o}tzli and Kunle Olukotun and
                 Christopher R{\'e}",
  title =        "{EmptyHeaded}: a Relational Engine for Graph
                 Processing",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "20:1--20:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3129246",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "There are two types of high-performance graph
                 processing engines: low- and high-level engines.
                 Low-level engines (Galois, PowerGraph, Snap) provide
                 optimized data structures and computation models but
                 require users to write low-level imperative code, hence
                 ensuring that efficiency is the burden of the user. In
                 high-level engines, users write in query languages like
                 datalog (SociaLite) or SQL (Grail). High-level engines
                 are easier to use but are orders of magnitude slower
                 than the low-level graph engines. We present
                 EmptyHeaded, a high-level engine that supports a rich
                 datalog-like query language and achieves performance
                 comparable to that of low-level engines. At the core of
                 EmptyHeaded's design is a new class of join algorithms
                 that satisfy strong theoretical guarantees, but have
                 thus far not achieved performance comparable to that of
                 specialized graph processing engines. To achieve high
                 performance, EmptyHeaded introduces a new join engine
                 architecture, including a novel query optimizer and
                 execution engine that leverage single-instruction
                 multiple data (SIMD) parallelism. With this
                 architecture, EmptyHeaded outperforms high-level
                 approaches by up to three orders of magnitude on graph
                 pattern queries, PageRank, and Single-Source Shortest
                 Paths (SSSP) and is an order of magnitude faster than
                 many low-level baselines. We validate that EmptyHeaded
                 competes with the best-of-breed low-level engine
                 (Galois), achieving comparable performance on PageRank
                 and at most 3$ \times $ worse performance on SSSP.
                 Finally, we show that the EmptyHeaded design can easily
                 be extended to accommodate a standard resource
                 description framework (RDF) workload, the LUBM
                 benchmark. On the LUBM benchmark, we show that
                 EmptyHeaded can compete with and sometimes outperform
                 two high-level, but specialized RDF baselines
                 (TripleBit and RDF-3X), while outperforming MonetDB by
                 up to three orders of magnitude and LogicBlox by up to
                 two orders of magnitude.",
  acknowledgement = ack-nhfb,
  articleno =    "20",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2017:DQL,
  author =       "Marcelo Arenas and Martin Ugarte",
  title =        "Designing a Query Language for {RDF}: Marrying Open
                 and Closed Worlds",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "21:1--21:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3129247",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "When querying an Resource Description Framework (RDF)
                 graph, a prominent feature is the possibility of
                 extending the answer to a query with optional
                 information. However, the definition of this feature in
                 SPARQL-the standard RDF query language-has raised some
                 important issues. Most notably, the use of this feature
                 increases the complexity of the evaluation problem, and
                 its closed-world semantics is in conflict with the
                 underlying open-world semantics of RDF. Many approaches
                 for fixing such problems have been proposed, the most
                 prominent being the introduction of the semantic notion
                 of weakly monotone SPARQL query. Weakly monotone SPARQL
                 queries have shaped the class of queries that conform
                 to the open-world semantics of RDF. Unfortunately,
                 finding an effective way of restricting SPARQL to the
                 fragment of weakly monotone queries has proven to be an
                 elusive problem. In practice, the most widely adopted
                 fragment for writing SPARQL queries is based on the
                 syntactic notion of well designedness. This notion has
                 proven to be a good approach for writing SPARQL
                 queries, but its expressive power has yet to be fully
                 understood. The starting point of this article is to
                 understand the relation between well-designed queries
                 and the semantic notion of weak monotonicity. It is
                 known that every well-designed SPARQL query is weakly
                 monotone; as our first contribution we prove that the
                 converse does not hold, even if an extension of this
                 notion based on the use of disjunction is considered.
                 Given this negative result, we embark on the task of
                 defining syntactic fragments that are weakly monotone
                 and have higher expressive power than the fragment of
                 well-designed queries. To this end, we move to a more
                 general scenario where infinite RDF graphs are also
                 allowed, so interpolation techniques studied for
                 first-order logic can be applied. With the use of these
                 techniques, we are able to define a new operator for
                 SPARQL that gives rise to a query language with the
                 desired properties (over finite and infinite RDF
                 graphs). It should be noticed that every query in this
                 fragment is weakly monotone if we restrict the
                 semantics to finite RDF graphs. Moreover, we use this
                 result to provide a simple characterization of the
                 class of monotone CONSTRUCT queries, that is, the class
                 of SPARQL queries that produce RDF graphs as output.
                 Finally, we pinpoint the complexity of the evaluation
                 problem for the query languages identified in the
                 article.",
  acknowledgement = ack-nhfb,
  articleno =    "21",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Barany:2017:DPP,
  author =       "Vince B{\'a}r{\'a}ny and Balder {Ten Cate} and Benny
                 Kimelfeld and Dan Olteanu and Zografoula Vagena",
  title =        "Declarative Probabilistic Programming with {Datalog}",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "22:1--22:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3132700",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Probabilistic programming languages are used for
                 developing statistical models. They typically consist
                 of two components: a specification of a stochastic
                 process (the prior) and a specification of observations
                 that restrict the probability space to a conditional
                 subspace (the posterior). Use cases of such formalisms
                 include the development of algorithms in machine
                 learning and artificial intelligence. In this article,
                 we establish a probabilistic-programming extension of
                 Datalog that, on the one hand, allows for defining a
                 rich family of statistical models, and on the other
                 hand retains the fundamental properties of
                 declarativity. Our proposed extension provides
                 mechanisms to include common numerical probability
                 functions; in particular, conclusions of rules may
                 contain values drawn from such functions. The semantics
                 of a program is a probability distribution over the
                 possible outcomes of the input database with respect to
                 the program. Observations are naturally incorporated by
                 means of integrity constraints over the extensional and
                 intensional relations. The resulting semantics is
                 robust under different chases and invariant to
                 rewritings that preserve logical equivalence.",
  acknowledgement = ack-nhfb,
  articleno =    "22",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Alvaro:2017:BCA,
  author =       "Peter Alvaro and Neil Conway and Joseph M. Hellerstein
                 and David Maier",
  title =        "{Blazes}: Coordination Analysis and Placement for
                 Distributed Programs",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "23:1--23:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3110214",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Distributed consistency is perhaps the most-discussed
                 topic in distributed systems today. Coordination
                 protocols can ensure consistency, but in practice they
                 cause undesirable performance unless used judiciously.
                 Scalable distributed architectures avoid coordination
                 whenever possible, but under-coordinated systems can
                 exhibit behavioral anomalies under fault, which are
                 often extremely difficult to debug. This raises
                 significant challenges for distributed system
                 architects and developers. In this article, we present
                 B lazes, a cross-platform program analysis framework
                 that (a) identifies program locations that require
                 coordination to ensure consistent executions, and (b)
                 automatically synthesizes application-specific
                 coordination code that can significantly outperform
                 general-purpose techniques. We present two case
                 studies, one using annotated programs in the Twitter
                 Storm system and another using the Bloom declarative
                 language.",
  acknowledgement = ack-nhfb,
  articleno =    "23",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Colazzo:2017:LTM,
  author =       "Dario Colazzo and Giorgio Ghelli and Carlo Sartiani",
  title =        "Linear Time Membership in a Class of Regular
                 Expressions with Counting, Interleaving, and Unordered
                 Concatenation",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "24:1--24:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3132701",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Regular Expressions (REs) are ubiquitous in database
                 and programming languages. While many applications make
                 use of REs extended with interleaving (shuffle) and
                 unordered concatenation operators, this extension badly
                 affects the complexity of basic operations, and,
                 especially, makes membership NP-hard, which is
                 unacceptable in most practical scenarios. In this
                 article, we study the problem of membership checking
                 for a restricted class of these extended REs, called
                 conflict-free REs, which are expressive enough to cover
                 the vast majority of real-world applications. We
                 present several polynomial algorithms for membership
                 checking over conflict-free REs. The algorithms are all
                 polynomial and differ in terms of adopted optimization
                 techniques and in the kind of supported operators. As a
                 particular application, we generalize the approach to
                 check membership of Extensible Markup Language trees
                 into a class of EDTDs (Extended Document Type
                 Definitions) that models the crucial aspects of DTDs
                 (Document Type Definitions) and XSD (XML Schema
                 Definitions) schemas. Results about an extensive
                 experimental analysis validate the efficiency of the
                 presented membership checking techniques.",
  acknowledgement = ack-nhfb,
  articleno =    "24",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Zhang:2017:PPD,
  author =       "Jun Zhang and Graham Cormode and Cecilia M. Procopiuc
                 and Divesh Srivastava and Xiaokui Xiao",
  title =        "{PrivBayes}: Private Data Release via {Bayesian}
                 Networks",
  journal =      j-TODS,
  volume =       "42",
  number =       "4",
  pages =        "25:1--25:??",
  month =        nov,
  year =         "2017",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3134428",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Nov 18 10:18:22 MST 2017",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Privacy-preserving data publishing is an important
                 problem that has been the focus of extensive study. The
                 state-of-the-art solution for this problem is
                 differential privacy, which offers a strong degree of
                 privacy protection without making restrictive
                 assumptions about the adversary. Existing techniques
                 using differential privacy, however, cannot effectively
                 handle the publication of high-dimensional data. In
                 particular, when the input dataset contains a large
                 number of attributes, existing methods require
                 injecting a prohibitive amount of noise compared to the
                 signal in the data, which renders the published data
                 next to useless. To address the deficiency of the
                 existing methods, this paper presents PrivBayes, a
                 differentially private method for releasing
                 high-dimensional data. Given a dataset D, PrivBayes
                 first constructs a Bayesian network N, which (i)
                 provides a succinct model of the correlations among the
                 attributes in D and (ii) allows us to approximate the
                 distribution of data in D using a set P of
                 low-dimensional marginals of D. After that, PrivBayes
                 injects noise into each marginal in P to ensure
                 differential privacy and then uses the noisy marginals
                 and the Bayesian network to construct an approximation
                 of the data distribution in D. Finally, PrivBayes
                 samples tuples from the approximate distribution to
                 construct a synthetic dataset, and then releases the
                 synthetic data. Intuitively, PrivBayes circumvents the
                 curse of dimensionality, as it injects noise into the
                 low-dimensional marginals in P instead of the
                 high-dimensional dataset D. Private construction of
                 Bayesian networks turns out to be significantly
                 challenging, and we introduce a novel approach that
                 uses a surrogate function for mutual information to
                 build the model more accurately. We experimentally
                 evaluate PrivBayes on real data and demonstrate that it
                 significantly outperforms existing solutions in terms
                 of accuracy.",
  acknowledgement = ack-nhfb,
  articleno =    "25",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2018:EUE,
  author =       "Christian S. Jensen",
  title =        "Editorial: Updates to the {Editorial Board}",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "1:1--1:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3183376",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Demertzis:2018:PPR,
  author =       "Ioannis Demertzis and Stavros Papadopoulos and
                 Odysseas Papapetrou and Antonios Deligiannakis and
                 Minos Garofalakis and Charalampos Papamanthou",
  title =        "Practical Private Range Search in Depth",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "2:1--2:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3167971",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We consider a data owner that outsources its dataset
                 to an untrusted server. The owner wishes to enable the
                 server to answer range queries on a single attribute,
                 without compromising the privacy of the data and the
                 queries. There are several schemes on ``practical''
                 private range search (mainly in database venues) that
                 attempt to strike a trade-off between efficiency and
                 security. Nevertheless, these methods either lack
                 provable security guarantees or permit unacceptable
                 privacy leakages. In this article, we take an
                 interdisciplinary approach, which combines the rigor of
                 security formulations and proofs with efficient data
                 management techniques. We construct a wide set of novel
                 schemes with realistic security/performance trade-offs,
                 adopting the notion of Searchable Symmetric Encryption
                 (SSE), primarily proposed for keyword search. We reduce
                 range search to multi-keyword search using
                 range-covering techniques with tree-like indexes, and
                 formalize the problem as Range Searchable Symmetric
                 Encryption (RSSE). We demonstrate that, given any
                 secure SSE scheme, the challenge boils down to (i)
                 formulating leakages that arise from the index
                 structure and (ii) minimizing false positives incurred
                 by some schemes under heavy data skew. We also explain
                 an important concept in the recent SSE bibliography,
                 namely locality, and design generic and specialized
                 ways to attribute locality to our RSSE schemes.
                 Moreover, we are the first to devise secure schemes for
                 answering range aggregate queries, such as range sums
                 and range min/max. We analytically detail the
                 superiority of our proposals over prior work and
                 experimentally confirm their practicality.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Chung:2018:EIU,
  author =       "Yeounoh Chung and Michael Lind Mortensen and Carsten
                 Binnig and Tim Kraska",
  title =        "Estimating the Impact of Unknown Unknowns on Aggregate
                 Query Results",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "3:1--3:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3167970",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "It is common practice for data scientists to acquire
                 and integrate disparate data sources to achieve higher
                 quality results. But even with a perfectly cleaned and
                 merged data set, two fundamental questions remain: (1)
                 Is the integrated data set complete? and (2) What is
                 the impact of any unknown (i.e., unobserved) data on
                 query results? In this work, we develop and analyze
                 techniques to estimate the impact of the unknown data
                 (a.k.a., unknown unknowns ) on simple aggregate
                 queries. The key idea is that the overlap between
                 different data sources enables us to estimate the
                 number and values of the missing data items. Our main
                 techniques are parameter-free and do not assume prior
                 knowledge about the distribution; we also propose a
                 parametric model that can be used instead when the data
                 sources are imbalanced. Through a series of
                 experiments, we show that estimating the impact of
                 unknown unknowns is invaluable to better assess the
                 results of aggregate queries over integrated data
                 sources.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Shaikhha:2018:BEQ,
  author =       "Amir Shaikhha and Yannis Klonatos and Christoph Koch",
  title =        "Building Efficient Query Engines in a High-Level
                 Language",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "4:1--4:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3183653",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Abstraction without regret refers to the vision of
                 using high-level programming languages for systems
                 development without experiencing a negative impact on
                 performance. A database system designed according to
                 this vision offers both increased productivity and high
                 performance instead of sacrificing the former for the
                 latter as is the case with existing, monolithic
                 implementations that are hard to maintain and extend.
                 In this article, we realize this vision in the domain
                 of analytical query processing. We present LegoBase, a
                 query engine written in the high-level programming
                 language Scala. The key technique to regain efficiency
                 is to apply generative programming: LegoBase performs
                 source-to-source compilation and optimizes database
                 systems code by converting the high-level Scala code to
                 specialized, low-level C code. We show how generative
                 programming allows to easily implement a wide spectrum
                 of optimizations, such as introducing data partitioning
                 or switching from a row to a column data layout, which
                 are difficult to achieve with existing low-level query
                 compilers that handle only queries. We demonstrate that
                 sufficiently powerful abstractions are essential for
                 dealing with the complexity of the optimization effort,
                 shielding developers from compiler internals and
                 decoupling individual optimizations from each other. We
                 evaluate our approach with the TPC-H benchmark and show
                 that (a) with all optimizations enabled, our
                 architecture significantly outperforms a commercial
                 in-memory database as well as an existing query
                 compiler. (b) Programmers need to provide just a few
                 hundred lines of high-level code for implementing the
                 optimizations, instead of complicated low-level code
                 that is required by existing query compilation
                 approaches. (c) These optimizations may potentially
                 come at the cost of using more system memory for
                 improved performance. (d) The compilation overhead is
                 low compared to the overall execution time, thus making
                 our approach usable in practice for compiling query
                 engines.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Libkin:2018:TNA,
  author =       "Leonid Libkin and Juan L. Reutter and Adri{\'a}n Soto
                 and Domagoj Vrgoc",
  title =        "{TriAL}: a Navigational Algebra for {RDF}
                 Triplestores",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "5:1--5:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3154385",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/string-matching.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Navigational queries over RDF data are viewed as one
                 of the main applications of graph query languages, and
                 yet the standard model of graph databases ---
                 essentially labeled graphs --- is different from the
                 triples-based model of RDF. While encodings of RDF
                 databases into graph data exist, we show that even the
                 most natural ones are bound to lose some functionality
                 when used in conjunction with graph query
                 languages. The solution is to work directly with
                 triples, but then many properties taken for granted in
                 the graph database context (e.g., reachability) lose
                 their natural meaning. Our goal is to introduce
                 languages that work directly over triples and are
                 closed, i.e., they produce sets of triples, rather than
                 graphs. Our basic language is called TriAL, or Triple
                 Algebra: it guarantees closure properties by replacing
                 the product with a family of join operations. We extend
                 TriAL with recursion and explain why such an extension
                 is more intricate for triples than for graphs. We
                 present a declarative language, namely a fragment of
                 datalog, capturing the recursive algebra. For both
                 languages, the combined complexity of query evaluation
                 is given by low-degree polynomials. We compare our
                 language with previously studied graph query languages
                 such as adaptations of XPath, regular path queries, and
                 nested regular expressions; many of these languages are
                 subsumed by the recursive triple algebra. We also
                 provide an implementation of recursive TriAL on top of
                 a relational query engine, and we show its usefulness
                 by running a wide array of navigational queries over
                 real-world RDF data, while at the same time testing how
                 our implementation compares to existing RDF systems.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Cao:2018:BQR,
  author =       "Yang Cao and Wenfei Fan and Floris Geerts and Ping
                 Lu",
  title =        "Bounded Query Rewriting Using Views",
  journal =      j-TODS,
  volume =       "43",
  number =       "1",
  pages =        "6:1--6:??",
  month =        apr,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3183673",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Apr 11 18:02:25 MDT 2018",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "A query $Q$ in a language $L$ has a bounded rewriting
                 using a set of $L$-definable views if there exists a
                 query $ Q^'$ in $L$ such that given any dataset $D$, $
                 Q(D)$ can be computed by $ Q_'$ that accesses only
                 cached views and a small fraction $ D_Q$ of $D$. We
                 consider datasets $D$ that satisfy a set of access
                 constraints, which are a combination of simple
                 cardinality constraints and associated indices, such
                 that the size $ |D_Q|$ of $ D_Q$ and the time to
                 identify $ D_Q$ are independent of $ |D|$, no matter
                 how big $D$ is. In this article, we study the problem
                 for deciding whether a query has a bounded rewriting
                 given a set $V$ of views and a set $A$ of access
                 constraints. We establish the complexity of the problem
                 for various query languages $L$, from $
                 \Sigma_3^p$-complete for conjunctive queries (CQ) to
                 undecidable for relational algebra (FO). We show that
                 the intractability for CQ is rather robust even for
                 acyclic CQ with fixed $V$ and $A$, and characterize
                 when the problem is in PTIME. To make practical use of
                 bounded rewriting, we provide an effective syntax for
                 FO queries that have a bounded rewriting. The syntax
                 characterizes a key subclass of such queries without
                 sacrificing the expressive power, and can be checked in
                 PTIME. Finally, we investigate $ L_1$-to-$ L_2$ bounded
                 rewriting, when Q in $ L_1$ is allowed to be rewritten
                 into a query $ Q^'$ in another language $ L_2$. We show
                 that this relaxation does not simplify the analysis of
                 bounded query rewriting using views.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Berkholz:2018:AFQ,
  author =       "Christoph Berkholz and Jens Keppeler and Nicole
                 Schweikardt",
  title =        "Answering {FO+MOD} Queries under Updates on Bounded
                 Degree Databases",
  journal =      j-TODS,
  volume =       "43",
  number =       "2",
  pages =        "7:1--7:??",
  month =        sep,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3232056",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:12 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "We investigate the query evaluation problem for fixed
                 queries over fully dynamic databases, where tuples can
                 be inserted or deleted. The task is to design a dynamic
                 algorithm that immediately reports the new result of a
                 fixed query after every database update. We consider
                 queries in first-order logic (FO) and its extension
                 with modulo-counting quantifiers (FO+MOD) and show that
                 they can be efficiently evaluated under updates,
                 provided that the dynamic database does not exceed a
                 certain degree bound. In particular, we construct a
                 data structure that allows us to answer a Boolean
                 FO+MOD query and to compute the size of the result of a
                 non-Boolean query within constant time after every
                 database update. Furthermore, after every database
                 update, we can update the data structure in constant
                 time such that afterwards we are able to test within
                 constant time for a given tuple whether or not it
                 belongs to the query result, to enumerate all tuples in
                 the new query result, and to enumerate the difference
                 between the old and the new query result with constant
                 delay between the output tuples. The preprocessing time
                 needed to build the data structure is linear in the
                 size of the database. Our results extend earlier work
                 on the evaluation of first-order queries on static
                 databases of bounded degree and rely on an effective
                 Hanf normal form for FO+MOD recently obtained by
                 Heimberg, Kuske, and Schweikardt (LICS 2016).",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Barcelo:2018:EES,
  author =       "Pablo Barcel{\'o} and Markus Kr{\"o}ll and Reinhard
                 Pichler and Sebastian Skritek",
  title =        "Efficient Evaluation and Static Analysis for
                 Well-Designed Pattern Trees with Projection",
  journal =      j-TODS,
  volume =       "43",
  number =       "2",
  pages =        "8:1--8:??",
  month =        sep,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3233983",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:12 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Conjunctive queries (CQs) fail to provide an answer
                 when the pattern described by the query does not
                 exactly match the data. CQs might thus be too
                 restrictive as a querying mechanism when data is
                 semistructured or incomplete. The semantic web
                 therefore provides a formalism-known as (projected)
                 well-designed pattern trees (pWDPTs)-that tackles this
                 problem: pWDPTs allow us to formulate queries that
                 match parts of the query over the data if available,
                 but do not ignore answers of the remaining query
                 otherwise. Here we abstract away the specifics of
                 semantic web applications and study pWDPTs over
                 arbitrary relational schemas. Since the language of
                 pWDPTs subsumes CQs, their evaluation problem is
                 intractable. We identify structural properties of
                 pWDPTs that lead to (fixed-parameter) tractability of
                 various variants of the evaluation problem. We also
                 show that checking if a pWDPT is equivalent to one in
                 our tractable class is in 2EXPTIME. As a corollary, we
                 obtain fixed-parameter tractability of evaluation for
                 pWDPTs with such good behavior. Our techniques also
                 allow us to develop a theory of approximations for
                 pWDPTs.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Lazerson:2018:LMD,
  author =       "Arnon Lazerson and Daniel Keren and Assaf Schuster",
  title =        "Lightweight Monitoring of Distributed Streams",
  journal =      j-TODS,
  volume =       "43",
  number =       "2",
  pages =        "9:1--9:??",
  month =        sep,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3226113",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:12 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "As data becomes dynamic, large, and distributed, there
                 is increasing demand for what have become known as
                 distributed stream algorithms. Since continuously
                 collecting the data to a central server and processing
                 it there is infeasible, a common approach is to define
                 local conditions at the distributed nodes, such that-as
                 long as they are maintained-some desirable global
                 condition holds. Previous methods derived local
                 conditions focusing on communication efficiency. While
                 proving very useful for reducing the communication
                 volume, these local conditions often suffer from heavy
                 computational burden at the nodes. The computational
                 complexity of the local conditions affects both the
                 runtime and the energy consumption. These are
                 especially critical for resource-limited devices like
                 smartphones and sensor nodes. Such devices are becoming
                 more ubiquitous due to the recent trend toward smart
                 cities and the Internet of Things. To accommodate for
                 high data rates and limited resources of these devices,
                 it is crucial that the local conditions be quickly and
                 efficiently evaluated. Here we propose a novel
                 approach, designated CB (for Convex/Concave Bounds). CB
                 defines local conditions using suitably chosen convex
                 and concave functions. Lightweight and simple, these
                 local conditions can be rapidly checked on the fly.
                 CB's superiority over the state-of-the-art is
                 demonstrated in its reduced runtime and power
                 consumption, by up to six orders of magnitude in some
                 cases. As an added bonus, CB also reduced communication
                 overhead in all the tested application scenarios.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Qi:2018:RQU,
  author =       "Jianzhong Qi and Fei Zuo and Hanan Samet and Jia Cheng
                 Yao",
  title =        "{$K$}-Regret Queries Using Multiplicative Utility
                 Functions",
  journal =      j-TODS,
  volume =       "43",
  number =       "2",
  pages =        "10:1--10:??",
  month =        sep,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3230634",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:12 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The k -regret query aims to return a size- k subset S
                 of a database D such that, for any query user that
                 selects a data object from this size- k subset S rather
                 than from database D, her regret ratio is minimized.
                 The regret ratio here is modeled by the relative
                 difference in the optimality between the locally
                 optimal object in S and the globally optimal object in
                 D. The optimality of a data object in turn is modeled
                 by a utility function of the query user. Unlike
                 traditional top- k queries, the k -regret query does
                 not minimize the regret ratio for a specific utility
                 function. Instead, it considers a family of infinite
                 utility functions F, and aims to find a size- k subset
                 that minimizes the maximum regret ratio of any utility
                 function in F. Studies on k -regret queries have
                 focused on the family of additive utility functions,
                 which have limitations in modeling individuals'
                 preferences and decision-making processes, especially
                 for a common observation called the diminishing
                 marginal rate of substitution (DMRS). We introduce k
                 -regret queries with multiplicative utility functions,
                 which are more expressive in modeling the DMRS, to
                 overcome those limitations. We propose a query
                 algorithm with bounded regret ratios. To showcase the
                 applicability of the algorithm, we apply it to a
                 special family of multiplicative utility functions, the
                 Cobb--Douglas family of utility functions, and a
                 closely related family of utility functions, the
                 Constant Elasticity of Substitution family of utility
                 functions, both of which are frequently used utility
                 functions in microeconomics. After a further study of
                 the query properties, we propose a heuristic algorithm
                 that produces even smaller regret ratios in practice.
                 Extensive experiments on the proposed algorithms
                 confirm that they consistently achieve small maximum
                 regret ratios.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kimelfeld:2018:RFC,
  author =       "Benny Kimelfeld and Christopher R{\'e}",
  title =        "A Relational Framework for Classifier Engineering",
  journal =      j-TODS,
  volume =       "43",
  number =       "3",
  pages =        "11:1--11:??",
  month =        nov,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3268931",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:13 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In the design of analytical procedures and machine
                 learning solutions, a critical and time-consuming task
                 is that of feature engineering, for which various
                 recipes and tooling approaches have been developed. In
                 this article, we embark on the establishment of
                 database foundations for feature engineering. We
                 propose a formal framework for classification in the
                 context of a relational database. The goal of this
                 framework is to open the way to research and techniques
                 to assist developers with the task of feature
                 engineering by utilizing the database's modeling and
                 understanding of data and queries and by deploying the
                 well-studied principles of database management. As a
                 first step, we demonstrate the usefulness of this
                 framework by formally defining three key algorithmic
                 challenges. The first challenge is that of
                 separability, which is the problem of determining the
                 existence of feature queries that agree with the
                 training examples. The second is that of evaluating the
                 VC dimension of the model class with respect to a given
                 sequence of feature queries. The third challenge is
                 identifiability, which is the task of testing for a
                 property of independence among features that are
                 represented as database queries. We give preliminary
                 results on these challenges for the case where features
                 are defined by means of conjunctive queries, and, in
                 particular, we study the implication of various
                 traditional syntactic restrictions on the inherent
                 computational complexity.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Schwentick:2018:DCU,
  author =       "Thomas Schwentick and Nils Vortmeier and Thomas
                 Zeume",
  title =        "Dynamic Complexity under Definable Changes",
  journal =      j-TODS,
  volume =       "43",
  number =       "3",
  pages =        "12:1--12:??",
  month =        nov,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3241040",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:13 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In the setting of dynamic complexity, the goal of a
                 dynamic program is to maintain the result of a fixed
                 query for an input database that is subject to changes,
                 possibly using additional auxiliary relations. In other
                 words, a dynamic program updates a materialized view
                 whenever a base relation is changed. The update of
                 query result and auxiliary relations is specified using
                 first-order logic or, equivalently, relational algebra.
                 The original framework by Patnaik and Immerman only
                 considers changes to the database that insert or delete
                 single tuples. This article extends the setting to
                 definable changes, also specified by first-order
                 queries on the database, and generalizes previous
                 maintenance results to these more expressive change
                 operations. More specifically, it is shown that the
                 undirected reachability query is first-order
                 maintainable under single-tuple changes and first-order
                 defined insertions, likewise the directed reachability
                 query for directed acyclic graphs is first-order
                 maintainable under insertions defined by
                 quantifier-free first-order queries. These results rely
                 on bounded bridge properties, which basically say that,
                 after an insertion of a defined set of edges, for each
                 connected pair of nodes there is some path with a
                 bounded number of new edges. While this bound can be
                 huge, in general, it is shown to be small for insertion
                 queries defined by unions of conjunctive queries. To
                 illustrate that the results for this restricted setting
                 could be practically relevant, they are complemented by
                 an experimental study that compares the performance of
                 dynamic programs with complex changes, dynamic programs
                 with single changes, and with recomputation from
                 scratch. The positive results are complemented by
                 several inexpressibility results. For example, it is
                 shown that-unlike for single-tuple insertions-dynamic
                 programs that maintain the reachability query under
                 definable, quantifier-free changes strictly need update
                 formulas with quantifiers. Finally, further positive
                 results unrelated to reachability are presented: it is
                 shown that for changes definable by parameter-free
                 first-order formulas, all LOGSPACE-definable (and even
                 AC$^1$-definable) queries can be maintained by
                 first-order dynamic programs.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Arenas:2018:ELQ,
  author =       "Marcelo Arenas and Georg Gottlob and Andreas Pieris",
  title =        "Expressive Languages for Querying the {Semantic Web}",
  journal =      j-TODS,
  volume =       "43",
  number =       "3",
  pages =        "13:1--13:??",
  month =        nov,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3238304",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:13 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "The problem of querying RDF data is a central issue
                 for the development of the Semantic Web. The query
                 language SPARQL has become the standard language for
                 querying RDF since its W3C standardization in 2008.
                 However, the 2008 version of this language missed some
                 important functionalities: reasoning capabilities to
                 deal with RDFS and OWL vocabularies, navigational
                 capabilities to exploit the graph structure of RDF
                 data, and a general form of recursion much needed to
                 express some natural queries. To overcome these
                 limitations, a new version of SPARQL, called SPARQL
                 1.1, was released in 2013, which includes entailment
                 regimes for RDFS and OWL vocabularies, and a mechanism
                 to express navigation patterns through regular
                 expressions. Unfortunately, there are a number of
                 useful navigation patterns that cannot be expressed in
                 SPARQL 1.1, and the language lacks a general mechanism
                 to express recursive queries. To the best of our
                 knowledge, no efficient RDF query language that
                 combines the above functionalities is known. It is the
                 aim of this work to fill this gap. To this end, we
                 focus on a core fragment of the OWL 2 QL profile of OWL
                 2 and show that every SPARQL query enriched with the
                 above features can be naturally translated into a query
                 expressed in a language that is based on an extension
                 of Datalog, which allows for value invention and
                 stratified negation. However, the query evaluation
                 problem for this language is highly intractable, which
                 is not surprising since it is expressive enough to
                 encode some inherently hard queries. We identify a
                 natural fragment of it, and we show it to be tractable
                 and powerful enough to define SPARQL queries enhanced
                 with the desired functionalities.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Polychroniou:2018:DJD,
  author =       "Orestis Polychroniou and Wangda Zhang and Kenneth A.
                 Ross",
  title =        "Distributed Joins and Data Placement for Minimal
                 Network Traffic",
  journal =      j-TODS,
  volume =       "43",
  number =       "3",
  pages =        "14:1--14:??",
  month =        nov,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3241039",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:13 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Network communication is the slowest component of many
                 operators in distributed parallel databases deployed
                 for large-scale analytics. Whereas considerable work
                 has focused on speeding up databases on modern
                 hardware, communication reduction has received less
                 attention. Existing parallel DBMSs rely on algorithms
                 designed for disks with minor modifications for
                 networks. A more complicated algorithm may burden the
                 CPUs but could avoid redundant transfers of tuples
                 across the network. We introduce track join, a new
                 distributed join algorithm that minimizes network
                 traffic by generating an optimal transfer schedule for
                 each distinct join key. Track join extends the
                 trade-off options between CPU and network. Track join
                 explicitly detects and exploits locality, also allowing
                 for advanced placement of tuples beyond hash
                 partitioning on a single attribute. We propose a novel
                 data placement algorithm based on track join that
                 minimizes the total network cost of multiple joins
                 across different dimensions in an analytical workload.
                 Our evaluation shows that track join outperforms hash
                 join on the most expensive queries of real workloads
                 regarding both network traffic and execution time.
                 Finally, we show that our data placement optimization
                 approach is both robust and effective in minimizing the
                 total network cost of joins in analytical workloads.",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Abuzaid:2018:MPA,
  author =       "Firas Abuzaid and Peter Bailis and Jialin Ding and
                 Edward Gan and Samuel Madden and Deepak Narayanan and
                 Kexin Rong and Sahaana Suri",
  title =        "{MacroBase}: Prioritizing Attention in Fast Data",
  journal =      j-TODS,
  volume =       "43",
  number =       "4",
  pages =        "15:1--15:??",
  month =        dec,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3276463",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "As data volumes continue to rise, manual inspection is
                 becoming increasingly untenable. In response, we
                 present MacroBase, a data analytics engine that
                 prioritizes end-user attention in high-volume fast data
                 streams. MacroBase enables efficient, accurate, and
                 modular analyses that highlight and aggregate important
                 and unusual behavior, acting as a search engine for
                 fast data. MacroBase is able to deliver
                 order-of-magnitude speedups over alternatives by
                 optimizing the combination of explanation (i.e.,
                 feature selection) and classification tasks and by
                 leveraging a new reservoir sampler and heavy-hitters
                 sketch specialized for fast data streams. As a result,
                 MacroBase delivers accurate results at speeds of up to
                 2M events per second per query on a single core. The
                 system has delivered meaningful results in production,
                 including at a telematics company monitoring hundreds
                 of thousands of vehicles.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dayan:2018:OBF,
  author =       "Niv Dayan and Manos Athanassoulis and Stratos Idreos",
  title =        "Optimal {Bloom} Filters and Adaptive Merging for
                 {LSM}-Trees",
  journal =      j-TODS,
  volume =       "43",
  number =       "4",
  pages =        "16:1--16:??",
  month =        dec,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3276980",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "In this article, we show that key-value stores backed
                 by a log-structured merge-tree (LSM-tree) exhibit an
                 intrinsic tradeoff between lookup cost, update cost,
                 and main memory footprint, yet all existing designs
                 expose a suboptimal and difficult to tune tradeoff
                 among these metrics. We pinpoint the problem to the
                 fact that modern key-value stores suboptimally co-tune
                 the merge policy, the buffer size, and the Bloom
                 filters' false-positive rates across the LSM-tree's
                 different levels. We present Monkey, an LSM-tree based
                 key-value store that strikes the optimal balance
                 between the costs of updates and lookups with any given
                 main memory budget. The core insight is that worst-case
                 lookup cost is proportional to the sum of the
                 false-positive rates of the Bloom filters across all
                 levels of the LSM-tree. Contrary to state-of-the-art
                 key-value stores that assign a fixed number of
                 bits-per-element to all Bloom filters, Monkey allocates
                 memory to filters across different levels so as to
                 minimize the sum of their false-positive rates. We show
                 analytically that Monkey reduces the asymptotic
                 complexity of the worst-case lookup I/O cost, and we
                 verify empirically using an implementation on top of
                 RocksDB that Monkey reduces lookup latency by an
                 increasing margin as the data volume grows (50--80\%
                 for the data sizes we experimented with). Furthermore,
                 we map the design space onto a closed-form model that
                 enables adapting the merging frequency and memory
                 allocation to strike the best tradeoff among lookup
                 cost, update cost and main memory, depending on the
                 workload (proportion of lookups and updates), the
                 dataset (number and size of entries), and the
                 underlying hardware (main memory available, disk vs.
                 flash). We show how to use this model to answer what-if
                 design questions about how changes in environmental
                 parameters impact performance and how to adapt the
                 design of the key-value store for optimal
                 performance.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Meneghetti:2018:LQA,
  author =       "Niccol{\`o} Meneghetti and Oliver Kennedy and Wolfgang
                 Gatterbauer",
  title =        "Learning From Query-Answers: a Scalable Approach to
                 Belief Updating and Parameter Learning",
  journal =      j-TODS,
  volume =       "43",
  number =       "4",
  pages =        "17:1--17:??",
  month =        dec,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3277503",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Tuple-independent and disjoint-independent
                 probabilistic databases (TI- and DI-PDBs) represent
                 uncertain data in a factorized form as a product of
                 independent random variables that represent either
                 tuples (TI-PDBs) or sets of tuples (DI-PDBs). When the
                 user submits a query, the database derives the marginal
                 probabilities of each output-tuple, exploiting the
                 underlying assumptions of statistical independence.
                 While query processing in TI- and DI-PDBs has been
                 studied extensively, limited research has been
                 dedicated to the problems of updating or deriving the
                 parameters from observations of query results.
                 Addressing this problem is the main focus of this
                 article. We first introduce Beta Probabilistic
                 Databases (B-PDBs), a generalization of TI-PDBs
                 designed to support both (i) belief updating and (ii)
                 parameter learning in a principled and scalable way.
                 The key idea of B-PDBs is to treat each parameter as a
                 latent, Beta-distributed random variable. We show how
                 this simple expedient enables both belief updating and
                 parameter learning in a principled way, without
                 imposing any burden on regular query processing.
                 Building on B-PDBs, we then introduce Dirichlet
                 Probabilistic Databases (D-PDBs), a generalization of
                 DI-PDBs with similar properties. We provide the
                 following key contributions for both B- and D-PDBs: (i)
                 We study the complexity of performing Bayesian belief
                 updates and devise efficient algorithms for certain
                 tractable classes of queries; (ii) we propose a soft-EM
                 algorithm for computing maximum-likelihood estimates of
                 the parameters; (iii) we present an algorithm for
                 efficiently computing conditional probabilities,
                 allowing us to efficiently implement B- and D-PDBs via
                 a standard relational engine; and (iv) we support our
                 conclusions with extensive experimental results.",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2018:PSG,
  author =       "Wenfei Fan and Wenyuan Yu and Jingbo Xu and Jingren
                 Zhou and Xiaojian Luo and Qiang Yin and Ping Lu and
                 Yang Cao and Ruiqi Xu",
  title =        "Parallelizing Sequential Graph Computations",
  journal =      j-TODS,
  volume =       "43",
  number =       "4",
  pages =        "18:1--18:??",
  month =        dec,
  year =         "2018",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3282488",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This article presents GRAPE, a parallel GRAPhEngine
                 for graph computations. GRAPE differs from prior
                 systems in its ability to parallelize existing
                 sequential graph algorithms as a whole, without the
                 need for recasting the entire algorithm into a new
                 model. Underlying GRAPE are a simple programming model
                 and a principled approach based on fixpoint computation
                 that starts with partial evaluation and uses an
                 incremental function as the intermediate consequence
                 operator. We show that users can devise existing
                 sequential graph algorithms with minor additions, and
                 GRAPE parallelizes the computation. Under a monotonic
                 condition, the GRAPE parallelization guarantees to
                 converge at correct answers as long as the sequential
                 algorithms are correct. Moreover, we show that
                 algorithms in MapReduce, BSP, and PRAM can be optimally
                 simulated on GRAPE. In addition to the ease of
                 programming, we experimentally verify that GRAPE
                 achieves comparable performance to the state-of-the-art
                 graph systems using real-life and synthetic graphs.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Kipf:2019:SAF,
  author =       "Andreas Kipf and Varun Pandey and Jan B{\"o}ttcher and
                 Lucas Braun and Thomas Neumann and Alfons Kemper",
  title =        "Scalable Analytics on Fast Data",
  journal =      j-TODS,
  volume =       "44",
  number =       "1",
  pages =        "1:1--1:??",
  month =        jan,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3283811",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Today's streaming applications demand increasingly
                 high event throughput rates and are often subject to
                 strict latency constraints. To allow for more complex
                 workloads, such as window-based aggregations, streaming
                 systems need to support stateful event processing. This
                 introduces new challenges for streaming engines as the
                 state needs to be maintained in a consistent and
                 durable manner and simultaneously accessed by complex
                 queries for real-time analytics. Modern streaming
                 systems, such as Apache Flink, do not allow for
                 efficiently exposing the state to analytical queries.
                 Thus, data engineers are forced to keep the state in
                 external data stores, which significantly increases the
                 latencies until events become visible to analytical
                 queries. Proprietary solutions have been created to
                 meet data freshness constraints. These solutions are
                 expensive, error-prone, and difficult to maintain.
                 Main-memory database systems, such as HyPer, achieve
                 extremely low query response times while maintaining
                 high update rates, which makes them well-suited for
                 analytical streaming workloads. In this article, we
                 explore extensions to database systems to match the
                 performance and usability of streaming systems.",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Li:2019:WJX,
  author =       "Feifei Li and Bin Wu and Ke Yi and Zhuoyue Zhao",
  title =        "Wander Join and {XDB}: Online Aggregation via Random
                 Walks",
  journal =      j-TODS,
  volume =       "44",
  number =       "1",
  pages =        "2:1--2:??",
  month =        jan,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3284551",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Joins are expensive, and online aggregation over joins
                 was proposed to mitigate the cost, which offers users a
                 nice and flexible tradeoff between query efficiency and
                 accuracy in a continuous, online fashion. However, the
                 state-of-the-art approach, in both internal and
                 external memory, is based on ripple join, which is
                 still very expensive and even needs unrealistic
                 assumptions (e.g., tuples in a table are stored in
                 random order). This article proposes a new approach,
                 the wander join algorithm, to the online aggregation
                 problem by performing random walks over the underlying
                 join graph. We also design an optimizer that chooses
                 the optimal plan for conducting the random walks
                 without having to collect any statistics a priori.
                 Compared with ripple join, wander join is particularly
                 efficient for equality joins involving multiple tables,
                 but also supports $ \theta $-joins. Selection
                 predicates and group-by clauses can be handled as well.
                 To demonstrate the usefulness of wander join, we have
                 designed and implemented XDB (approXimate DB) by
                 integrating wander join into various systems including
                 PostgreSQL, Spark, and a stand-alone plug-in version
                 using PL/SQL. The design and implementation of XDB has
                 demonstrated wander join's practicality in a
                 full-fledged database system. Extensive experiments
                 using the TPC-H benchmark have demonstrated the
                 superior performance of wander join over ripple join.",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bai:2019:HMD,
  author =       "Ran Bai and Wing Kai Hon and Eric Lo and Zhian He and
                 Kenny Zhu",
  title =        "Historic Moments Discovery in Sequence Data",
  journal =      j-TODS,
  volume =       "44",
  number =       "1",
  pages =        "3:1--3:??",
  month =        jan,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3276975",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Many emerging applications are based on finding
                 interesting subsequences from sequence data. Finding
                 ``prominent streaks,'' a set of the longest contiguous
                 subsequences with values all above (or below) a certain
                 threshold, from sequence data is one of that kind that
                 receives much attention. Motivated from real
                 applications, we observe that prominent streaks alone
                 are not insightful enough but require the discovery of
                 something we coined as ``historic moments'' as
                 companions. In this article, we present an algorithm to
                 efficiently compute historic moments from sequence
                 data. The algorithm is incremental and space optimal,
                 meaning that when facing new data arrival, it is able
                 to efficiently refresh the results by keeping minimal
                 information. Case studies show that historic moments
                 can significantly improve the insights offered by
                 prominent streaks alone. Furthermore, experiments show
                 that our algorithm can outperform the baseline in both
                 time and space.",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Alexandrov:2019:ROE,
  author =       "Alexander Alexandrov and Georgi Krastev and Volker
                 Markl",
  title =        "Representations and Optimizations for Embedded
                 Parallel Dataflow Languages",
  journal =      j-TODS,
  volume =       "44",
  number =       "1",
  pages =        "4:1--4:??",
  month =        jan,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3281629",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 29 17:36:16 MST 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "Parallel dataflow engines such as Apache Hadoop,
                 Apache Spark, and Apache Flink are an established
                 alternative to relational databases for modern data
                 analysis applications. A characteristic of these
                 systems is a scalable programming model based on
                 distributed collections and parallel transformations
                 expressed by means of second-order functions such as
                 map and reduce. Notable examples are Flink's DataSet
                 and Spark's RDD programming abstractions. These
                 programming models are realized as EDSLs-domain
                 specific languages embedded in a general-purpose host
                 language such as Java, Scala, or Python. This approach
                 has several advantages over traditional external DSLs
                 such as SQL or XQuery. First, syntactic constructs from
                 the host language (e.g., anonymous functions syntax,
                 value definitions, and fluent syntax via method
                 chaining) can be reused in the EDSL. This eases the
                 learning curve for developers already familiar with the
                 host language. Second, it allows for seamless
                 integration of library methods written in the host
                 language via the function parameters passed to the
                 parallel dataflow operators. This reduces the effort
                 for developing analytics dataflows that go beyond pure
                 SQL and require domain-specific logic. At the same
                 time, however, state-of-the-art parallel dataflow EDSLs
                 exhibit a number of shortcomings. First, one of the
                 main advantages of an external DSL such as SQL-the
                 high-level, declarative Select-From-Where syntax-is
                 either lost completely or mimicked in a non-standard
                 way. Second, execution aspects such as caching, join
                 order, and partial aggregation have to be decided by
                 the programmer. Optimizing them automatically is very
                 difficult due to the limited program context available
                 in the intermediate representation of the DSL. In this
                 article, we argue that the limitations listed above are
                 a side effect of the adopted type-based embedding
                 approach. As a solution, we propose an alternative EDSL
                 design based on quotations. We present a DSL embedded
                 in Scala and discuss its compiler pipeline,
                 intermediate representation, and some of the enabled
                 optimizations. We promote the algebraic type of bags in
                 union representation as a model for distributed
                 collections and its associated structural recursion
                 scheme and monad as a model for parallel collection
                 processing. At the source code level, Scala's
                 comprehension syntax over a bag monad can be used to
                 encode Select-From-Where expressions in a standard way.
                 At the intermediate representation level, maintaining
                 comprehensions as a first-class citizen can be used to
                 simplify the design and implementation of holistic
                 dataflow optimizations that accommodate for nesting and
                 control-flow. The proposed DSL design therefore
                 reconciles the benefits of embedded parallel dataflow
                 DSLs with the declarativity and optimization potential
                 of external DSLs like SQL.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Fan:2019:DG,
  author =       "Wenfei Fan and Ping Lu",
  title =        "Dependencies for Graphs",
  journal =      j-TODS,
  volume =       "44",
  number =       "2",
  pages =        "5:1--5:??",
  month =        apr,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3287285",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:20 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3287285",
  abstract =     "This article proposes a class of dependencies for
                 graphs, referred to as graph entity dependencies
                 (GEDs). A GED is defined as a combination of a graph
                 pattern and an attribute dependency. In a uniform
                 format, GEDs can express graph functional dependencies
                 with constant literals to catch inconsistencies, and
                 keys carrying id literals to identify entities
                 (vertices) in a graph. We revise the chase for GEDs and
                 prove its Church--Rosser property. We characterize GED
                 satisfiability and implication, and establish the
                 complexity of these problems and the validation problem
                 for GEDs, in the presence and absence of constant
                 literals and id literals. We also develop a sound,
                 complete and independent axiom system for finite
                 implication of GEDs. In addition, we extend GEDs with
                 built-in predicates or disjunctions, to strike a
                 balance between the expressive power and complexity. We
                 settle the complexity of the satisfiability,
                 implication, and validation problems for these
                 extensions.",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Hu:2019:OOM,
  author =       "Xiao Hu and Ke Yi and Yufei Tao",
  title =        "Output-Optimal Massively Parallel Algorithms for
                 Similarity Joins",
  journal =      j-TODS,
  volume =       "44",
  number =       "2",
  pages =        "6:1--6:??",
  month =        apr,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3311967",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:20 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3311967",
  abstract =     "Parallel join algorithms have received much attention
                 in recent years due to the rapid development of
                 massively parallel systems such as MapReduce and Spark.
                 In the database theory community, most efforts have
                 been focused on studying worst-case optimal algorithms.
                 However, the worst-case optimality of these join
                 algorithms relies on the hard instances having very
                 large output sizes. In the case of a two-relation join,
                 the hard instance is just a Cartesian product, with an
                 output size that is quadratic in the input size. In
                 practice, however, the output size is usually much
                 smaller. One recent parallel join algorithm by Beame et
                 al. has achieved output-optimality (i.e., its cost is
                 optimal in terms of both the input size and the output
                 size), but their algorithm only works for a 2-relation
                 equi-join and has some imperfections. In this article,
                 we first improve their algorithm to true optimality.
                 Then we design output-optimal algorithms for a large
                 class of similarity joins. Finally, we present a lower
                 bound, which essentially eliminates the possibility of
                 having output-optimal algorithms for any join on more
                 than two relations.",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Dautrich:2019:IIT,
  author =       "Jonathan L. {Dautrich, Jr.} and Chinya V.
                 Ravishankar",
  title =        "Inferring Insertion Times and Optimizing Error
                 Penalties in Time-decaying {Bloom} Filters",
  journal =      j-TODS,
  volume =       "44",
  number =       "2",
  pages =        "7:1--7:??",
  month =        apr,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3284552",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:20 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3284552",
  abstract =     "Current Bloom Filters tend to ignore Bayesian priors
                 as well as a great deal of useful information they
                 hold, compromising the accuracy of their responses.
                 Incorrect responses cause users to incur penalties that
                 are both application- and item-specific, but current
                 Bloom Filters are typically tuned only for static
                 penalties. Such shortcomings are problematic for all
                 Bloom Filter variants, but especially so for
                 Time-decaying Bloom Filters, in which the memory of
                 older items decays over time, causing both false
                 positives and false negatives. We address these issues
                 by introducing inferential filters, which integrate
                 Bayesian priors and information latent in filters to
                 make penalty-optimal, query-specific decisions. We also
                 show how to properly infer insertion times in such
                 filters. Our methods are general, but here we
                 illustrate their application to inferential
                 time-decaying filters to support novel query types and
                 sliding window queries with dynamic error penalties. We
                 present inferential versions of the Timing Bloom Filter
                 and Generalized Bloom Filter. Our experiments on real
                 and synthetic datasets show that our methods reduce
                 penalties for incorrect responses to sliding-window
                 queries in these filters by up to 70\% when penalties
                 are dynamic.",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Gummidi:2019:SSC,
  author =       "Srinivasa Raghavendra Bhuvan Gummidi and Xike Xie and
                 Torben Bach Pedersen",
  title =        "A Survey of Spatial Crowdsourcing",
  journal =      j-TODS,
  volume =       "44",
  number =       "2",
  pages =        "8:1--8:??",
  month =        apr,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3291933",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:20 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3291933",
  abstract =     "Widespread use of advanced mobile devices has led to
                 the emergence of a new class of crowdsourcing called
                 spatial crowdsourcing. Spatial crowdsourcing advances
                 the potential of a crowd to perform tasks related to
                 real-world scenarios involving physical locations,
                 which were not feasible with conventional crowdsourcing
                 methods. The main feature of spatial crowdsourcing is
                 the presence of spatial tasks that require workers to
                 be physically present at a particular location for task
                 fulfillment. Research related to this new paradigm has
                 gained momentum in recent years, necessitating a
                 comprehensive survey to offer a bird's-eye view of the
                 current state of spatial crowdsourcing literature. In
                 this article, we discuss the spatial crowdsourcing
                 infrastructure and identify the fundamental differences
                 between spatial and conventional crowdsourcing.
                 Furthermore, we provide a comprehensive view of the
                 existing literature by introducing a taxonomy,
                 elucidate the issues/challenges faced by different
                 components of spatial crowdsourcing, and suggest
                 potential research directions for the future.",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Damme:2019:CES,
  author =       "Patrick Damme and Annett Ungeth{\"u}m and Juliana
                 Hildebrandt and Dirk Habich and Wolfgang Lehner",
  title =        "From a Comprehensive Experimental Survey to a
                 Cost-based Selection Strategy for Lightweight Integer
                 Compression Algorithms",
  journal =      j-TODS,
  volume =       "44",
  number =       "3",
  pages =        "9:1--9:??",
  month =        jun,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3323991",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/datacompression.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3323991",
  abstract =     "Lightweight integer compression algorithms are
                 frequently applied in in-memory database systems to
                 tackle the growing gap between processor speed and main
                 memory bandwidth. In recent years, the vectorization of
                 basic techniques such as delta coding and null
                 suppression has considerably enlarged the corpus of
                 available algorithms. As a result, today there is a
                 large number of algorithms to choose from, while
                 different algorithms are tailored to different data
                 characteristics. However, a comparative evaluation of
                 these algorithms with different data and hardware
                 characteristics has never been sufficiently conducted
                 in the literature. To close this gap, we conducted an
                 exhaustive experimental survey by evaluating several
                 state-of-the-art lightweight integer compression
                 algorithms as well as cascades of basic techniques. We
                 systematically investigated the influence of data as
                 well as hardware properties on the performance and the
                 compression rates. The evaluated algorithms are based
                 on publicly available implementations as well as our
                 own vectorized reimplementations. We summarize our
                 experimental findings leading to several new insights
                 and to the conclusion that there is no single-best
                 algorithm. Moreover, in this article, we also introduce
                 and evaluate a novel cost model for the selection of a
                 suitable lightweight integer compression algorithm for
                 a given dataset.",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Bonifati:2019:IMS,
  author =       "Angela Bonifati and Ugo Comignani and Emmanuel Coquery
                 and Romuald Thion",
  title =        "Interactive Mapping Specification with Exemplar
                 Tuples",
  journal =      j-TODS,
  volume =       "44",
  number =       "3",
  pages =        "10:1--10:??",
  month =        jun,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3321485",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3321485",
  abstract =     "While schema mapping specification is a cumbersome
                 task for data curation specialists, it becomes
                 unfeasible for non-expert users, who are unacquainted
                 with the semantics and languages of the involved
                 transformations. In this article, we present an
                 interactive framework for schema mapping specification
                 suited for non-expert users. The underlying key
                 intuition is to leverage a few exemplar tuples to infer
                 the underlying mappings and iterate the inference
                 process via simple user interactions under the form of
                 Boolean queries on the validity of the initial exemplar
                 tuples. The approaches available so far are mainly
                 assuming pairs of complete universal data examples,
                 which can be solely provided by data curation experts,
                 or are limited to poorly expressive mappings. We
                 present a quasi-lattice-based exploration of the space
                 of all possible mappings that satisfy arbitrary user
                 exemplar tuples. Along the exploration, we challenge
                 the user to retain the mappings that fit the user's
                 requirements at best and to dynamically prune the
                 exploration space, thus reducing the number of user
                 interactions. We prove that after the refinement
                 process, the obtained mappings are correct and
                 complete. We present an extensive experimental analysis
                 devoted to measure the feasibility of our interactive
                 mapping strategies and the inherent quality of the
                 obtained mappings.",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Beedkar:2019:UFF,
  author =       "Kaustubh Beedkar and Rainer Gemulla and Wim Martens",
  title =        "A Unified Framework for Frequent Sequence Mining with
                 Subsequence Constraints",
  journal =      j-TODS,
  volume =       "44",
  number =       "3",
  pages =        "11:1--11:??",
  month =        jun,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3321486",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3321486",
  abstract =     "Frequent sequence mining methods often make use of
                 constraints to control which subsequences should be
                 mined. A variety of such subsequence constraints has
                 been studied in the literature, including length, gap,
                 span, regular-expression, and hierarchy constraints. In
                 this article, we show that many subsequence
                 constraints-including and beyond those considered in
                 the literature-can be unified in a single framework. A
                 unified treatment allows researchers to study jointly
                 many types of subsequence constraints (instead of each
                 one individually) and helps to improve usability of
                 pattern mining systems for practitioners. In more
                 detail, we propose a set of simple and intuitive
                 ``pattern expressions'' to describe subsequence
                 constraints and explore algorithms for efficiently
                 mining frequent subsequences under such general
                 constraints. Our algorithms translate pattern
                 expressions to succinct finite-state transducers, which
                 we use as computational model, and simulate these
                 transducers in a way suitable for frequent sequence
                 mining. Our experimental study on real-world datasets
                 indicates that our algorithms-although more general-are
                 efficient and, when used for sequence mining with prior
                 constraints studied in literature, competitive to (and
                 in some cases superior to) state-of-the-art specialized
                 methods.",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Deutsch:2019:VHA,
  author =       "Alin Deutsch and Yuliang Li and Victor Vianu",
  title =        "Verification of Hierarchical Artifact Systems",
  journal =      j-TODS,
  volume =       "44",
  number =       "3",
  pages =        "12:1--12:??",
  month =        jun,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3321487",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3321487",
  abstract =     "Data-driven workflows, of which IBM's Business
                 Artifacts are a prime exponent, have been successfully
                 deployed in practice, adopted in industrial standards,
                 and have spawned a rich body of research in academia,
                 focused primarily on static analysis. The present work
                 represents a significant advance on the problem of
                 artifact verification by considering a much richer and
                 more realistic model than in previous work,
                 incorporating core elements of IBM's successful
                 Guard-Stage-Milestone model. In particular, the model
                 features task hierarchy, concurrency, and richer
                 artifact data. It also allows database key and foreign
                 key dependencies, as well as arithmetic constraints.
                 The results show decidability of verification and
                 establish its complexity, making use of novel
                 techniques including a hierarchy of Vector Addition
                 Systems and a variant of quantifier elimination
                 tailored to our context.",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Seidemann:2019:CHP,
  author =       "Marc Seidemann and Nikolaus Glombiewski and Michael
                 K{\"o}rber and Bernhard Seeger",
  title =        "{ChronicleDB}: a High-Performance Event Store",
  journal =      j-TODS,
  volume =       "44",
  number =       "4",
  pages =        "13:1--13:??",
  month =        oct,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3342357",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3342357",
  abstract =     "Reactive security monitoring, self-driving cars, the
                 Internet of Things (IoT), and many other novel
                 applications require systems for both writing events
                 arriving at very high and fluctuating rates to
                 persistent storage as well as supporting analytical ad
                 hoc queries. As standard database systems are not
                 capable of delivering the required write performance,
                 log-based systems, key-value stores, and other
                 write-optimized data stores have emerged recently.
                 However, the drawbacks of these systems are a fair
                 query performance and the lack of suitable instant
                 recovery mechanisms in case of system failures. In this
                 article, we present ChronicleDB, a novel database
                 system with a storage layout tailored for high write
                 performance under fluctuating data rates and powerful
                 indexing capabilities to support a variety of queries.
                 In addition, ChronicleDB offers low-cost fault
                 tolerance and instant recovery within milliseconds.
                 Unlike previous work, ChronicleDB is designed either as
                 a serverless library to be tightly integrated in an
                 application or as a standalone database server. Our
                 results of an experimental evaluation with real and
                 synthetic data reveal that ChronicleDB clearly
                 outperforms competing systems with respect to both
                 write and query performance.",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Brijder:2019:EPQ,
  author =       "Robert Brijder and Floris Geerts and Jan {Van Den
                 Bussche} and Timmy Weerwag",
  title =        "On the Expressive Power of Query Languages for
                 Matrices",
  journal =      j-TODS,
  volume =       "44",
  number =       "4",
  pages =        "15:1--15:??",
  month =        oct,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3331445",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3331445",
  abstract =     "We investigate the expressive power of MATLANG, a
                 formal language for matrix manipulation based on common
                 matrix operations and linear algebra. The language can
                 be extended with the operation inv for inverting a
                 matrix. In MATLANG + inv, we can compute the transitive
                 closure of directed graphs, whereas we show that this
                 is not possible without inversion. Indeed, we show that
                 the basic language can be simulated in the relational
                 algebra with arithmetic operations, grouping, and
                 summation. We also consider an operation eigen for
                 diagonalizing a matrix. It is defined such that for
                 each eigenvalue a set of mutually orthogonal
                 eigenvectors is returned that span the eigenspace of
                 that eigenvalue. We show that inv can be expressed in
                 MATLANG + eigen. We put forward the open question
                 whether there are Boolean queries about matrices, or
                 generic queries about graphs, expressible in MATLANG +
                 eigen but not in MATLANG + inv. Finally, the evaluation
                 problem for MATLANG + eigen is shown to be complete for
                 the complexity class $ \exists R $.",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Martens:2019:DES,
  author =       "Wim Martens and Tina Trautner",
  title =        "Dichotomies for Evaluating Simple Regular Path
                 Queries",
  journal =      j-TODS,
  volume =       "44",
  number =       "4",
  pages =        "16:1--16:??",
  month =        oct,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3331446",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3331446",
  abstract =     "Regular path queries (RPQs) are a central component of
                 graph databases. We investigate decision and
                 enumeration problems concerning the evaluation of RPQs
                 under several semantics that have recently been
                 considered: arbitrary paths, shortest paths, paths
                 without node repetitions (simple paths), and paths
                 without edge repetitions (trails). Whereas arbitrary
                 and shortest paths can be dealt with efficiently,
                 simple paths and trails become computationally
                 difficult already for very small RPQs. We study RPQ
                 evaluation for simple paths and trails from a
                 parameterized complexity perspective and define a class
                 of simple transitive expressions that is prominent in
                 practice and for which we can prove dichotomies for the
                 evaluation problem. We observe that, even though simple
                 path and trail semantics are intractable for RPQs in
                 general, they are feasible for the vast majority of
                 RPQs that are used in practice. At the heart of this
                 study is a result of independent interest: the two
                 disjoint paths problem in directed graphs is W[1]-hard
                 if parameterized by the length of one of the two
                 paths.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Wang:2019:EAA,
  author =       "Sibo Wang and Renchi Yang and Runhui Wang and Xiaokui
                 Xiao and Zhewei Wei and Wenqing Lin and Yin Yang and
                 Nan Tang",
  title =        "Efficient Algorithms for Approximate Single-Source
                 Personalized {PageRank} Queries",
  journal =      j-TODS,
  volume =       "44",
  number =       "4",
  pages =        "18:1--18:??",
  month =        oct,
  year =         "2019",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3360902",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Oct 29 10:55:21 MDT 2019",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/pagerank.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/ft_gateway.cfm?id=3360902",
  abstract =     "Given a graph G, a source node s, and a target node t,
                 the personalized PageRank ( PPR ) of t with respect to
                 s is the probability that a random walk starting from s
                 terminates at t. An important variant of the PPR query
                 is single-source PPR ( SSPPR ), which enumerates all
                 nodes in G and returns the top- k nodes with the
                 highest PPR values with respect to a given source s.
                 PPR in general and SSPPR in particular have important
                 applications in web search and social networks, e.g.,
                 in Twitter's Who-To-Follow recommendation service.
                 However, PPR computation is known to be expensive on
                 large graphs and resistant to indexing. Consequently,
                 previous solutions either use heuristics, which do not
                 guarantee result quality, or rely on the strong
                 computing power of modern data centers, which is
                 costly. Motivated by this, we propose effective
                 index-free and index-based algorithms for approximate
                 PPR processing, with rigorous guarantees on result
                 quality. We first present FORA, an approximate SSPPR
                 solution that combines two existing methods-Forward
                 Push (which is fast but does not guarantee quality) and
                 Monte Carlo Random Walk (accurate but slow)-in a simple
                 and yet non-trivial way, leading to both high accuracy
                 and efficiency. Further, FORA includes a simple and
                 effective indexing scheme, as well as a module for top-
                 k selection with high pruning power. Extensive
                 experiments demonstrate that the proposed solutions are
                 orders of magnitude more efficient than their
                 respective competitors. Notably, on a billion-edge
                 Twitter dataset, FORA answers a top-500 approximate
                 SSPPR query within 1s, using a single commodity
                 server.",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "http://portal.acm.org/browse_dl.cfm?idx=J777",
}

@Article{Jensen:2020:EUE,
  author =       "Christian S. Jensen",
  title =        "Editorial: Updates to the {Editorial Board}",
  journal =      j-TODS,
  volume =       "45",
  number =       "1",
  pages =        "1e:1--1e:1",
  month =        mar,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3381020",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 4 07:13:27 MST 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3381020",
  acknowledgement = ack-nhfb,
  articleno =    "1e",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{McCamish:2020:GTA,
  author =       "Ben McCamish and Vahid Ghadakchi and Arash Termehchy
                 and Behrouz Touri and Eduardo Cotilla-Sanchez and Liang
                 Huang and Soravit Changpinyo",
  title =        "A Game-theoretic Approach to Data Interaction",
  journal =      j-TODS,
  volume =       "45",
  number =       "1",
  pages =        "1:1--1:44",
  month =        mar,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3351450",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 4 07:13:27 MST 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3351450",
  abstract =     "As most users do not precisely know the structure
                 and/or the content of databases, their queries do not
                 exactly reflect their information needs. The database
                 management system (DBMS) may interact with users and
                 use their feedback on the returned results \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Zhang:2020:FDD,
  author =       "Dan Zhang and Ryan McKenna and Ios Kotsogiannis and
                 George Bissias and Michael Hay and Ashwin
                 Machanavajjhala and Gerome Miklau",
  title =        "{$ \epsilon $KTELO}: a Framework for Defining
                 Differentially Private Computations",
  journal =      j-TODS,
  volume =       "45",
  number =       "1",
  pages =        "2:1--2:44",
  month =        mar,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3362032",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 4 07:13:27 MST 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3362032",
  abstract =     "The adoption of differential privacy is growing, but
                 the complexity of designing private, efficient, and
                 accurate algorithms is still high. We propose a novel
                 programming framework and system, $ \epsilon $KTELO for
                 implementing both existing and new privacy \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Florenzano:2020:EEA,
  author =       "Fernando Florenzano and Cristian Riveros and
                 Mart{\'\i}n Ugarte and Stijn Vansummeren and Domagoj
                 Vrgoc",
  title =        "Efficient Enumeration Algorithms for Regular Document
                 Spanners",
  journal =      j-TODS,
  volume =       "45",
  number =       "1",
  pages =        "3:1--3:42",
  month =        mar,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3351451",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 4 07:13:27 MST 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3351451",
  abstract =     "Regular expressions and automata models with capture
                 variables are core tools in rule-based information
                 extraction. These formalisms, also called regular
                 document spanners, use regular languages to locate the
                 data that a user wants to extract from a \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Livshits:2020:COR,
  author =       "Ester Livshits and Benny Kimelfeld and Sudeepa Roy",
  title =        "Computing Optimal Repairs for Functional
                 Dependencies",
  journal =      j-TODS,
  volume =       "45",
  number =       "1",
  pages =        "4:1--4:46",
  month =        mar,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3360904",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 4 07:13:27 MST 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3360904",
  abstract =     "We investigate the complexity of computing an optimal
                 repair of an inconsistent database, in the case where
                 integrity constraints are Functional Dependencies
                 (FDs). We focus on two types of repairs: an optimal
                 subset repair (optimal S-repair), which is \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Zhang:2020:SRF,
  author =       "Huanchen Zhang and Hyeontaek Lim and Viktor Leis and
                 David G. Andersen and Michael Kaminsky and Kimberly
                 Keeton and Andrew Pavlo",
  title =        "Succinct Range Filters",
  journal =      j-TODS,
  volume =       "45",
  number =       "2",
  pages =        "5:1--5:31",
  month =        jul,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3375660",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jul 19 08:40:15 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3375660",
  abstract =     "We present the Succinct Range Filter (SuRF), a fast
                 and compact data structure for approximate membership
                 tests. Unlike traditional Bloom filters, SuRF supports
                 both single-key lookups and common range queries:
                 open-range queries, closed-range queries, \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fan:2020:AAP,
  author =       "Wenfei Fan and Ping Lu and Wenyuan Yu and Jingbo Xu
                 and Qiang Yin and Xiaojian Luo and Jingren Zhou and
                 Ruochun Jin",
  title =        "Adaptive Asynchronous Parallelization of Graph
                 Algorithms",
  journal =      j-TODS,
  volume =       "45",
  number =       "2",
  pages =        "6:1--6:45",
  month =        jul,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3397491",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jul 19 08:40:15 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3397491",
  abstract =     "This article proposes an Adaptive Asynchronous
                 Parallel (AAP) model for graph computations. As opposed
                 to Bulk Synchronous Parallel (BSP) and Asynchronous
                 Parallel (AP) models, AAP reduces both stragglers and
                 stale computations by dynamically adjusting \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Khamis:2020:LMR,
  author =       "Mahmoud Abo Khamis and Hung Q. Ngo and Xuanlong Nguyen
                 and Dan Olteanu and Maximilian Schleich",
  title =        "Learning Models over Relational Data Using Sparse
                 Tensors and Functional Dependencies",
  journal =      j-TODS,
  volume =       "45",
  number =       "2",
  pages =        "7:1--7:66",
  month =        jul,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3375661",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jul 19 08:40:15 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3375661",
  abstract =     "Integrated solutions for analytics over relational
                 databases are of great practical importance as they
                 avoid the costly repeated loop data scientists have to
                 deal with on a daily basis: select features from data
                 residing in relational databases using \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Kolaitis:2020:LNT,
  author =       "Phokion G. Kolaitis and Reinhard Pichler and Emanuel
                 Sallinger and Vadim Savenkov",
  title =        "On the Language of Nested Tuple Generating
                 Dependencies",
  journal =      j-TODS,
  volume =       "45",
  number =       "2",
  pages =        "8:1--8:59",
  month =        jul,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3369554",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jul 19 08:40:15 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3369554",
  abstract =     "During the past 15 years, schema mappings have been
                 extensively used in formalizing and studying such
                 critical data interoperability tasks as data exchange
                 and data integration. Much of the work has focused on
                 GLAV mappings, i.e., schema mappings \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fan:2020:CNI,
  author =       "Wenfei Fan and Xueli Liu and Ping Lu and Chao Tian",
  title =        "Catching Numeric Inconsistencies in Graphs",
  journal =      j-TODS,
  volume =       "45",
  number =       "2",
  pages =        "9:1--9:47",
  month =        jul,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3385031",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jul 19 08:40:15 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/abs/10.1145/3385031",
  abstract =     "Numeric inconsistencies are common in real-life
                 knowledge bases and social networks. To catch such
                 errors, we extend graph functional dependencies with
                 linear arithmetic expressions and built-in comparison
                 predicates, referred to as numeric graph \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Jermaine:2020:E,
  author =       "Chris Jermaine",
  title =        "Editorial",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "10:1--10:1",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3417730",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3417730",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Kara:2020:MTQ,
  author =       "Ahmet Kara and Hung Q. Ngo and Milos Nikolic and Dan
                 Olteanu and Haozhe Zhang",
  title =        "Maintaining Triangle Queries under Updates",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "11:1--11:46",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3396375",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3396375",
  abstract =     "We consider the problem of incrementally maintaining
                 the triangle queries with arbitrary free variables
                 under single-tuple updates to the input relations. We
                 introduce an approach called IVM$\epsilon$ that
                 exhibits a trade-off between the update time, the
                 space, \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Shaikhha:2020:SIL,
  author =       "Amir Shaikhha and Mohammed Elseidy and Stephan Mihaila
                 and Daniel Espino and Christoph Koch",
  title =        "Synthesis of Incremental Linear Algebra Programs",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "12:1--12:44",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3385398",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3385398",
  abstract =     "This article targets the Incremental View Maintenance
                 (IVM) of sophisticated analytics (such as statistical
                 models, machine learning programs, and graph
                 algorithms) expressed as linear algebra programs. We
                 present LAGO, a unified framework for linear \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Schirmer:2020:EDM,
  author =       "Philipp Schirmer and Thorsten Papenbrock and Ioannis
                 Koumarelas and Felix Naumann",
  title =        "Efficient Discovery of Matching Dependencies",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "13:1--13:33",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3392778",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3392778",
  abstract =     "Matching dependencies (MDs) are data profiling results
                 that are often used for data integration, data
                 cleaning, and entity matching. They are a
                 generalization of functional dependencies (FDs)
                 matching similar rather than same elements. As their
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Qi:2020:PTS,
  author =       "Jianzhong Qi and Yufei Tao and Yanchuan Chang and Rui
                 Zhang",
  title =        "Packing {$R$}-trees with Space-filling Curves:
                 Theoretical Optimality, Empirical Efficiency, and
                 Bulk-loading Parallelizability",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "14:1--14:47",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3397506",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3397506",
  abstract =     "The massive amount of data and large variety of data
                 distributions in the big data era call for access
                 methods that are efficient in both query processing and
                 index management, and over both practical and
                 worst-case workloads. To address this need, we
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fan:2020:DGF,
  author =       "Wenfei Fan and Chunming Hu and Xueli Liu and Ping Lu",
  title =        "Discovering Graph Functional Dependencies",
  journal =      j-TODS,
  volume =       "45",
  number =       "3",
  pages =        "15:1--15:42",
  month =        sep,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3397198",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sat Sep 26 07:22:54 MDT 2020",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3397198",
  abstract =     "This article studies discovery of Graph Functional
                 Dependencies (GFDs), a class of functional dependencies
                 defined on graphs. We investigate the fixed-parameter
                 tractability of three fundamental problems related to
                 GFD discovery. We show that the \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Nakandala:2020:IAC,
  author =       "Supun Nakandala and Kabir Nagrecha and Arun Kumar and
                 Yannis Papakonstantinou",
  title =        "Incremental and Approximate Computations for
                 Accelerating Deep {CNN} Inference",
  journal =      j-TODS,
  volume =       "45",
  number =       "4",
  pages =        "16:1--16:42",
  month =        dec,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3397461",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Mar 28 09:41:59 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3397461",
  abstract =     "Deep learning now offers state-of-the-art accuracy for
                 many prediction tasks. A form of deep learning called
                 deep convolutional neural networks (CNNs) are
                 especially popular on image, video, and time series
                 data. Due to its high computational cost, CNN
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Khamis:2020:FAQ,
  author =       "Mahmoud Abo Khamis and Ryan R. Curtin and Benjamin
                 Moseley and Hung Q. Ngo and Xuanlong Nguyen and Dan
                 Olteanu and Maximilian Schleich",
  title =        "Functional Aggregate Queries with Additive
                 Inequalities",
  journal =      j-TODS,
  volume =       "45",
  number =       "4",
  pages =        "17:1--17:41",
  month =        dec,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3426865",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Mar 28 09:41:59 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3426865",
  abstract =     "Motivated by fundamental applications in databases and
                 relational machine learning, we formulate and study the
                 problem of answering functional aggregate queries (FAQ)
                 in which some of the input factors are defined by a
                 collection of additive \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "17",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Ciaccia:2020:FSD,
  author =       "Paolo Ciaccia and Davide Martinenghi",
  title =        "Flexible Skylines: Dominance for Arbitrary Sets of
                 Monotone Functions",
  journal =      j-TODS,
  volume =       "45",
  number =       "4",
  pages =        "18:1--18:45",
  month =        dec,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3406113",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Mar 28 09:41:59 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3406113",
  abstract =     "Skyline and ranking queries are two popular,
                 alternative ways of discovering interesting data in
                 large datasets. Skyline queries are simple to specify,
                 as they just return the set of all non-dominated
                 tuples, thereby providing an overall view of \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "18",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Zimanyi:2020:MMD,
  author =       "Esteban Zim{\'a}nyi and Mahmoud Sakr and Arthur
                 Lesuisse",
  title =        "{MobilityDB}: a Mobility Database Based on
                 {PostgreSQL} and {PostGIS}",
  journal =      j-TODS,
  volume =       "45",
  number =       "4",
  pages =        "19:1--19:42",
  month =        dec,
  year =         "2020",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3406534",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Mar 28 09:41:59 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3406534",
  abstract =     "Despite two decades of research in moving object
                 databases and a few research prototypes that have been
                 proposed, there is not yet a mainstream system targeted
                 for industrial use. In this article, we present
                 MobilityDB, a moving object database that \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "19",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Traub:2021:SGE,
  author =       "Jonas Traub and Philipp Marian Grulich and Alejandro
                 Rodr{\'\i}guez Cu{\'e}llar and Sebastian Bre{\ss} and
                 Asterios Katsifodimos and Tilmann Rabl and Volker
                 Markl",
  title =        "{Scotty}: General and Efficient Open-source Window
                 Aggregation for Stream Processing Systems",
  journal =      j-TODS,
  volume =       "46",
  number =       "1",
  pages =        "1:1--1:46",
  month =        apr,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3433675",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 15 14:48:28 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3433675",
  abstract =     "Window aggregation is a core operation in data stream
                 processing. Existing aggregation techniques focus on
                 reducing latency, eliminating redundant computations,
                 or minimizing memory usage. However, each technique
                 operates under different assumptions \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Amarilli:2021:CDE,
  author =       "Antoine Amarilli and Pierre Bourhis and Stefan Mengel
                 and Matthias Niewerth",
  title =        "Constant-Delay Enumeration for Nondeterministic
                 Document Spanners",
  journal =      j-TODS,
  volume =       "46",
  number =       "1",
  pages =        "2:1--2:30",
  month =        apr,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3436487",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 15 14:48:28 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3436487",
  abstract =     "We consider the information extraction framework known
                 as document spanners and study the problem of
                 efficiently computing the results of the extraction
                 from an input document, where the extraction task is
                 described as a sequential variable-set \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Mitchell:2021:ESM,
  author =       "Rory Mitchell and Eibe Frank and Geoffrey Holmes",
  title =        "An Empirical Study of Moment Estimators for Quantile
                 Approximation",
  journal =      j-TODS,
  volume =       "46",
  number =       "1",
  pages =        "3:1--3:21",
  month =        apr,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3442337",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 15 14:48:28 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3442337",
  abstract =     "We empirically evaluate lightweight moment estimators
                 for the single-pass quantile approximation problem,
                 including maximum entropy methods and orthogonal series
                 with Fourier, Cosine, Legendre, Chebyshev and Hermite
                 basis functions. We show how to apply \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Meduri:2021:EML,
  author =       "Venkata Vamsikrishna Meduri and Kanchan Chowdhury and
                 Mohamed Sarwat",
  title =        "Evaluation of Machine Learning Algorithms in
                 Predicting the Next {SQL} Query from the Future",
  journal =      j-TODS,
  volume =       "46",
  number =       "1",
  pages =        "4:1--4:46",
  month =        apr,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3442338",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Thu Apr 15 14:48:28 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3442338",
  abstract =     "Prediction of the next SQL query from the user, given
                 her sequence of queries until the current timestep,
                 during an ongoing interaction session of the user with
                 the database, can help in speculative query processing
                 and increased interactivity. While existing machine
                 learning-(ML) based approaches use recommender
                 systems to suggest relevant queries to a user, there
                 has been no exhaustive study on applying temporal
                 predictors to predict the next user issued
                 query.\par

                 In this work, we experimentally compare ML algorithms
                 in predicting the immediate next future query in an
                 interaction workload, given the current user query or
                 the sequence of queries in a user session thus far. As
                 a part of this, we propose the adaptation of two
                 powerful temporal predictors: (a) Recurrent Neural
                 Networks (RNNs) and (b) a Reinforcement Learning
                 approach called Q-Learning that uses Markov Decision
                 Processes. We represent each query as a comprehensive
                 set of fragment embeddings that not only captures the
                 SQL operators, attributes, and relations but also the
                 arithmetic comparison operators and constants that
                 occur in the query. Our experiments on two real-world
                 datasets show the effectiveness of temporal predictors
                 against the baseline recommender systems in predicting
                 the structural fragments in a query w.r.t. both quality
                 and time. Besides showing that RNNs can be used to
                 synthesize novel queries, we find that exact Q-Learning
                 outperforms RNNs despite predicting the next query
                 entirely from the historical query logs.",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Carmeli:2021:ECU,
  author =       "Nofar Carmeli and Markus Kr{\"o}ll",
  title =        "On the Enumeration Complexity of Unions of Conjunctive
                 Queries",
  journal =      j-TODS,
  volume =       "46",
  number =       "2",
  pages =        "5:1--5:41",
  month =        jun,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3450263",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jun 6 07:07:25 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3450263",
  abstract =     "We study the enumeration complexity of Unions of
                 Conjunctive Queries (UCQs). We aim to identify the UCQs
                 that are tractable in the sense that the answer tuples
                 can be enumerated with a linear preprocessing phase and
                 a constant delay between every \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Mhedhbi:2021:OOT,
  author =       "Amine Mhedhbi and Chathura Kankanamge and Semih
                 Salihoglu",
  title =        "Optimizing One-time and Continuous Subgraph Queries
                 using Worst-case Optimal Joins",
  journal =      j-TODS,
  volume =       "46",
  number =       "2",
  pages =        "6:1--6:45",
  month =        jun,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3446980",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jun 6 07:07:25 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3446980",
  abstract =     "We study the problem of optimizing one-time and
                 continuous subgraph queries using the new worst-case
                 optimal join plans. Worst-case optimal plans evaluate
                 queries by matching one query vertex at a time using
                 multiway intersections. The core problem in \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Wei:2021:EFD,
  author =       "Ziheng Wei and Sebastian Link",
  title =        "Embedded Functional Dependencies and Data-completeness
                 Tailored Database Design",
  journal =      j-TODS,
  volume =       "46",
  number =       "2",
  pages =        "7:1--7:46",
  month =        jun,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3450518",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jun 6 07:07:25 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3450518",
  abstract =     "We establish a principled schema design framework for
                 data with missing values. The framework is based on the
                 new notion of an embedded functional dependency, which
                 is independent of the interpretation of missing values,
                 able to express completeness and \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Chen:2021:GIE,
  author =       "Yangjun Chen and Gagandeep Singh",
  title =        "Graph Indexing for Efficient Evaluation of
                 Label-constrained Reachability Queries",
  journal =      j-TODS,
  volume =       "46",
  number =       "2",
  pages =        "8:1--8:50",
  month =        jun,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3451159",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Sun Jun 6 07:07:25 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3451159",
  abstract =     "Given a directed edge labeled graph G, to check
                 whether vertex v is reachable from vertex u under a
                 label set S is to know if there is a path from u to v
                 whose edge labels across the path are a subset of S.
                 Such a query is referred to as a label-. \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Trummer:2021:SRB,
  author =       "Immanuel Trummer and Junxiong Wang and Ziyun Wei and
                 Deepak Maram and Samuel Moseley and Saehan Jo and
                 Joseph Antonakakis and Ankush Rayabhari",
  title =        "{SkinnerDB}: Regret-bounded Query Evaluation via
                 Reinforcement Learning",
  journal =      j-TODS,
  volume =       "46",
  number =       "3",
  pages =        "9:1--9:45",
  month =        sep,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3464389",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Sep 29 06:47:23 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3464389",
  abstract =     "SkinnerDB uses reinforcement learning for reliable
                 join ordering, exploiting an adaptive processing engine
                 with specialized join algorithms and data structures.
                 It maintains no data statistics and uses no cost or
                 cardinality models. Also, it uses no \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Song:2021:SDC,
  author =       "Shaoxu Song and Fei Gao and Aoqian Zhang and Jianmin
                 Wang and Philip S. Yu",
  title =        "Stream Data Cleaning under Speed and Acceleration
                 Constraints",
  journal =      j-TODS,
  volume =       "46",
  number =       "3",
  pages =        "10:1--10:44",
  month =        sep,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3465740",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Sep 29 06:47:23 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3465740",
  abstract =     "Stream data are often dirty, for example, owing to
                 unreliable sensor reading or erroneous extraction of
                 stock prices. Most stream data cleaning approaches
                 employ a smoothing filter, which may seriously alter
                 the data without preserving the original \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Lin:2021:EBL,
  author =       "Xuelian Lin and Shuai Ma and Jiahao Jiang and Yanchen
                 Hou and Tianyu Wo",
  title =        "Error Bounded Line Simplification Algorithms for
                 Trajectory Compression: an Experimental Evaluation",
  journal =      j-TODS,
  volume =       "46",
  number =       "3",
  pages =        "11:1--11:44",
  month =        sep,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3474373",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Sep 29 06:47:23 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3474373",
  abstract =     "Nowadays, various sensors are collecting, storing, and
                 transmitting tremendous trajectory data, and it is well
                 known that the storage, network bandwidth, and
                 computing resources could be heavily wasted if raw
                 trajectory data is directly adopted. Line \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Khamis:2021:BQC,
  author =       "Mahmoud Abo Khamis and Phokion G. Kolaitis and Hung Q.
                 Ngo and Dan Suciu",
  title =        "Bag Query Containment and Information Theory",
  journal =      j-TODS,
  volume =       "46",
  number =       "3",
  pages =        "12:1--12:39",
  month =        sep,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3472391",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Sep 29 06:47:23 MDT 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3472391",
  abstract =     "The query containment problem is a fundamental
                 algorithmic problem in data management. While this
                 problem is well understood under set semantics, it is
                 by far less understood under bag semantics. In
                 particular, it is a long-standing open question whether
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Ma:2021:DDS,
  author =       "Chenhao Ma and Yixiang Fang and Reynold Cheng and Laks
                 V. S. Lakshmanan and Wenjie Zhang and Xuemin Lin",
  title =        "On Directed Densest Subgraph Discovery",
  journal =      j-TODS,
  volume =       "46",
  number =       "4",
  pages =        "13:1--13:45",
  month =        dec,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3483940",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Dec 10 10:59:16 MST 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3483940",
  abstract =     "Given a directed graph G, the directed densest
                 subgraph (DDS) problem refers to the finding of a
                 subgraph from G, whose density is the highest among all
                 the subgraphs of G. The DDS problem is fundamental to a
                 wide range of applications, such as fraud \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Singh:2021:TRH,
  author =       "Shikha Singh and Prashant Pandey and Michael A. Bender
                 and Jonathan W. Berry and Mart{\'\i}n Farach-Colton and
                 Rob Johnson and Thomas M. Kroeger and Cynthia A.
                 Phillips",
  title =        "Timely Reporting of Heavy Hitters Using External
                 Memory",
  journal =      j-TODS,
  volume =       "46",
  number =       "4",
  pages =        "14:1--14:35",
  month =        dec,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3472392",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Dec 10 10:59:16 MST 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3472392",
  abstract =     "Given an input stream $S$ of size $N$, a $\phi$-heavy
                 hitter is an item that occurs at least $\phi N$ times
                 in $S$. The problem of finding heavy-hitters is
                 extensively studied in the database literature. We
                 study a real-time heavy-hitters variant in which an
                 element must \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Benedikt:2021:BEI,
  author =       "Michael Benedikt and Pierre Bourhis and Louis Jachiet
                 and Efthymia Tsamoura",
  title =        "Balancing Expressiveness and Inexpressiveness in View
                 Design",
  journal =      j-TODS,
  volume =       "46",
  number =       "4",
  pages =        "15:1--15:40",
  month =        dec,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3488370",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Dec 10 10:59:16 MST 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3488370",
  abstract =     "We study the design of data publishing mechanisms that
                 allow a collection of autonomous distributed data
                 sources to collaborate to support queries. A common
                 mechanism for data publishing is via views: functions
                 that expose derived data to users, usually \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Grez:2021:FFC,
  author =       "Alejandro Grez and Cristian Riveros and Mart{\'\i}n
                 Ugarte and Stijn Vansummeren",
  title =        "A Formal Framework for Complex Event Recognition",
  journal =      j-TODS,
  volume =       "46",
  number =       "4",
  pages =        "16:1--16:49",
  month =        dec,
  year =         "2021",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3485463",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Dec 10 10:59:16 MST 2021",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3485463",
  abstract =     "Complex event recognition (CER) has emerged as the
                 unifying field for technologies that require processing
                 and correlating distributed data sources in real time.
                 CER finds applications in diverse domains, which has
                 resulted in a large number of proposals \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Berger:2022:SEC,
  author =       "Gerald Berger and Georg Gottlob and Andreas Pieris and
                 Emanuel Sallinger",
  title =        "The Space-Efficient Core of {Vadalog}",
  journal =      j-TODS,
  volume =       "47",
  number =       "1",
  pages =        "1:1--1:46",
  month =        mar,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3488720",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue May 3 06:36:04 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3488720",
  abstract =     "Vadalog is a system for performing complex reasoning
                 tasks such as those required in advanced knowledge
                 graphs. The logical core of the underlying Vadalog
                 language is the warded fragment of tuple-generating
                 dependencies (TGDs). This formalism ensures \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Tong:2022:URP,
  author =       "Yongxin Tong and Yuxiang Zeng and Zimu Zhou and Lei
                 Chen and Ke Xu",
  title =        "Unified Route Planning for Shared Mobility: an
                 Insertion-based Framework",
  journal =      j-TODS,
  volume =       "47",
  number =       "1",
  pages =        "2:1--2:48",
  month =        mar,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3488723",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue May 3 06:36:04 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3488723",
  abstract =     "There has been a dramatic growth of shared mobility
                 applications such as ride-sharing, food delivery, and
                 crowdsourced parcel delivery. Shared mobility refers to
                 transportation services that are shared among users,
                 where a central issue is route planning. \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Binna:2022:HOT,
  author =       "Robert Binna and Eva Zangerle and Martin Pichl and
                 G{\"u}nther Specht and Viktor Leis",
  title =        "Height Optimized Tries",
  journal =      j-TODS,
  volume =       "47",
  number =       "1",
  pages =        "3:1--3:46",
  month =        mar,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3506692",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue May 3 06:36:04 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3506692",
  abstract =     "We present the Height Optimized Trie (HOT), a fast and
                 space-efficient in-memory index structure. The core
                 algorithmic idea of HOT is to dynamically vary the
                 number of bits considered at each node, which enables a
                 consistently high fanout and thereby good \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Aumuller:2022:SNN,
  author =       "Martin Aum{\"u}ller and Sariel Har-Peled and Sepideh
                 Mahabadi and Rasmus Pagh and Francesco Silvestri",
  title =        "Sampling a Near Neighbor in High Dimensions --- Who is
                 the Fairest of Them All?",
  journal =      j-TODS,
  volume =       "47",
  number =       "1",
  pages =        "4:1--4:40",
  month =        mar,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3502867",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue May 3 06:36:04 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3502867",
  abstract =     "Similarity search is a fundamental algorithmic
                 primitive, widely used in many computer science
                 disciplines. Given a set of points S and a radius
                 parameter $ r > 0 $, the $r$-near neighbor ($r$-NN)
                 problem asks for a data structure that, given any query
                 point $q$, \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "4",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Schmid:2022:CRP,
  author =       "Markus L. Schmid",
  title =        "Conjunctive Regular Path Queries with Capture Groups",
  journal =      j-TODS,
  volume =       "47",
  number =       "2",
  pages =        "5:1--5:52",
  month =        jun,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3514230",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 6 06:55:46 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3514230",
  abstract =     "In practice, regular expressions are usually extended
                 by so-called capture groups or capture variables, which
                 allow to capture a subexpression by a variable that can
                 be referenced in the regular expression in order to
                 describe repetitions of subwords. We \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "5",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fan:2022:IGC,
  author =       "Wenfei Fan and Chao Tian",
  title =        "Incremental Graph Computations: Doable and Undoable",
  journal =      j-TODS,
  volume =       "47",
  number =       "2",
  pages =        "6:1--6:44",
  month =        jun,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3500930",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 6 06:55:46 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3500930",
  acknowledgement = ack-nhfb,
  articleno =    "6",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Cheng:2022:MOP,
  author =       "Ji Cheng and Da Yan and Wenwen Qu and Xiaotian Hao and
                 Cheng Long and Wilfred Ng and Xiaoling Wang",
  title =        "Mining Order-preserving Submatrices under Data
                 Uncertainty: a Possible-world Approach and Efficient
                 Approximation Methods",
  journal =      j-TODS,
  volume =       "47",
  number =       "2",
  pages =        "7:1--7:57",
  month =        jun,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3524915",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 6 06:55:46 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3524915",
  abstract =     "Given a data matrix $D$, a submatrix $S$ of $D$ is an
                 order-preserving submatrix (OPSM) if there is a
                 permutation of the columns of $S$, ! under which the
                 entry values of each row in $S$ are strictly
                 increasing. OPSM mining is \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "7",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Arroyuelo:2022:OJU,
  author =       "Diego Arroyuelo and Gonzalo Navarro and Juan L.
                 Reutter and Javiel Rojas-Ledesma",
  title =        "Optimal Joins Using Compressed Quadtrees",
  journal =      j-TODS,
  volume =       "47",
  number =       "2",
  pages =        "8:1--8:53",
  month =        jun,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3514231",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jun 6 06:55:46 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3514231",
  abstract =     "Worst-case optimal join algorithms have gained a lot
                 of attention in the database literature. We now count
                 several algorithms that are optimal in the worst case,
                 and many of them have been implemented and validated in
                 practice. However, the implementation \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "8",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Carmeli:2022:AUC,
  author =       "Nofar Carmeli and Shai Zeevi and Christoph Berkholz
                 and Alessio Conte and Benny Kimelfeld and Nicole
                 Schweikardt",
  title =        "Answering (Unions of) Conjunctive Queries using Random
                 Access and Random-Order Enumeration",
  journal =      j-TODS,
  volume =       "47",
  number =       "3",
  pages =        "9:1--9:49",
  month =        sep,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3531055",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 19 08:39:08 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3531055",
  abstract =     "As data analytics becomes more crucial to digital
                 systems, so grows the importance of characterizing the
                 database queries that admit a more efficient
                 evaluation. We consider the tractability yardstick of
                 answer enumeration with a polylogarithmic delay
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "9",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Asudeh:2022:FRR,
  author =       "Abolfazl Asudeh and Gautam Das and H. V. Jagadish and
                 Shangqi Lu and Azade Nazi and Yufei Tao and Nan Zhang
                 and Jianwen Zhao",
  title =        "On Finding Rank Regret Representatives",
  journal =      j-TODS,
  volume =       "47",
  number =       "3",
  pages =        "10:1--10:37",
  month =        sep,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3531054",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 19 08:39:08 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3531054",
  abstract =     "Selecting the best items in a dataset is a common task
                 in data exploration. However, the concept of ``best''
                 lies in the eyes of the beholder: Different users may
                 consider different attributes more important and,
                 hence, arrive at different rankings. \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "10",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Zeng:2022:PS,
  author =       "Tianjing Zeng and Zhewei Wei and Ge Luo and Ke Yi and
                 Xiaoyong Du and Ji-Rong Wen",
  title =        "Persistent Summaries",
  journal =      j-TODS,
  volume =       "47",
  number =       "3",
  pages =        "11:1--11:42",
  month =        sep,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3531053",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 19 08:39:08 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3531053",
  abstract =     "A persistent data structure, also known as a
                 multiversion data structure in the database literature,
                 is a data structure that preserves all its previous
                 versions as it is updated over time. Every update
                 (inserting, deleting, or changing a data record) to
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "11",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Guo:2022:IMR,
  author =       "Qintian Guo and Sibo Wang and Zhewei Wei and Wenqing
                 Lin and Jing Tang",
  title =        "Influence Maximization Revisited: Efficient Sampling
                 with Bound Tightened",
  journal =      j-TODS,
  volume =       "47",
  number =       "3",
  pages =        "12:1--12:45",
  month =        sep,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3533817",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Aug 19 08:39:08 MDT 2022",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3533817",
  abstract =     "Given a social network G with n nodes and m edges, a
                 positive integer k, and a cascade model C, the
                 influence maximization (IM) problem asks for k nodes in
                 G such that the expected number of nodes influenced by
                 the k nodes under cascade model C is \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "12",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Ketsman:2022:DRL,
  author =       "Bas Ketsman and Christoph Koch and Frank Neven and
                 Brecht Vandevoort",
  title =        "Deciding Robustness for Lower {SQL} Isolation Levels",
  journal =      j-TODS,
  volume =       "47",
  number =       "4",
  pages =        "13:1--13:??",
  month =        dec,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3561049",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:50 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3561049",
  abstract =     "While serializability always guarantees application
                 correctness, lower isolation levels can be chosen to
                 improve transaction throughput at the risk of
                 introducing certain anomalies. A set of transactions is
                 robust against a given isolation level if every
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "13",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{TenCate:2022:CQU,
  author =       "Balder {Ten Cate} and Victor Dalmau",
  title =        "Conjunctive Queries: Unique Characterizations and
                 Exact Learnability",
  journal =      j-TODS,
  volume =       "47",
  number =       "4",
  pages =        "14:1--14:??",
  month =        dec,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3559756",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:50 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3559756",
  abstract =     "We answer the question of which conjunctive queries
                 are uniquely characterized by polynomially many
                 positive and negative examples and how to construct
                 such examples efficiently. As a consequence, we obtain
                 a new efficient exact learning algorithm for a
                 \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "14",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Wei:2022:PQT,
  author =       "Victor Junqiu Wei and Raymond Chi-Wing Wong and Cheng
                 Long and David Mount and Hanan Samet",
  title =        "Proximity Queries on Terrain Surface",
  journal =      j-TODS,
  volume =       "47",
  number =       "4",
  pages =        "15:1--15:??",
  month =        dec,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3563773",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:50 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3563773",
  abstract =     "Due to the advance of the geo-spatial positioning and
                 the computer graphics technology, digital terrain data
                 has become increasingly popular nowadays. Query
                 processing on terrain data has attracted considerable
                 attention from both the academic and the \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "15",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Do:2022:ESD,
  author =       "Thanh Do and Goetz Graefe and Jeffrey Naughton",
  title =        "Efficient Sorting, Duplicate Removal, Grouping, and
                 Aggregation",
  journal =      j-TODS,
  volume =       "47",
  number =       "4",
  pages =        "16:1--16:??",
  month =        dec,
  year =         "2022",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3568027",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:50 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3568027",
  abstract =     "Database query processing requires algorithms for
                 duplicate removal, grouping, and aggregation. Three
                 algorithms exist: in-stream aggregation is most
                 efficient by far but requires sorted input; sort-based
                 aggregation relies on external merge sort; and hash
                 aggregation relies on an in-memory hash table plus hash
                 partitioning to temporary storage. Cost-based query
                 optimization chooses which algorithm to use based on
                 several factors, including the sort order of the input,
                 input and output sizes, and the need for sorted output.
                 For example, hash-based aggregation is ideal for output
                 smaller than the available memory (e.g., Query 1 of
                 TPC-H), whereas sorting the entire input and
                 aggregating after sorting are preferable when both
                 aggregation input and output are large and the output
                 needs to be sorted for a subsequent operation such as a
                 merge join.\par

                 Unfortunately, the size information required for a
                 sound choice is often inaccurate or unavailable during
                 query optimization, leading to sub-optimal algorithm
                 choices. In response, this article introduces a new
                 algorithm for sort-based duplicate removal, grouping,
                 and aggregation. The new algorithm always performs at
                 least as well as both traditional hash-based and
                 traditional sort-based algorithms. It can serve as a
                 system's only aggregation algorithm for unsorted
                 inputs, thus preventing erroneous algorithm choices.
                 Furthermore, the new algorithm produces sorted output
                 that can speed up subsequent operations. Google's F1
                 Query uses the new algorithm in production workloads
                 that aggregate petabytes of data every day.",
  acknowledgement = ack-nhfb,
  articleno =    "16",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Carmeli:2023:TOD,
  author =       "Nofar Carmeli and Nikolaos Tziavelis and Wolfgang
                 Gatterbauer and Benny Kimelfeld and Mirek Riedewald",
  title =        "Tractable Orders for Direct Access to Ranked Answers
                 of Conjunctive Queries",
  journal =      j-TODS,
  volume =       "48",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3578517",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:51 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3578517",
  abstract =     "We study the question of when we can provide direct
                 access to the k-th answer to a Conjunctive Query (CQ)
                 according to a specified order over the answers in time
                 logarithmic in the size of the database, following a
                 preprocessing step that constructs a \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "1",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Do:2023:RES,
  author =       "Thanh Do and Goetz Graefe",
  title =        "Robust and Efficient Sorting with Offset-value
                 Coding",
  journal =      j-TODS,
  volume =       "48",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3570956",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:51 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3570956",
  abstract =     "Sorting and searching are large parts of database
                 query processing, e.g., in the forms of index creation,
                 index maintenance, and index lookup, and comparing
                 pairs of keys is a substantial part of the effort in
                 sorting and searching. We have worked on \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "2",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Huang:2023:ECS,
  author =       "Ruihong Huang and Jianmin Wang and Shaoxu Song and
                 Xuemin Lin and Xiaochen Zhu and Jian Pei",
  title =        "Efficiently Cleaning Structured Event Logs: a Graph
                 Repair Approach",
  journal =      j-TODS,
  volume =       "48",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3571281",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Fri Mar 31 10:14:51 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3571281",
  abstract =     "Event data are often dirty owing to various recording
                 conventions or simply system errors. These errors may
                 cause serious damage to real applications, such as
                 inaccurate provenance answers, poor profiling results,
                 or concealing interesting patterns from \ldots{}",
  acknowledgement = ack-nhfb,
  articleno =    "3",
  ajournal =     "ACM Trans. Database Syst.",
  ajournal =     "ACM Trans. Database Syst.",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fakas:2023:PSD,
  author =       "Georgios J. Fakas and Georgios Kalamatianos",
  title =        "Proportionality on Spatial Data with Context",
  journal =      j-TODS,
  volume =       "48",
  number =       "2",
  pages =        "4:1--4:??",
  month =        jun,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3588434",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 3 07:30:46 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3588434",
  abstract =     "More often than not, spatial objects are associated
                 with some context, in the form of text, descriptive
                 tags (e.g., points of interest, flickr photos), or
                 linked entities in semantic graphs (e.g., Yago2,
                 DBpedia). Hence, location-based retrieval should be
                 \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "4",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Hu:2023:RDW,
  author =       "Donghui Hu and Qing Wang and Song Yan and Xiaojun Liu
                 and Meng Li and Shuli Zheng",
  title =        "Reversible Database Watermarking Based on
                 Order-preserving Encryption for Data Sharing",
  journal =      j-TODS,
  volume =       "48",
  number =       "2",
  pages =        "5:1--5:??",
  month =        jun,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3589761",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 3 07:30:46 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3589761",
  abstract =     "In the era of big data, data sharing not only boosts
                 the economy of the world but also brings about problems
                 of privacy disclosure and copyright infringement. The
                 collected data may contain users' sensitive
                 information; thus, privacy protection should be
                 \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "5",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Chen:2023:EBO,
  author =       "Yaxing Chen and Qinghua Zheng and Zheng Yan",
  title =        "Efficient Bi-objective {SQL} Optimization for Enclaved
                 Cloud Databases with Differentially Private Padding",
  journal =      j-TODS,
  volume =       "48",
  number =       "2",
  pages =        "6:1--6:??",
  month =        jun,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3597021",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Jul 3 07:30:46 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3597021",
  abstract =     "Hardware-enabled enclaves have been applied to
                 efficiently enforce data security and privacy
                 protection in cloud database services. Such enclaved
                 systems, however, are reported to suffer from I/O-size
                 (also referred to as communication-volume)-based
                 side-\ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "6",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Pavan:2023:MCM,
  author =       "A. Pavan and N. V. Vinodchandran and Arnab
                 Bhattacharyya and Kuldeep S. Meel",
  title =        "Model Counting Meets {$ F_0 $} Estimation",
  journal =      j-TODS,
  volume =       "48",
  number =       "3",
  pages =        "7:1--7:??",
  month =        sep,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3603496",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Sep 18 08:53:16 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3603496",
  abstract =     "Constraint satisfaction problems (CSPs) and data
                 stream models are two powerful abstractions to capture
                 a wide variety of problems arising in different domains
                 of computer science. Developments in the two
                 communities have mostly occurred independently and
                 \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "7",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Sarkar:2023:ETP,
  author =       "Subhadeep Sarkar and Tarikul Islam Papon and Dimitris
                 Staratzis and Zichen Zhu and Manos Athanassoulis",
  title =        "Enabling Timely and Persistent Deletion in
                 {LSM}-Engines",
  journal =      j-TODS,
  volume =       "48",
  number =       "3",
  pages =        "8:1--8:??",
  month =        sep,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3599724",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Sep 18 08:53:16 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3599724",
  abstract =     "Data-intensive applications have fueled the evolution
                 of log-structured merge (LSM) based key-value engines
                 that employ the out-of-place paradigm to support high
                 ingestion rates with low read/write interference. These
                 benefits, however, come at the cost \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "8",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Leventidis:2023:DHD,
  author =       "Aristotelis Leventidis and Laura {Di Rocco} and
                 Wolfgang Gatterbauer and Ren{\'e}e J. Miller and Mirek
                 Riedewald",
  title =        "{DomainNet}: Homograph Detection and Understanding in
                 Data Lake Disambiguation",
  journal =      j-TODS,
  volume =       "48",
  number =       "3",
  pages =        "9:1--9:??",
  month =        sep,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3612919",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Mon Sep 18 08:53:16 MDT 2023",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3612919",
  abstract =     "Modern data lakes are heterogeneous in the vocabulary
                 that is used to describe data. We study a problem of
                 disambiguation in data lakes: How can we determine if a
                 data value occurring more than once in the lake has
                 different meanings and is therefore a \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "9",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Shangqi:2023:POM,
  author =       "Lu Shangqi and Wim Martens and Matthias Niewerth and
                 Yufei Tao",
  title =        "Partial Order Multiway Search",
  journal =      j-TODS,
  volume =       "48",
  number =       "4",
  pages =        "10:1--10:??",
  month =        dec,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3626956",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 9 07:07:31 MST 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3626956",
  abstract =     "Partial order multiway search (POMS) is a fundamental
                 problem that finds applications in crowdsourcing,
                 distributed file systems, software testing, and more.
                 This problem involves an interaction between an
                 algorithm A and an oracle, conducted on a \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "10",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Herodotou:2023:CBD,
  author =       "Herodotos Herodotou and Elena Kakoulli",
  title =        "Cost-based Data Prefetching and Scheduling in Big Data
                 Platforms over Tiered Storage Systems",
  journal =      j-TODS,
  volume =       "48",
  number =       "4",
  pages =        "11:1--11:??",
  month =        dec,
  year =         "2023",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3625389",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Tue Jan 9 07:07:31 MST 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3625389",
  abstract =     "The use of storage tiering is becoming popular in
                 data-intensive compute clusters due to the recent
                 advancements in storage technologies. The Hadoop
                 Distributed File System, for example, now supports
                 storing data in memory, SSDs, and HDDs, while OctopusFS
                 \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "11",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Gottlob:2024:FPH,
  author =       "Georg Gottlob and Matthias Lanzinger and Cem Okulmus
                 and Reinhard Pichler",
  title =        "Fast Parallel Hypertree Decompositions in Logarithmic
                 Recursion Depth",
  journal =      j-TODS,
  volume =       "49",
  number =       "1",
  pages =        "1:1--1:??",
  month =        mar,
  year =         "2024",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3638758",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 13 07:26:47 MDT 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3638758",
  abstract =     "Various classic reasoning problems with natural
                 hypergraph representations are known to be tractable if
                 a hypertree decomposition (HD) of low width exists. The
                 resulting algorithms are attractive for practical use
                 in fields like databases and constraint \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "1",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Fan:2024:LEA,
  author =       "Wenfei Fan and Ping Lu and Kehan Pang and Ruochun Jin
                 and Wenyuan Yu",
  title =        "Linking Entities across Relations and Graphs",
  journal =      j-TODS,
  volume =       "49",
  number =       "1",
  pages =        "2:1--2:??",
  month =        mar,
  year =         "2024",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3639363",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 13 07:26:47 MDT 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3639363",
  abstract =     "This article proposes a notion of parametric
                 simulation to link entities across a relational
                 database D and a graph G. Taking functions and
                 thresholds for measuring vertex closeness, path
                 associations, and important properties as parameters,
                 parametric \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "2",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Wang:2024:AHT,
  author =       "Zhaoguo Wang and Chuzhe Tang and Xiaodong Zhang and
                 Qianmian Yu and Binyu Zang and Haibing Guan and Haibo
                 Chen",
  title =        "Ad Hoc Transactions through the Looking Glass: an
                 Empirical Study of Application-Level Transactions in
                 {Web} Applications",
  journal =      j-TODS,
  volume =       "49",
  number =       "1",
  pages =        "3:1--3:??",
  month =        mar,
  year =         "2024",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3638553",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 13 07:26:47 MDT 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3638553",
  abstract =     "Many transactions in web applications are constructed
                 ad hoc in the application code. For example, developers
                 might explicitly use locking primitives or validation
                 procedures to coordinate critical code fragments. We
                 refer to database operations \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "3",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

@Article{Currim:2024:IRC,
  author =       "Sabah Currim and Richard T. Snodgrass and Young-Kyoon
                 Suh",
  title =        "Identifying the Root Causes of {DBMS} Suboptimality",
  journal =      j-TODS,
  volume =       "49",
  number =       "1",
  pages =        "4:1--4:??",
  month =        mar,
  year =         "2024",
  CODEN =        "ATDSD3",
  DOI =          "https://doi.org/10.1145/3636425",
  ISSN =         "0362-5915 (print), 1557-4644 (electronic)",
  ISSN-L =       "0362-5915",
  bibdate =      "Wed Mar 13 07:26:47 MDT 2024",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "https://dl.acm.org/doi/10.1145/3636425",
  abstract =     "The query optimization phase within a database
                 management system (DBMS) ostensibly finds the fastest
                 query execution plan from a potentially large set of
                 enumerated plans, all of which correctly compute the
                 same result of the specified query. Sometimes
                 \ldots{}",
  acknowledgement = ack-nhfb,
  ajournal =     "ACM Trans. Database Syst.",
  articleno =    "4",
  fjournal =     "ACM Transactions on Database Systems",
  journal-URL =  "https://dl.acm.org/loi/tods",
}

%%% ====================================================================
%%% Cross-referenced entries must come last:
@InProceedings{Litwin:1980:LHN,
  author =       "W. Litwin",
  title =        "Linear Hashing: a New Tool for File and Table
                 Addressing",
  crossref =     "Lochovsky:1980:SIC",
  pages =        "212--223",
  year =         "1980",
  bibdate =      "Tue Jul 19 00:55:06 1994",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Reprinted in
                 \cite[p.~570--581]{Stonebraker:1988:RDS}.",
  annote =       "On excess collisions create new bucket and modify
                 hash. Leads to constant performance.",
}

@Article{Regnier:1985:AGF,
  author =       "Mireille Regnier",
  title =        "Analysis of Grid File Algorithms",
  journal =      j-BIT,
  volume =       "25",
  number =       "2",
  pages =        "335--357",
  month =        jun,
  year =         "1985",
  CODEN =        "BITTEL, NBITAB",
  DOI =          "https://doi.org/10.1007/BF01934379",
  ISSN =         "0006-3835 (print), 1572-9125 (electronic)",
  ISSN-L =       "0006-3835",
  MRclass =      "68P10",
  MRnumber =     "86m:68019",
  bibdate =      "Wed Jan 4 18:52:19 MST 2006",
  bibsource =    "Compendex database;
                 https://www.math.utah.edu/pub/tex/bib/bit.bib;
                 https://www.math.utah.edu/pub/tex/bib/hash.bib;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  URL =          "http://www.springerlink.com/openurl.asp?genre=article&issn=0006-3835&volume=25&issue=2&spage=335",
  abstract =     "Grid File is a generic name for geometric algorithms
                 used to map multiple-key indices onto primary files or
                 databases. Two existing methods by other authors
                 \cite{Fagin:1979:EHF,Larson:1978:DH} are generalized to
                 the multidimensional environment. The presence of a
                 hashing function is the primary difference between the
                 two algorithms; it is recommended to provide uniformity
                 when presented with biased key distributions.",
  acknowledgement = ack-nhfb,
  affiliationaddress = "Inst Natl de Recherche en Informatique et en
                 Automatique, Chesnay, Fr",
  classification = "723",
  fjournal =     "BIT (Nordisk tidskrift for informationsbehandling)",
  journal-URL =  "http://link.springer.com/journal/10543",
  journalabr =   "BIT (Copenhagen)",
  keywords =     "computer programming --- Algorithms; data processing;
                 database systems; dynamic data structures; File
                 Organization; grid file algorithms; hashing; multi-key
                 access",
}

@Proceedings{Kerr:1975:PIC,
  editor =       "Douglas S. Kerr",
  booktitle =    "{Proceedings of the International Conference on Very
                 Large Data Bases, Framingham, MA, USA, September
                 22--24, 1975}",
  title =        "{Proceedings of the International Conference on Very
                 Large Data Bases, Framingham, MA, USA, September
                 22--24, 1975}",
  publisher =    pub-ACM,
  address =      pub-ACM:adr,
  pages =        "viii + 592",
  year =         "1975",
  ISSN =         "0278-2596",
  LCCN =         "QA76.9.D3 I55 1975",
  bibdate =      "Fri Sep 16 12:12:29 1994",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "ACM SIGMOD v. 1, no. 1, September 1975.",
  price =        "US\$15.00",
  acknowledgement = ack-nhfb,
}

@Proceedings{Lochovsky:1980:SIC,
  editor =       "Frederick H. Lochovsky and ?. Taylor",
  key =          "VLDB'80",
  booktitle =    "{Sixth International Conference on Very Large Data
                 Bases: reprinted from Very large data bases}",
  title =        "{Sixth International Conference on Very Large Data
                 Bases: reprinted from Very large data bases}",
  publisher =    pub-IEEE,
  address =      pub-IEEE:adr,
  pages =        "ix + 435",
  year =         "1980",
  bibdate =      "Wed Jul 20 12:37:29 1994",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "IEEE Catalog Number 80CH1534-7C. ACM Order Number
                 471800. Long Beach Order Number 322.",
  acknowledgement = ack-nhfb,
}

@Proceedings{Schkolnick:1983:ICV,
  editor =       "Mario Schkolnick and C. (Costantino) Thanos",
  booktitle =    "{9th International Conference on Very Large Data Bases
                 (Florence, Italy, October 31--November 2, 1983)}",
  title =        "{9th International Conference on Very Large Data Bases
                 (Florence, Italy, October 31--November 2, 1983)}",
  publisher =    "VLDB Endowment",
  address =      "P.O. Box 2245, Saratoga, CA, USA",
  pages =        "xiii + 416",
  year =         "1983",
  ISBN =         "0-934613-15-X",
  ISBN-13 =      "978-0-934613-15-6",
  LCCN =         "QA 76.9 D3 I61 1983",
  bibdate =      "Sat Dec 7 13:05:35 MST 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "This conference is sponsored by VLDB Endowment and
                 co-sponsored by IFIP et al.",
  acknowledgement = ack-nhfb,
  keywords =     "database management -- congresses",
}

@Proceedings{ACM:1985:PFA,
  editor =       "{ACM}",
  key =          "ACM-PODS'85",
  booktitle =    "{Proceedings of the Fourth ACM SIGACT-SIGMOD Symposium
                 on Principles of Database Systems, March 25--27, 1985,
                 Portland, Oregon}",
  title =        "{Proceedings of the Fourth ACM SIGACT-SIGMOD Symposium
                 on Principles of Database Systems, March 25--27, 1985,
                 Portland, Oregon}",
  publisher =    pub-ACM,
  address =      pub-ACM:adr,
  pages =        "275",
  year =         "1985",
  ISBN =         "0-89791-153-9",
  ISBN-13 =      "978-0-89791-153-5",
  LCCN =         "QA 76.9 D3 A296 1985",
  bibdate =      "Sat Sep 17 10:24:09 1994",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
}

@Book{Kambayashi:1986:TIC,
  editor =       "Yahiko Kambayashi and Wesley Chu and Georges Gardarin
                 and Setsuo Ohsuga",
  booktitle =    "{Twelfth International Conference on Very Large Data
                 Bases, Proceedings (VLDB '86)}",
  title =        "{Twelfth International Conference on Very Large Data
                 Bases, Proceedings (VLDB '86)}",
  publisher =    pub-MORGAN-KAUFMANN,
  address =      pub-MORGAN-KAUFMANN:adr,
  pages =        "xiii + 512",
  year =         "1986",
  ISBN =         "0-934613-18-4",
  ISBN-13 =      "978-0-934613-18-7",
  LCCN =         "QA 76.9 D3 I61 1986",
  bibdate =      "Tue Nov 10 07:59:52 1998",
  bibsource =    "Compendex database;
                 https://www.math.utah.edu/pub/tex/bib/tods.bib",
  abstract =     "This conference proceeding contains 54 papers, 3 of
                 them are in abstract form only. The purpose of these
                 papers is the promotion of an understanding of current
                 research, and the exchange of ideas, experiences and
                 future directions in database systems. The main
                 subjects are office systems, database machines,
                 computer aided design\slash computer aided
                 manufacturing, query processing, relational operation
                 optimization, extended data models, historical database
                 systems, distributed database systems, knowledge-based
                 systems and managing of unformatted data, such as text
                 and graphical images.",
  acknowledgement = ack-nhfb,
  classification = "723; 903; 912",
  conference =   "Twelfth International Conference on Very Large Data
                 Bases, Proceedings (VLDB '86)",
  conferenceyear = "1986",
  keywords =     "Artificial Intelligence --- Applications; Computer
                 Metatheory --- Formal Logic; Data Processing, Business
                 --- Data Description; Database Integrity; Database
                 Systems; Information Retrieval Systems --- Evaluation;
                 Knowledge Base Systems; Logic Programming; Management
                 --- Information Systems; Query Processing; Research",
  meetingabr =   "Twelfth Int Conf Very Large Data Bases Proc VLDB 86",
  meetingaddress = "Kyoto, Jpn",
  meetingdate =  "Aug 25--28 1986",
  meetingdate2 = "08/25--28/86",
  sponsor =      "VLDB Endowment, Jpn; IFIP; INRIA; Information
                 Processing Soc of Japan; DARPA; et al",
}

@Book{Stonebraker:1988:RDS,
  editor =       "Michael Stonebraker",
  booktitle =    "Readings in Database Systems",
  title =        "Readings in Database Systems",
  publisher =    pub-MORGAN-KAUFMANN,
  address =      pub-MORGAN-KAUFMANN:adr,
  pages =        "xii + 644",
  year =         "1988",
  ISBN =         "0-934613-65-6",
  ISBN-13 =      "978-0-934613-65-1",
  LCCN =         "QA76.9.D3 R4 1988",
  bibdate =      "Tue Jul 19 00:53:02 1994",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  note =         "Contains reprint of \cite{Litwin:1980:LHN}.",
  price =        "US\$29.95",
  abstract =     "The purpose of this collection is to assemble
                 significant research contributions so they are easily
                 access-ible to anyone interested in database research.
                 It is appropriate for use as an introduction for
                 students or professionals from industry, and as a
                 reference volume to anyone active in database systems
                 \ldots. It is intended to serve as a core of material
                 that any DBMS professional should be familiar with.
                 Moreover, any industrial practitioner or graduate
                 student who wishes to be current on the important
                 research themes would be well advised to read these
                 papers.",
  acknowledgement = ack-nhfb,
  bookpages =    "xii + 644",
}

@Proceedings{Garcia-Molina:1990:PAS,
  editor =       "H{\'e}ctor Garc{\'\i}a-Molina and H. V. Jagadish",
  booktitle =    "{Proceedings of the 1990 ACM SIGMOD International
                 Conference on Management of Data, May 23--25, 1990,
                 Atlantic City, NJ}",
  title =        "{Proceedings of the 1990 ACM SIGMOD International
                 Conference on Management of Data, May 23--25, 1990,
                 Atlantic City, NJ}",
  volume =       "19(2)",
  publisher =    pub-ACM,
  address =      pub-ACM:adr,
  pages =        "xii + 388",
  month =        jun,
  year =         "1990",
  ISBN =         "0-89791-365-5",
  ISBN-13 =      "978-0-89791-365-2",
  LCCN =         "QA 76.9 D3 S53 v.19 no.2 1990",
  bibdate =      "Mon Dec 09 07:53:58 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  series =       "SIGMOD Record",
  acknowledgement = ack-nhfb,
}

@Book{Zdonik:1990:ROO,
  editor =       "B. Zdonik and David Maier",
  booktitle =    "Readings in Object-Oriented Database Systems",
  title =        "Readings in Object-Oriented Database Systems",
  publisher =    pub-MORGAN-KAUFMANN,
  address =      pub-MORGAN-KAUFMANN:adr,
  pages =        "x + 629",
  year =         "1990",
  ISBN =         "1-55860-000-0",
  ISBN-13 =      "978-1-55860-000-3",
  LCCN =         "QA76.9.D3 R42 1990",
  bibdate =      "Mon Dec 09 07:31:31 1996",
  bibsource =    "https://www.math.utah.edu/pub/tex/bib/tods.bib",
  price =        "US\$29.95",
  series =       "Morgan Kaufmann series in data management systems",
  acknowledgement = ack-nhfb,
}