Skip to main content

User account menu

  • Log in
DBS-Logo

Database Group Leipzig

within the department of computer science

ScaDS-Logo Logo of the University of Leipzig

Main navigation

  • Home
  • Study
    • Exams
      • Hinweise zu Klausuren
    • Courses
      • Current
    • Modules
    • LOTS-Training
    • Abschlussarbeiten
    • Masterstudiengang Data Science
    • Oberseminare
    • Problemseminare
    • Top-Studierende
  • Research
    • Projects
      • Benchmark datasets for entity resolution
      • FAMER
      • HyGraph
      • Privacy-Preserving Record Linkage
      • GRADOOP
    • Publications
    • Prototypes
    • Annual reports
    • Cooperations
    • Graduations
    • Colloquia
    • Conferences
  • Team
    • Erhard Rahm
    • Member
    • Former employees
    • Associated members
    • Gallery

Evaluation of Closely Coupled Systems for High Performance Database Processing

Breadcrumb

  • Home
  • Research
  • Publications
  • Evaluation of Closely Coupled Systems for High Performance Database Processing

Rahm, E.

Evaluation of Closely Coupled Systems for High Performance Database Processing

Proc. 13th Int. Conf. on Distributed Computing Systems, Pittsburgh, USA, May 1993, IEEE Computer Society Press, pp. 301-310

1993

Paper

Abstract

Closely coupled systems aim at a more efficient communication and cooperation between processing nodes compared to loosely coupled systems. This can be achieved by using globally shared semiconductor memory to speed up the exchange of messages or to store global data structures. For distributed database processing, the database sharing (shared disk) architecture can benefit most from such a close coupling. The author presents a detailed simulation study of closely coupled database sharing systems. A shared store called global extended memory (GEM) was used for system-wide concurrency and coherency control, and to improve input/output (I/O) performance. The performance of such an architecture is evaluated and compared with loosely coupled database sharing systems employing the primary copy approach for concurrency and coherency control. In particular, the impact of different update strategies (FORCE vs. NOFORCE) and workload allocation schemes (random vs. affinity-based routing) is studied. The use of shared disk caches implementing a global database buffer is also considered. Simulation results are presented for synthetically generated debit-credit workloads and a real-life workload represented by a database trace.

Recent publications

  • 2025 / 9: Generating Semantically Enriched Mobility Data from Travel Diaries
  • 2025 / 8: Slice it up: Unmasking User Identities in Smartwatch Health Data
  • 2025 / 6: SecUREmatch: Integrating Clerical Review in Privacy-Preserving Record Linkage
  • 2025 / 6: Leveraging foundation models and goal-dependent annotations for automated cell confluence assessment
  • 2025 / 5: Federated Learning With Individualized Privacy Through Client Sampling

Footer menu

  • Directions
  • Contact
  • Impressum