cm20145 architectures and implementations dr alwyn barry dr joanna bryson

58
CM20145 CM20145 Architectures and Architectures and Implementations Implementations Dr Alwyn Barry Dr Joanna Bryson

Post on 20-Dec-2015

218 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

CM20145CM20145Architectures and Architectures and ImplementationsImplementations

Dr Alwyn BarryDr Joanna Bryson

Page 2: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Last Time 1 – Rules to WatchLast Time 1 – Rules to Watch 1NF: attributes not atomic. 2NF: non-key attribute FD on part of

key. 3NF: one non-key attribute FD on

another. Boyce-Codd NF: overlapping but

otherwise independent candidate keys. 4NF: multiple, independent multi-valued

attributes. 5NF: unnecessary table. Domain Key / NF: all constraints either

domain or key

Page 3: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Last Time 2 – ConceptsLast Time 2 – Concepts

Functional Dependencies: Axioms & Closure.

Lossless-join decomposition. Design Process. Normalization Problems.

Soon: Architectures and Implementations

Page 4: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

ImplementImplementing Databasesing Databases

BABAR Database Group in front of

Data Storage EquipmentPhoto - Diana Rogers

Previous lectures were about how you design and program databases.

Most of the rest of the term is about how they work.

Important to understand tradeoffs when: choosing systems to buy, enforcing security &

reliability, or optimizing time or space.

Page 5: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

ConsiderationsConsiderations

BABAR Database Group in front of

Data Storage EquipmentPhoto - Diana Rogers

Making things faster or more reliable often requires more equipment.

What if two people try to change the same thing at the same time?

Speed, size and reliability often trade off with each other.

How does the equipment work together?

Page 6: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Introduction to Transactions. Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems. Parallel Systems. Distributed Systems.

Page 7: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Transaction ConcernsTransaction Concerns A transaction is a unit of program execution that

accesses and possibly updates various data items. A transaction starts with a consistent database. During transaction execution the database may

be inconsistent. When the transaction is committed, the

database must be consistent. Two main issues to deal with:

Failures, e.g. hardware failures and system crashes.

Concurrency, for simultaneous execution of multiple transactions.

©Silberschatz, Korth and Sudarshan

Modifications & additions by S Bird, J Bryson

Page 8: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Example: A Fund TransferExample: A Fund TransferTransfer $50 from account A to B:

1. read(A)2. A := A – 503. write(A)4. read(B)5. B := B + 506. write(B)

Atomicity: if the transaction fails after step 3 and before step 6, the system must ensure that no updates are reflected in the database.

Consistency: the sum of A and B is unchanged by the execution of the transaction.

Isolation: between steps 3-6, no other transaction should access the partially updated database, or it would see an inconsistent state (A + B < A + B).

Durability: once the user is notified that the transaction is complete, the database updates must persist despite failures.

Page 9: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Very Quick Introduction to Transactions.

Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems. Parallel Systems. Distributed Systems.

Page 10: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage Media Physical Storage Media ConcernsConcerns Speed with which the data can be

accessed. Cost per unit of data. Reliability:

data loss on power failure or system crash, physical failure of the storage device.

We can differentiate storage into: volatile storage (loses contents when power

is switched off), non-volatile storage (doesn’t).

©Silberschatz, Korth and Sudarshan

Modifications & additions by J Bryson

Page 11: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage MediaPhysical Storage Media

Cache: Fastest and most costly form of storage;

volatile; managed by the computer system hardware.

Main memory: Fast access (10s to 100s of nanoseconds; 1

nanosecond = 10–9 seconds). Generally too small (or too expensive) to

store the entire database: capacities of up to a few Gigabytes widely used

currently, Capacities have gone up and per-byte costs have

decreased steadily and rapidly (roughly factor of 2 every 2 to 3 years).

Also volatile.

Page 12: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage Media (Cont.)Physical Storage Media (Cont.) Flash memory – or EEPROM

(Electrically Erasable Programmable Read-Only Memory)

Data survives power failure – non-volatile. Data can be written at a location only once,

but location can be erased and written to again. Limited number of write/erase cycles possible. Erasing must be done to an entire bank of

memory Reads are roughly as fast as main memory,

but writes are slow (few microseconds), erase is slower.

Cost per unit unit similar to main memory. Widely used in embedded devices such as

digital cameras

Page 13: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage Media (Cont.)Physical Storage Media (Cont.) Magnetic disk

Data is stored on spinning disk, and read/written magnetically.

Primary medium for the long-term storage of data; typically stores entire database.

Data must be moved from disk to main memory for access, and written back for storage.

Much slower access than main memory, but much cheaper, rapid improvements.

Direct access – possible to read data on disk in any order (unlike magnetic tape).

Survives most power failures and system crashes – disk failure can destroy data, but fairly rare.

Page 14: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Magnetic Hard Disk MechanismMagnetic Hard Disk Mechanism

NOTE: Diagram is schematic, and simplifies the structure of actual disk drives

Page 15: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage Media (Cont.)Physical Storage Media (Cont.) Optical Storage (e.g. CD, DVD)

Non-volatile, data is read optically from a spinning disk using a laser.

Write-one, read-many (WORM) optical disks used for archival storage (CD-R and DVD-R)

Multiple write versions also available (CD-RW, DVD-RW, and DVD-RAM)

Reads and writes are slower than with magnetic disk.

Juke-box systems available for storing large volumes of data: large numbers of removable disks, a few drives,

and a mechanism for automatic loading/unloading of disks.

Page 16: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Physical Storage Media (Cont.)Physical Storage Media (Cont.)

Tape storage non-volatile, used primarily to backup for

recovery from disk failure, and for archival data.

Sequential access – much slower than disk.

Very high capacity (40 to 300 GB tapes available)

Tape can be removed from drive storage costs much cheaper than magnetic disk, but drives are expensive.

Jukeboxes store massive amounts of data: hundreds of terabytes (1 terabyte = 109 bytes) to

even a petabyte (1 petabyte = 1012 bytes)

Page 17: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Storage HierarchyStorage Hierarchy

1. Primary storage: Fastest media but volatile (e.g. cache, main memory).

2. Secondary storage: next level in hierarchy, non-volatile, moderately fast access time (e.g. flash, magnetic disks).

also called on-line storage

3. Tertiary storage: lowest level in hierarchy, non-volatile, slow access time (e.g. magnetic tape, optical storage).

also called off-line storage

Page 18: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Very Quick Introduction to Transactions.

Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems. Parallel Systems. Distributed Systems.

Page 19: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Architecture ConcernsArchitecture Concerns

SpeedCostReliabilityMaintainability

Structure: Centralized Client/Server Parallel Distributed

Page 20: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Centralized SystemsCentralized Systems Run on a single computer system and do not

interact with other computer systems. General-purpose computer system: one to a few

CPUs and a number of device controllers that are connected through a common bus that provides access to shared memory.

Single-user system (e.g., personal computer or workstation): desk-top unit, usually only one CPU, 1-2 hard disks; OS may support only one user.

Multi-user system: May have more disks, more memory, multiple CPUs; Must have a multi-user OS. Serve a large number of users who are connected to the

system via terminals. Often called server systems.

Page 21: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

A Centralized Computer SystemA Centralized Computer System

Page 22: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Very Quick Introduction to Transactions.

Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems.

transaction servers – used in relational database systems, and

data servers – used in object-oriented database systems.

Parallel Systems. Distributed Systems.

Page 23: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Client-Server SystemsClient-Server Systems

PC and networking revolutions allow distributed processing.

Easiest to distribute only part of DB functionality.

Server systems satisfy requests generated at m client systems, e.g.

Page 24: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Client-Server Systems (Cont.)Client-Server Systems (Cont.) Database functionality can be divided into:

Back-end: access structures, query evaluation & optimization, concurrency control and recovery.

Front-end: consists of tools such as forms, report-writers, and graphical user interface facilities.

The interface between the front-end and the back-end is through SQL or through an application program interface (API).

Page 25: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Transaction ServersTransaction Servers Also called query server systems or SQL

server systems. Clients send requests to server, where the

transactions are executed, and results shipped back to the client.

Requests specified in SQL, and communicated to the server through a remote procedure call (RPC).

Transactional RPC allows many RPC calls to collectively form a transaction.

Open Database Connectivity (ODBC) is a C language API “standard” from Microsoft.

JDBC standard similar, for Java.

Page 26: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Trans. Server Process StructureTrans. Server Process Structure

A typical transaction server consists of multiple processes accessing data in shared memory: Server processes

Receive user queries (transactions), execute them and send results back.

May be multithreaded, allowing a single process to execute several user queries concurrently.

Typically multiple multithreaded server processes.

Lock manager process Handles concurrency (lecture 13).

Database writer process Output modified buffer blocks to disks continually.

Page 27: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Trans. Server Processes (Cont.)Trans. Server Processes (Cont.)

Log writer process Server processes simply add log records to

log record buffer. Log writer process outputs log records to

stable storage. (lecture 16)

Checkpoint process Performs periodic checkpoints.

Process monitor process Monitors other processes, and takes recovery

actions if any of the other processes fail. E.g. aborting any transactions being executed

by a server process and restarting it.

Page 28: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Trans. Server Processes (Cont.)Trans. Server Processes (Cont.)

Page 29: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Data ServersData Servers Used in many object-oriented DBs. Ship data to client machines where processing

is performed, and then ship results back to the server machine.

This architecture requires full back-end functionality at the clients.

Used in LANs, where high speed connection between clients and server, client machines comparable in processing power to

the server machine, and tasks to be executed are compute intensive.

Issues: Page-Shipping versus Item-Shipping Locking Data Caching Lock Caching

Page 30: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Very Quick Introduction to Transactions.

Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems.

transaction servers – used in relational database systems, and

data servers – used in object-oriented database systems.

Parallel Systems. Distributed Systems.

Page 31: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Parallel SystemsParallel Systems Parallel database systems consist of

multiple processors and multiple disks connected by a fast network. Coarse-grain parallel machine consists of a

small number of powerful processors Massively parallel or fine grain parallel

machine utilizes thousands of smaller processors.

Two main performance measures: throughput – the number of tasks that can

be completed in a given time interval. response time – the amount of time taken

to complete a single task.

Page 32: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Speed up: a fixed-sized problem executing on a small system is given to a system which is N-times larger. Measured by: speedup = small system elapsed time large system elapsed time Speedup is linear if equation equals N.

Scale up: increase the size of both the problem and the system N-times larger system used to perform N-times

larger job Measured by: scaleup = small system small problem elapsed time big system big problem elapsed time Scale up is linear if equation equals 1.

Speed Up and Scale UpSpeed Up and Scale Up

Page 33: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Speed upSpeed up

Page 34: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Scale upScale up

Page 35: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Limits to Speeding / Scaling UpLimits to Speeding / Scaling UpSpeed up and scale up are often sublinear due to: Start up costs: Cost of starting up multiple

processes may dominate computation time. Interference: Processes accessing shared

resources (e.g. system bus, disks, locks) compete spend time waiting on other processes rather than doing useful work.

Skew: Increasing the degree of parallelism increases the variance in service times of parallely executing tasks. Overall execution time determined by slowest parallely executing tasks.

Page 36: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Interconnection ArchitecturesInterconnection Architectures

Bus. System components send data on and receive data from a single communication bus; Does not scale well with increasing parallelism.

Mesh. Components are arranged as nodes in a grid, and each component is connected to all adjacent components Communication links grow with growing number of

components, and so scales better. But may require 2n hops to send message to a node (or

n with wraparound connections at edge of grid). Hypercube. Components are numbered in binary;

components are connected to one another if their binary representations differ in exactly one bit. n components are connected to log(n) other components

and can reach each other via at most log(n) links; reduces communication delays.

Page 37: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Interconnection ArchitecturesInterconnection Architectures

Page 38: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Parallel Database ArchitecturesParallel Database Architectures

Shared memory Shared disk Shared nothing – processors share neither

a common memory nor common disk. Hierarchical – hybrid of the above

architectures.

Page 39: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

OverviewOverview

Very Quick Introduction to Transactions.

Introduction to Storage. Architecture:

Centralized Systems. Client-Server Systems.

transaction servers – used in relational database systems, and

data servers – used in object-oriented database systems.

Parallel Systems. Distributed Systems.

Page 40: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Distributed SystemsDistributed Systems Data spread over multiple machines

(also referred to as sites or nodes). Network interconnects the machines. Data shared by users on multiple

machines.

Page 41: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Distributed DatabasesDistributed Databases Homogeneous distributed databases

Same software/schema on all sites, data may be partitioned among sites

Goal: provide a view of a single database, hiding details of distribution

Heterogeneous distributed databases Different software/schema on different sites Goal: integrate existing databases to provide useful

functionality Differentiate between local and global

transactions A local transaction accesses data in the single site at

which the transaction was initiated. A global transaction either accesses data in a site

different from the one at which the transaction was initiated or accesses data in several different sites.

Page 42: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Trade-offs in Distrib. SystemsTrade-offs in Distrib. Systems Sharing data – users at one site able to access

the data residing at some other sites. Autonomy – each site is able to retain a degree

of control over data stored locally. Higher system availability through redundancy

– data can be replicated at remote sites, and system can function even if a site fails.

Disadvantage: added complexity required to ensure proper coordination among sites. Software development cost. Greater potential for bugs. Increased processing overhead.

Page 43: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

SummarySummary Introduction to Transactions & Storage Architecture concerns: speed, cost,

reliability, maintainability. Architecture Types:

Centralized Client/Server

transaction servers – used in relational database systems, and

data servers – used in object-oriented database systems

Parallel Distributed

Next: Integrity and Security

Page 44: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Reading & ExercisesReading & Exercises

Reading Silberschatz Chapters 18, 11 (more

of Ch. 11 in lecture 16). Connolly & Begg have little on

hardware, but lots on distributed databases (section 6).

Exercises: Silberschatz 18.4, 18.7-9.

Page 45: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

End of TalkEnd of Talk

The following slides are ones I might use next year. They were not presented in class

and you are not responsible for them.

JJB Nov 2004

Page 46: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Formalisms and Your MarksFormalisms and Your Marks

Doing proofs is not required for a passing mark in this course.

Since CM20145 is part of a Computer Science degree, proofs could very well help undergraduates go over the 70% mark. “Evidence of bringing in knowledge from

other course curriculum and other texts” – your assignment.

“go beyond normal expectations” – ibid. Could also apply to MScs.

Page 47: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

FD ClosureFD Closure Given a set F of FDs, other FDs are logically implied.

E.g. If A B and B C, we can infer that A C The set of all FDs implied by F is the closure of F, written F+ . Find F+ by applying Armstrong’s Axioms:

if , then (reflexivity) if , then (augmentation) if , and , then (transitivity)

Additional rules (derivable from Armstrong’s Axioms): If and holds, then holds (union) If holds, then holds and holds (decomposition) If holds and holds, then holds (pseudotransitivity)

Page 48: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

FD Closure: ExampleFD Closure: Example R = (A, B, C, G, H, I)

F = { A B A C CG H CG I B H}

some members of F+

A H by transitivity from A B and B H

AG I by augmenting A C with G, to get AG CG

and then transitivity with CG I CG HI

by union rule with CG H and CG I

Page 49: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Computing FD ClosureComputing FD Closure

To compute the closure of a set of FDs F: F+ = F

repeatfor each FD f in F+

apply reflexivity and augmentation rules on f

add the resulting FDs to F+

for each pair of FDs f1and f2 in F+

if f1 and f2 can be combined using transitivity

then add the resulting FD to F+

until F+ does not change any further

(NOTE: More efficient algorithms exist)Most slides in this talk

©Silberschatz, Korth and Sudarshan

Modifications & additions by S Bird, J Bryson

Page 50: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Third Normal Form (3NF)Third Normal Form (3NF)

Violated when a nonkey column is a fact about another nonkey column.

A column is not fully functionally dependent on the primary key.

R is 3NF iff R is 2NF and has no transitive dependencies. EXCHANGE RATE violates in this case.

STOCK

*stock codefirm namestock price

stock quantitystock dividend

stock PE

NATION

*nation codenation name

exchange rate

STOCK

STOCK CODE

NATION EXCHANGE RATE

MG USA 0.67

IR AUS 0.46

©This one’s from Watson!

Page 51: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Shared MemoryShared Memory

Processors and disks have access to a common memory, typically via a bus or through an interconnection network.

Extremely efficient communication between processors — data in shared memory can be accessed by any processor without having to move it using software.

Downside – architecture is not scalable beyond 32 or 64 processors since the bus or the interconnection network becomes a bottleneck

Widely used for lower degrees of parallelism (4 to 8).

Page 52: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Shared DiskShared Disk All processors can directly access all disks via an

interconnection network, but the processors have private memories. The memory bus is not a bottleneck Architecture provides a degree of fault-tolerance — if

a processor fails, the other processors can take over its tasks since the database is resident on disks that are accessible from all processors.

Examples: IBM Sysplex and DEC clusters (now part of Compaq) running Rdb (now Oracle Rdb) were early commercial users

Downside: bottleneck now occurs at interconnection to the disk subsystem.

Shared-disk systems can scale to a somewhat larger number of processors, but communication between processors is slower.

Page 53: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Shared NothingShared Nothing

Node consists of a processor, memory, and one or more disks. Processors at one node communicate with another processor at another node using an interconnection network. A node functions as the server for the data on the disk or disks the node owns.

Examples: Teradata, Tandem, Oracle-n CUBE Data accessed from local disks (and local memory

accesses) do not pass through interconnection network, thereby minimizing the interference of resource sharing.

Shared-nothing multiprocessors can be scaled up to thousands of processors without interference.

Main drawback: cost of communication and non-local disk access; sending data involves software interaction at both ends.

Page 54: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

HierarchicalHierarchical Combines characteristics of shared-memory,

shared-disk, and shared-nothing architectures. Top level is a shared-nothing architecture. Each node of the system could be a:

shared-memory system with a few processors. shared-disk system, and each of the systems sharing

a set of disks could be a shared-memory system.

Reduce the complexity of programming such systems by distributed virtual-memory architectures (also called non-uniform memory architectures.)

Page 55: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Implementation Issues for Distro DBImplementation Issues for Distro DB Atomicity needed even for transactions that update data at

multiple sites. Transaction cannot be committed at one site and aborted at

another The two-phase commit protocol (2PC) used to ensure

atomicity. Basic idea: each site executes transaction till just before

commit, and the leaves final decision to a coordinator Each site must follow decision of coordinator: even if there is a

failure while waiting for coordinators decision To do so, updates of transaction are logged to stable storage and

transaction is recorded as “waiting” More details in Sectin 19.4.1

2PC is not always appropriate: other transaction models based on persistent messaging, and workflows, are also used.

Distributed concurrency control (and deadlock detection) required.

Replication of data items required for improving data availability.

Details of above in Chapter 19.

Page 56: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Batch and Transaction ScaleupBatch and Transaction Scaleup Batch scale up:

A single large job; typical of most database queries and scientific simulation.

Use an N-times larger computer on N-times larger problem.

Transaction scale up: Numerous small queries submitted by

independent users to a shared database; typical transaction processing and timesharing systems.

N-times as many users submitting requests (hence, N-times as many requests) to an N-times larger database, on an N-times larger computer.

Well-suited to parallel execution.

Page 57: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Network TypesNetwork Types

Local-area networks (LANs) – composed of processors that are distributed over small geographical areas, such as a single building or a few adjacent buildings.

Wide-area networks (WANs) – composed of processors distributed over a large geographical area.

Discontinuous connection – WANs, such as those based on periodic dial-up (using, e.g., UUCP), that are connected only for part of the time.

Continuous connection – WANs, such as the Internet, where hosts are connected to the network at all times.

Page 58: CM20145 Architectures and Implementations Dr Alwyn Barry Dr Joanna Bryson

Networks Types (Cont.)Networks Types (Cont.)

WANs with continuous connection are needed for implementing distributed database systems

Groupware applications such as Lotus notes can work on WANs with discontinuous connection: Data is replicated. Updates are propagated to replicas periodically. No global locking is possible, and copies of data

may be independently updated. Non-serializable executions can thus result.

Conflicting updates may have to be detected, and resolved in an application dependent manner.