risc+reward_1988

Upload: brian-boyle

Post on 08-Apr-2018

214 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/7/2019 RISC+Reward_1988

    1/24

    RISC and Reward

    MetaTrends:

    of RISC and Reward

    In times of crisis and change, Knowledge is Power.

    John Fitzgerald Kennedy 1961

    Predicting is hard especially the future.

    Niels Bohr (Copenhagen) 1938

    On Foreknowledge

    Wouldnt it be wonderful to know the future before it happens?

    At the very least, it could be immensely profitable to know the next

    number to come up on the roulette wheel or the winner of tomorrows

    Daily Double. Such simple greed is more than the grist for the writers of

    old Twilight Zoneor new Back to the Future IIscripts. After all, it was the

    roll of the dice and the spin of the wheel that prompted Blaise Pascal to

    develop the first mathematical theories of Probability and Statistics in the

    early Seventeenth Century and the lure of the track and the table that

    seduced the first computer programmer, the Countess Ada Lovelace, in

    the early Nineteenth.

    A century later, the lure of a peek at the future inspired the Dow

    Theory, the quantitative analysis of the peaks, the valleys and the

    opportunities of the securities trading floor, an application area far more

    profitable and hence far more respectable (until recently) than the casino

    floor. The raw data fueling this engine of risk and reward have been

    published each trading day since the 1920s duly accompanied by itseponymous Industrial Average in the Dow-Jones Corporations ever-

    respectable Wall Street Journal, theRacing Formof the more modern

    breed of speculator.

    As we near the end of the Twentieth Centurys literally exponential

    growth of science, technology and the changes wrought by them, those

  • 8/7/2019 RISC+Reward_1988

    2/24

    most critically dependent on the ambiguous oracles of change are those

    that stand to gain or lose the most on the roll of the technological die. It is

    the investor, the entrepreneur and the user of these complex patterns of

    information whether inside and outside the giant corporate and

    government entities who most vitally depend on the new art and

    science of managing complexity.

    It is the thesis of this author and of his company from experience in

    researching and forecasting the future for those entities and individuals

    that the current evolutionary trend towards open systems is not an end in

    itself. Indeed, we believe it is even more than a means for managing the

    inevitable complexity of this complex information age. Instead, we see it

    as the tip of an emerging iceberg, a manifestation of an inescapable

    Darwinian force, the fabled Invisible Hand, reshaping our global

    economy and society into a new connective topology or, in current

    parlance, a new paradigm or returning to a very old, even prehistoric

    paradigm. It is a practical consequence of our thesis that the shape of the

    world under this fundamental new paradigm can be foreseen, anticipated

    and acted upon.

  • 8/7/2019 RISC+Reward_1988

    3/24

    On Connections

    If I have seen far it is because I have stood on the shoulders of

    giants.

    Isaac Newton (London) 1703

    In other disciplines, people make progress by standing on each

    others shoulders

    in Computing, its seems, we make progress only by standing on each

    others toes.

    Edsgar Djikstra (ACM Pacific Conference, San Francisco) 1975

    As the video-journalist-historian-philosopher James Burke has so ably

    pointed out, the entire history (and, perhaps, prehistory) of the human race

    is a history of connections. Human speech and its descendant forms of

    communicationmake possible the cooperation that is the basis of our

    societies, from the first protohuman tribes in paleolithic Africa to the

    satellite-linked world economy of today. With the development of the

    written word and number, communication in space was augmented by

    communication in time from past to present to future and the growthof human culture and knowledge began its exponential growth, building

    upon the knowledge of the past rather than reinventing it each generation.

    For much of the last five-hundred years, from the Age of Exploration

    onward, the physical aspects of connection transportation more than

    communication historically has captured more of the attention of

    society and had more visible impact on the economy. Just as in the

    popular view, The Wheel is the most significant early invention, so the

    roads of the Roman Empire, then the sailing ships of the various colonialEmpires, the steamship, the railroad, the automobile and the aircraft have

    shaped our society both in deed and name. Weve had theSteam Age,

    the Auto Age, the Jet Ageand the Space Age all within a single recent

    lifetime. There seems to have been no comparable reference to the

    Printing Age, the Telegraph Age, the Telephone Ageor any information-

  • 8/7/2019 RISC+Reward_1988

    4/24

    based equivalents until the currently referencedComputer Ageor

    Information Age each implying something more tangible, less transitory,

    less ephemeral than the abstraction of pure communication.

    In analyzing this elusive societal phenomenon the aversion of theaverage human mind to recognize or acknowledge that which it cannot

    grasp we have found as many explanations as there are experts, as

    many opinions as there are observers. But from a purely practical and

    directly utilitarian stance, at least one viewpoint recommends itself as a

    successful technique in predicting the future, at least as it may be

    extrapolated in a credible (if counterintuitive) way, from the past. The

    trend, or perhaps a Metatrend(with apologies to John Naisbitt) which

    emerges from this analysis is: The place to watch for fundamental

    transformingchanges is at the interface between Goods and Services,

    where new enablingtechnologies permit values once added by Services

    now to be added by Products.

  • 8/7/2019 RISC+Reward_1988

    5/24

    On Forecasting

    No more than 500,000 automobiles will ever be owned by families in

    the United States,

    as this is the greatest number of households which can afford the

    requisite chauffeurs.

    Unknown Market Predictor 1906

    If telephone usage in these United States were to continue growing at

    its current pace,

    in fifty years the entire population would need to be employed as

    telephone operators.

    Anonymous Forecasting Wit 1928

    Of course I am interested in the future I plan to spend the rest of

    my life there.

    Isaac Asimov (New York) 1963

    Among the less pretentious and self-satisfied members of the marketresearch and technological forecasting communities (possibly an empty

    set), the paraphrased predictions above stand as excellent examples of

    the many dangers inherent in linear extrapolation. Modern forecasters

    be they academic economists, government statisticians or private

    prognosticators (i.e., financial analysts and employees of syndicated

    market research firms) may smile at these naive efforts. But on semi-

    log paper (or its Lotus 1-2-3 equivalent) our complex formulae for

    CAAGRs (Compound Aggregate Annual Growth Rates) often plot to

    equally boring straight lines, bland linear extrapolations transposed into

    exponentials. Past events, on the other hand the histories of actual

    shipments, installations and use, rarely look so clean or well-behaved

    look very bumpy and irregular by contrast.

  • 8/7/2019 RISC+Reward_1988

    6/24

    This is not to say that this persistent error of the professionals is readily

    avoidable or that Economics is called the dismal science because of its

    dismal record of accuracy its predictions have compiled. The amateur

    predictor, whatever his stripe, is on the whole even worse: the only truly

    consistent Dow-Jones datum predicting the way the market will not go is

    the odd-lots trading figure, the paw-print of the small investor. The

    convoluted sine waves of the seesawing market where to be even

    slightly out of phase, to be even slightly behind the information curve, and

    hence perpetually wrong are only one form of the complex exponential

    which the unaided human mind seems predisposed to linearize.

    Mathematically, the same tendency to flatten the emotionally

    incomprehensible and ever-steepening exponential growth curve compels

    us to draw straight lines through some point in the near future. Thus wealways overestimate progress for the near future but grossly

    underestimate the possibilities of the long run. The latter accounts for the

    popular presss amusing technological predictions of decades past

    personal commuter helicopters for intracity travel, ultra-streamlined 200-

    mph cars for highways, 500-mph transcontinental trains and twelve-

    propeller flying boats or atomic-powered airships for intercontinental trips.

    Most significantly, both of the grossly underestimating

    prognostications regarding the automobiles and the telephones futuremarkets were right in essence but wrong in interpretations based on their

    metalinear mode of thinking from which we have no scientific escape

    even today, only the artistic flight of the true visionary. In reality, both of

    the predictions were correct in principle, but the essential value-added

    component of each of these two industries were transformed (in their

    essence) from reliance upon raw service functions to reliance on

    products plus a support infrastructure, specifically:

    Henry Fords economic advances in mass production and economies

    of scale were paralleled by the technological advances of the electric

    starter, synchromesh gears, pneumatic tire, automatic spark-

    advance and later market-expanding enhancements like the

    automatic transmission and power steering coupled with the

    infrastructure of paved highways and refined petroleum followed by

  • 8/7/2019 RISC+Reward_1988

    7/24

    the automotive fuel and service network and interstate highway

    system, converging to create an environment wherein we have all

    become our own chauffeurs.

    Bell Telephone Laboratories comparable advances in telephony,

    beginning with the stepping relays that displaced the switchboard

    and culminating in todays digitized, computer-routed, satellite-

    relayed, Touch-Tone direct international dialing, have now

    transformed the vast majority of the original service functions into a

    global network of mass-produced products all interconnected by a

    vast support structure in which we have all become our own

    telephone operators.

    The forgotten forecasters may indeed have the last laugh on their

    latter-day critics, for thoughts and deeds based on linear or even

    metalinearextrapolation are subject to the penalties set forth in The Origin

    of Species. The American railroads once constituted a rich, inbred

    oligopoly that looked no farther for its competition than the other side of

    the tracks. In the passenger markets, they never looked upwards to the

    unfriendly skies of primitive air travel; nor, in the freight transport markets,

    did they recognize the burgeoning network of publicly funded highways

    shared by long-haul trucks with an explosion of private autos. The Swiss

    arrogance in their early dismissal of the externallydigital watch nearly costthem an entire industry when quartz technology moved in to replace the

    timepiece's internals.

    Lest we be too quick to judge, in our own backyard, the U S

    semiconductor industry in the 70s virtually abandoned the slow high-

    density, low-powered CMOS technology to the Japanese for

    calculators, watches and other toy applications. Meanwhile, we pursued

    the bipolar and ECL leading edge of computer speed and power for nearly

    a decade until stopped by the twin bottlenecks of ultra-large, ultra-high-speed systems: the thermodynamic limits of speed-power product and

    physical limit of the speed of light. Somewhat belatedly we noticed that

    power consumption is ultimately only the flip side of power dissipation, and

    device density is only the obverse of device-to-device path length. Closer

    still to home, most of us remember how the giant mainframes of IBM were

  • 8/7/2019 RISC+Reward_1988

    8/24

    blind-sided by Digitals first minicomputers, which the King of Computing

    once took no more seriously than, say, Digital took the first personal

    computers two decades later. And so,ad infinitum.

  • 8/7/2019 RISC+Reward_1988

    9/24

    On Complexity

    It is not the case that our problems are so complex, but that we have

    such small heads.

    Edsgar Djikstra in The Magic Number 7, plus or minus 2 1974

    It isnt ignorance that gets us in trouble its what we know that aint

    so.

    Will Rogers (Tulsa, Oklahoma) 1932

    To restate our primary thesis that the critical transitions in the

    history of technology, the greatest impacts on industries, economies andsocieties, occur when service-based functions are transformed in to

    product-based functions. In the Industrial Age context of the past few

    centuries, mechanizationtransformed predominantly human-based tasks

    into primarily machine-based ones more accurately machine-

    augmented, since machine power was still under the control of human

    intelligence. In the present information age context, whereautomation

    seeks to provide machine augmentation of human intelligence, the

    success of these attempts often depends upon the systems underlying

    philosophy whether the design seeks toaugmentor to replacehuman

    capabilities.

    Virtually without exception, only the augmentation approach achieves

    success: even the autoteller is less an automated replacement for a human

    bank tellerthan an augmented and simplified system interface for the bank

    customer. This redefines the basic design task (and the proble

    encountered) from a computationfocus on information management, to a

    communicationfocus on complexity management. Communication problems

    whether between humans and programs, programs and systems, systems

    and machines, or between machines in a network, programs in a system, or

    humans in an organization are all problems in complexity. Consider five

    arenas of competition in our industry, five points of concern, taken in sequence

    from bottom to top, beginning at the hardwarelevel, namely:

  • 8/7/2019 RISC+Reward_1988

    10/24

    RISC: Reduced Instruction Set Computing Impact on computer

    speed & standards

    Open Software: Portability, scalability & interoperability of systems,

    software & data

    Open Systems: Impact & exploitation of networking,

    multiprocessing & parallelism

    GUIS: Graphic User Interface Systems Impact on human user

    speed & standards

    Systems Integration: All of the above + managing the complexity

    they necessitate

    Fundamentally, each of these five topics is either the solution to aspecific complexity management problem or an example of such a

    problem or both. While this analysis could treat each of the five topics,

    space and time permit us to examine in detail only the first the

    problems, challenges and opportunities of RISC with an eye to tracing

    the common thread connecting all this conferences topics, namely the

    management of complexity.

  • 8/7/2019 RISC+Reward_1988

    11/24

    If complexity is the common problem then is software the common

    solution or just another problem? From the viewpoint of a hardware

    engineer or corporate management, software is at best a black art and at

    worst a black hole, devouring countless millions of man-months and

    woman-weeks each year with no guarantees of delivery on time, within

    budget or even as specified. Endless analyses of the software

    bottleneck all cite the need for better tools and greater discipline. Rarely

    do they address the possibility that we may be tackling problems of

    excessive complexity, those beyond our psychological limits upon the

    number of items and interrelations we can mentally manipulate: the seven

    (plus or minus) cited by computer-science savant Edsgar Djikstra. In

    short, the first step towards our goal may be change in the way we thinkof

    software (in specific) and complexity (in general.)

    Were we seeking an alternative definition of the RISC buzzword at a

    more general level, Reduced Intrinsic System Complexity would fit

    perfectly with Gerald Weinbergs popular writings on general systems

    theory, including An Introduction to General Systems Thinking, and Are

    Your Lights On, but really beginning most appropriately with The

    Psychology of Computer Programming, in 1969, when he was an IBM

    fellow. One of his basic premises is that predictability occurs at either end

    of the spectrum of complexity Newtons Laws of Mechanics holding

    sway in simple domains with few elements, Boyles and Boltzmanns Laws

    of Statistical Mechanics guaranteeing outcomes in innumerably complex

    domains with Murphys Laws governing the intervening domain of

    complex system behavior.

    In examining the four critical technologies besides RISC (original

    definition), we find that all five approaches attempt to circumvent Murphys

    Laws a goal that may seem like thwarting the Second Law of

    Thermodynamics. In reality each methodology attempts toreduce the

    number of possible states the system can assume in Information

    Theory terms to minimize its entropy to a complexity level our little

    Size 72 heads can manage:

  • 8/7/2019 RISC+Reward_1988

    12/24

    In Open Software, the IEEE 1003 POSIX standard and its

    progeny the N.I.S.T. Federal Information Processing Standard

    used in United States government contracts, the X/Open Common

    Application Environment embraced by European agencies and the

    ISO Joint Technical Committee 1, Subcommittee 22 Working Group

    proposal that hopes to grow up to be a Standard for all Seasons (and

    reasons) are primarily Application Program Interface

    specifications. Each seeks to spell out clearly and precisely what

    functions and services an application can expect and an operating

    system must offer the closest software equivalent to a hardware

    bus specification. Each frees the user from the complexities of how

    a program was written: forwhichsystem (portability) of whatsize

    (scalability) or evenwhen(maintainability). The converse(interoperability) treats similar complexities for datarather than

    programs.

    In Open Systems, the celebrated seven-layer ISO Open Systems

    Interconnect model guarantees to each level more critically, to

    the programmers writing for it a consistent and relatively simple

    interface to the level below it and above it and the right to ignore

    everything else as if each lower level were well-behaved

    hardware. The ultimate goal, particularly in the related areas ofmultiple-processor and parallel execution, is not only to free the user

    from the complexity of wherein the network the processing occurs or

    data resides, but from whichor how manyprocessors are used.

    In Graphic User Interface Systems, the portability-scalability-

    interoperability motif (lower case, again) is extended to managing the

    complexity of the human interface:

    a consistent look and feel makes expertise portableacross

    applications and reduces to a humanly manageable number the

    skills needed to use the system;

    a progressive command and control interface makes skills

    scalableacross levels of expertise, leading from simplicity for

    the novice to power for the expert;

  • 8/7/2019 RISC+Reward_1988

    13/24

    a regular repertoire of data models makes information

    interoperablebetween applications, from simple cut-and-paste

    to sophisticated interchange formats;

    a modular user interface makes systemsmaintainable,

    presenting a fixed level of abstraction independent of the

    underlying application, system or platform.

    In Systems Integration, arguably the most lucrative opportunity in

    the computer industry today, managing complexity is not the means

    to an end it isthe end itself. As H. Ross Perot proved with EDS

    and its billion-dollar facilities-management empire, To people with

    headaches, sell aspirin. The standard dismissal of this opportunity

    by manufacturing companies is that the field is a labor-intensive

    Service industry with finicky customers, fickle employees, faulty

    software and failure-prone hardware. But Perot realized that there

    are economies of scale possible in serving a large number of similar

    customers with fundamentally the same needs, that while it may not

    be feasible economically to build a complex set of tools, procedures

    and skills for one site, it may very well be practical to create such

    products if amortized across many customers.

    The RISC revolution offers a prime example of the turmoil of an

    industry in transition ours. Indeed, the computer industry is naturally

    fractious, both inter- and intra-company an understatement of the first

    magnitude and this alone could almostaccount for the way in which

    several large companies transition to RISC has nearly torn each of them

    asunder. IBM lost savants Joel Birnbaum (to Hewlett-Packards RISC

    design leadership), Glen Henry and Charlie Sauer (to Dell Computers

    secret task force) and even team-leader Andy Heller (to Kleiner/Perkins

    and even more secret services) and top manager Frank King (to Lotus).

    Digital lost VMS-creator Dave Cutler (to Microsoft) in its internal RISC

    wars, and the fabled loyalty of H-P customers was put to the ultimate test

    in the years of delay for Spectrum/PA. Still, it appears that something

    more fundamental is going on in the dark places of the mind.

  • 8/7/2019 RISC+Reward_1988

    14/24

    On Management

    A slow sort of country! [said the Red Queen to Alice.] Here, you

    see, it takes all the

    running you can do just to keep in the same place. If you want to get

    somewhere else,

    you must run at least twice as fast as that!

    Lewis Carrolls Through the Looking Glass 1872

    Managing complexity seems to be the common thread weaving

    together todays critical problems, the key issues of our industry

    perhaps in all industries or perhaps in the world. If that sounds like a

    grandiose, self-serving projection of our own concerns onto the face of the

    planet, could not one legitimately define the failure of the Communist

    system in Eastern Europe as a failure of central planning itself, an inability

    to manage the intrinsic complexity? In the recent hostilities in Iraq, were

    not the first targets attacked the command and control centers, the

    enemys complexity management systems; were not a substantial

    fraction of the Allied casualties from friendly fire a failure to manage

    Wars incredible complexities?

    From the Challenger disaster to the Hubble myopia, can we infer that

    the complexity of coordinating the best and the brightest individual and

    corporate talents is beyond NASA? Is not much of the yellow peril cited

    by the American Automotive, Steel and even High-tech industries, our own

    inability to match Japans uses of just-in-time inventory management, of

    statistical quality control systems, of advanced manufacturing lines lending

    themselves to robotics and automation, of dealer feedback systems tied

    directly to manufacturing design?

    Are these management problems an inevitable outcome of the

    complexity of our age? Three classical approaches to reducing

    complexity to manageable dimensions are neither unique to our times nor

    to our industries, but are found in the works of three managers who

  • 8/7/2019 RISC+Reward_1988

    15/24

    operated in the complex and competitive environments of more than

    twenty centuries past:

    Divide and Conquer: Julius Caesar, inBella Gallica, begins his

    classic description of the conquest of Gaul, ancient France, in 49 BC,with a simple declarative sentence: Gaul is divided into three parts.

    Even today, satellite photography would not yield such an insight; it

    was Caesar's understanding of the problem of conquering Gaul as

    divisible into three simpler problems that was part of his managerial

    genius.

    Know Thy Adversary: Sun Tzu, inThe Art of War, his definitive

    treatise of 202 BC, favors intelligence (military and otherwise) in

    approaching the problem. He ranks an understanding of theenvironment weather, terrain and logistics as comparable in

    strategic value to a knowledge of the strength and disposition of

    opposing forces.

    Keep It Simple & Straightforward: Alexander the Great, in legend,

    resolved the complexity of theGordian Knotby the simplest

    approach: a bold stroke of his sword. (It is rumored that even B-

    school graduates learn some variant of the KISS acronym.)

  • 8/7/2019 RISC+Reward_1988

    16/24

    On RISC

    RISC which is typically parsed as Reduced Instruction-Set

    Computer

    might be better understood if parsed as Reduced-Instruction-Set

    Computer.

    Dan Prenner (IBM, Yorktown Heights) 1989

    RISC the technology of the Reduced Instruction Set Computer in its

    various forms seems to have flashed upon the hardware talk circuit as

    the hottest topic of recent years, generating at least as much fire and

    smoke controversy and confusion as it has heat. This is not to saythat this new technology is not truly hot: measured in the rawMIPS

    the equally ambiguous acronymic units of Millions of [computer]

    Instructions Per Second RISC processors blaze past everything but the

    supercomputers they internally resemble. Rather, it is that RISC,in and

    of itself, is neither particularly new, novel nor even necessary except as

    the latest solution to a recurring problem once again, managing

    complexity.

    Reduced Instruction-Set Complexity the more liberal interpretationespoused by one of IBMs RISC pioneers in the quote above is only the

    most recent in a long sequence of hardwaresolutions to a problem of

    softwarecomplexity management that seems to emerge inevitably once

    a computer company reaches a critical mass or threshold complexity. At

    the end of the 1950s, with a decade-and-a-halfs worth of mutually

    incompatible system architectures and associated software to support,

    IBM was the first to reach this threshold or at least the first to recognize

    the problem and attempt to solve it. The solution began with an analysis

    of the mix of machine instructions actually used in running programs

    virtually the same analysis that led to RISC two decades later. Then,

    however, the usage pie chart suggested a single, multipurpose, full-

    circle (360-degree) virtual architecture. This democratic instruction set

    was ultimately implemented on a whole family of machines of that name

    by the all-to-common technique of each of the machines emulating the

  • 8/7/2019 RISC+Reward_1988

    17/24

    target instruction set but by design rather than default, using a control

    store or microprogram.

    At just about that time, Digital Equipment Corporation, lacking the

    golden albatross the large installed base to service, loyal customersto support and market to cannibalize burst upon the scene with the

    minicomputer. But by the end of their own decade and a half, with more

    than a half-billion dollars in annual sales and more than 30,000 copies of

    the PDP-8, PDP-11 and siblings installed worldwide, DEC similarly found

    themselves with the unmanageable complexity of too many machine types

    and operating systems. (The PDP-11 alone harbored RT-11, RSTS, RSX-

    11D, RSX-11M and variants.) They, too, examined the instruction set and

    leapt into the 32-bit domain with the VAX (Virtual Architecture eXtended)

    the classicComplexInstruction-Set Computer. Arguably, of several

    hundred proprietary CISC systems offered in 1976, only those of todays

    first- and second-place manufacturers (the IBM 360 [extended] and Digital

    VAX architectures) can survive into the next century.

    While it was half a decade later that the RISC principles (and acronym)

    were formally set forth by U.C. Berkeleys Glenn Patterson, the concepts

    were not new: Digitals 12-bit PDP-8 minicomputer and, a decade later,

    Intels 8-bit 8008 microcomputer each had an Instruction set reduced to fit

    their tiny brains the same problem noted in Djikstras quote. At the

    opposite end of the spectrum, supercomputer pioneer Seymour Cray had

    stripped the Control Data 6600 and 7600 instruction sets of all

    unnecessary complexities, facilitating instruction execution overlap

    (better known today as pipelining) even before 1970. Thus by 1986, when

    Hewlett-Packard had passed its own 15-year adolescence in the computer

    industry, the young but proven RISC approach Spectrum or Precision

    Architecture was the replacement of choice for the aging HP-1000,

    -2000, and -3000 instruction sets.

    Early that same year, RISC pioneer IBM fearing its golden albatross

    might be blown off course by the gale-force processor born in IBM

    temporary building 801, but wary of the rising winds of change

    launched the RT/PC as a (safely underinflated) trial balloon into emerging

    workstation markets. There it sank, hovered, drifted into commercial

  • 8/7/2019 RISC+Reward_1988

    18/24

    markets and ultimately rose above all but the very peaks of that terrain

    but never quite reaching the lofty slope of ideal IBM sales curves. Soon

    the semiconductor houses joined the fray, not only second-tier Fairchild

    and AMD, but Intel and Motorola placed their dominant microprocessor-

    specific market shares at RISC. In each case, whether RISC pioneer or

    risk-averse follower, whether carried by covered wagon or bandwagon,

    the companies involved were in their own eyes, at least

    unquestionably hardwaremanufacturers. With the advent of Electronic

    CAD and CAE (Computer-Aided-Design and -Engineering, respectively)

    they now had tools to build tiny devices of incredible complexity and/or

    speed, and each designer recognized the need to use it or lose it

    competitively. But to a hardware designer that AND/OR was an illogic

    gate, the near-mystical rate-limiting software bottleneck. To a hardwarevendor, the requisite softwareto managethis new hardware complexity

    was not within their control possibly not in anyones control.

    For despite the much-heralded birth of CASE (Computer-Aided

    Software Engineering) as Mark Twain might have opined the rumors

    of its birth were greatly exaggerated, the principal problems to be

    solved were no longer in the clean and pure realm of hardwarebut

    among the dark and murky black arts of software. The precipitating

    rationale for the 85%-commercial/15%-technical IBM 360 family wasconceived, the original reason the ultra-CISC VAX was born and the initial

    reason the H-P Spectrum saw the light of day was to solve what is actually

    a software problem the never-ending saga and ever-increasing costs of

    maintaining multiple OSs. Not surprisingly, the hardware vendors

    solution was not to address the software complexity problem directly but to

    move the complexity downward into a domain they didunderstand.

    Following Henry Fords engineering solution to the one-color problem,

    they slashed the Gordian knot of softwares uncertainty, complexity and

    flexibility by moving it into the hardware.

    Ironically, for well over a decade there had already been an alternative

    solution that captured the very essence of the VLSI revolution, the

    advantages of designing around standard, modular components

    interconnected in a uniform and controlled fashion without nearly the loss

  • 8/7/2019 RISC+Reward_1988

    19/24

    of flexibility, expansibility and power. Today such approaches are called

    open systemsor (less ambiguously) open software(lower case, please)

    but at that time simply called the UNIXoperating system. Even more

    ironically, the RISC revolution, which was largely a delayed but equal

    and opposite reaction to the hardware overcomplication and speed limits

    imposed by the CISC solution, pushed the responsibility for complexity

    back into the software, specifically into the compiler, the unquestioned

    forteof the UNIX OS.

    Most ironically, the number-one proponent and provider of advanced

    UNIX system software outside of Bell Laboratories by early 1988 was

    next-generation enfant terribleSun Microsystems. Though barelysix

    years old but already exceeding a billion dollars per year in revenue they

    saw the impending catastrophe by the time they had begun to implement

    their second machine architecture, the Intel-based 386i. In full

    accordance with politically correct open thinking, they preemptively

    developed and published the SPARCbinary design specification, a

    general-purpose Scalable Processor Architecture instruction set, a RISC

    for all reasons that everyone could (and properly should) employ, or so

    they sincerely believed. The true irony is that this movemayvery well

    have been a step backwards into the hardware solution by the most

    successful champions of the UNIX operating system. Despitemisconceptions of UNIX systems being for engineers and scientists, the

    UNIX forteis its standard arsenal of software tools compilers, editors,

    analyzers, debuggers, specialized languages, utilities the best software

    solutions to complex software problems.

    A glib diagnosis of Suns apparent aberration would attribute it to PC

    envy a morbid fixation on the market size of the IBM PC and its clones.

    The difficulty arises not from simple ambition but from the single-minded

    explanation of the PCs success almost entirely in terms of its plug-and-play, one-size-fits-all binary compatibility. This simplification manages to

    neglect such factors as relative market sophistications (or the PC markets

    lack thereof), total cost of ownership and usage or the level of complexity

    appropriateness to the task at hand. Realistically, PCs are priced and

    promoted as truly general-purpose devices with the all-too-subtle benefits

  • 8/7/2019 RISC+Reward_1988

    20/24

    of reducing complexity by reducing capability. When experienced

    computer users purchase a system from supercomputer to workstation

    they have a fairly good idea of the machines intended use. When a

    naive user buys a PC, even a sophisticated corporate user may have only

    a partial idea of the ends and purposes to which the PC will ultimately be

    put. This market difference, more than any technical limitation,

    determines that the majority of the PC software will be obtained in the

    aftermarket, with the volumes characteristic of such products. A more

    serious armchair psychoanalysis of the corporation if such a thing is

    possible suggests that the corporate self-image is that of a hardware

    company seeking hardwaresolutions wherever possible, even though fully

    85 percent of the technical staff (and at least that fraction of the value-

    added and differentiation from rival products) is insoftware.

  • 8/7/2019 RISC+Reward_1988

    21/24

    On Resolution

    When you have eliminated the impossible, whatever remains,

    however improbable,

    must be the truth.

    Mr. Sherlock Holmes in A. Conan Doyles The Sign of Four 1890

    Our analysis suggests that perhaps the turbulence we are

    experiencing signifies the boundary of a new domain, a new kind of

    sound barrier, a transition from the classical physics of compressible

    fluids, of the controllable complexity of the Hardware Paradigm. Far from

    an impassable wall in the sky, the complexity barriermay be a gatewayto a new domain where the laws of nature are simply different not

    better or worse. It may be that in this new domain complexities can no

    longer be handled sequentially and linearly but must be solved in parallel

    and nondeterministically, with all the uncertainty, chaos and surprise that

    nonlinear systems imply. Thus when (notif) the function of software

    today very much a service is sufficiently understood and sufficiently

    valuedto be treated as a product as hardware is today then not

    just the computer industry but the world itself may be a very different place

    indeed. But where shall we find the key to this transformation? How will it

    come about if we continue to solve the easy problems we understand

    instead of the hard problems we dont? And most certainly, the software

    problems are the hard problems. Just like the inebriate searching under

    the lamppost for the keys lost in the bushes, we continue to look where

    the light is better.

    Ultimately, the economic and intellectual capital must be ventured, the

    design tools sharpened, and the disciplines imposed. Butfirstthe

    fundamental way we thinkabout software must undergo a transformation,

    a paradigm shift back a dozen millennia, to before the comfortable, linear

    predictable risk-averse world of agriculture and citification. We must

    become familiar once again with the chaotic yet strangely attractive

    world of the hunter-gatherer, the exploratory nature mapped into our

    genes. Without yielding to fatalism or mysticism we may have to learn to

  • 8/7/2019 RISC+Reward_1988

    22/24

    accept the impossibility of absolute predictability once a critical level of

    complexity is reached, to go with the flow of uncertainty rather than

    attempting to conquer it. Even if we are uncomfortable with the idea of

    God playing dice at the macroscopic level, we may need to forego strict

    determinism as Djikstra proposes in his guarded command language

    to give up procedurally instructing the machine how to accomplish our

    goals, specifying only what objectives and endswe wish to accomplish,

    leaving successively lower and decreasing levels of complexity to

    determine the means. Whether achieved by by applicative fourth-

    generation languages, by object-oriented programming techniques, by

    standardized application, intersystem and human interfaces or (most

    likely) all of the above combined with ideas yet undreamed, we need to

    learn the larger lesson inherent in the RISC phenomenon and the X-windows phenomenon and the Open Systems and Software phenomena

    as applied to managing complexity.

  • 8/7/2019 RISC+Reward_1988

    23/24

    That lesson is that, increasingly, it is insufficient to escape from the

    linear only to project our trend lines into a geometric or even an

    exponential future. We must make a fundamental shift in our thinking, its

    dimension, if we are to have any hope of capturing the metatrendsthat

    allow us to stand on the shoulders rather than the toes of our intellectual

    forebears. One examination of a trend in our own industry should at least

    illustrate this concept. Consider, literally, the evolution indimensionof

    the relation between human and computer:

    Dimension 0: (circa 1960): apoint punched in a card in a deck

    slipped through the slot to a white-robed high priest of the batch-

    oriented mainframe, responding to ones JCL commands in hours

    or days and using a simple linear file system;

    Dimension 1: (circa 1970): aline of text typed interactively on a

    Teletype on-line to a time-shared minicomputer with response in

    seconds or minutes, a command-lineinterpreter, a simple line-

    editor and a more powerful hierarchical database;

    Dimension 2: (circa 1980): ascreen of text randomly accessible

    on a single-task personal computer with a response in fractions of

    seconds or seconds, a menu-bar command structure, character-

    oriented word processor and relational database;

    Dimension 3: (circa 1990): multiple overlapped windows of

    graphics and text on a multitasking workstation with a response in

    milliseconds, an object-oriented icon-based command

    structure, a WYSIWYG desktop publishing system and an

    entity-object database seamlessly incorporating text, graphics and

    images;

    Dimension 4: (circa 2000): ...left as an exercise for the readers

    own capabilities...

  • 8/7/2019 RISC+Reward_1988

    24/24

    Dr. Brian Boyle, currently Director of Research for NOVON Research

    Group in Berkeley, California, holds an MD (Board Certified in Radiology,

    Specialty in Nuclear Medicine and Medical Imaging,) and a PhD in

    Medical Information Science from University of California, San Francisco,

    an MSEE in Computer Science from the U.C, Berkeley, and a BS in

    Biophysics and in Experimental Psycholgy from Harvey Mudd College.

    Prior to founding NOVON in 1980 to investigate prospects for distributed

    Non-Von Neumann architectures, Dr. Boyle held positions in academia

    and industry and consulted for the scientific, commercial and government

    sectors. As Manager of Varian Associates Instrument Data Systems

    Division, he has worked with the UNIX Operating System since 1974.

    He investigated its usage and future managing the Systems and Software

    Group of McGraw-Hill's high-technology market research arm until 1985,when he returned to NOVON for its independent status as a research,

    consulting and publishing organization with the charter of exploring the

    next generation of systems and software technologies. An active member

    of the ACM, AAMI, IDBMA (Pick OS), IEEE, Usenix and Uniforum, he has

    participated in the creation of the IEEE 1003 POSIX standard since 1981

    and founded (in 1983) the Uniforum Internationalization Technical

    Advisory Committee, efforts he continues today as a frequent invited

    speaker at academic, technical and trade conferences worldwide.