The Risks Digest

The RISKS Digest

Forum on Risks to the Public in Computers and Related Systems

ACM Committee on Computers and Public Policy, Peter G. Neumann, moderator

Volume 12 Issue 49

Monday 14 October 1991

Contents

Nuclear Computer Safety Fears
Antony Upward
Computer Error by Policeman
Antony Upward
Thermostat failure mode
Liudvikas Bukys
I really like banks -- world wide!
Boyd Roberts
I'm sorry, the computer says your credit is bad
David Bremner
"Who Flies the Plane?"
Ken Tindell
Risks of Enterprise-Wide Phone Systems
David Fiedler
AT&T Outage
Jerry Schwarz
Re: "AT&T `Deeply Distressed'
Flint Pellett
Re: External risks to computer systems
Peter Mellor
Re: Keeping people in the loop
George W. Leach
Re: ``Friendly'' (?) viruses
Brandon S. Allbery
Paul Smee
Re: buggy software
James B. Shearer
Re: Security Criteria, Evaluation ...
David States
Re: Software migration at Johnson Space Center
Richard H. Miller
Tim Parker
Info on RISKS (comp.risks)

Nuclear Computer Safety Fears

KPMG - Antony Upward,IVC
14 Oct 91 11:03 GMT

>From the London Independent on Sunday, October 13, 1991

Computer Watch On Nuclear Plant Raises Safety Fears
by Susan Watts, Technology Correspondent

Fears are growing that computer software designed to protect the Sizewell B
nuclear reactor from serious accidents is too complex to check.

Sizewell B will be the first nuclear power station in the UK to rely so heavily
on computers in its primary protection system.  A computer-controlled safety
system was seen as superior to one operated by people because of the risk of
human error.

But Nuclear Electric has told the Independent on Sunday that the system for
Sizewell B, based around some 300-400 micro-processors, is made up of modules
which in total constitute more than 100,000 lines of code.

A number of software engineers close to the project are known to have suggested
that the software is now so unmanageable that is should be scrapped, and the
whole system built again.

The Nuclear Installations Inspectorate (NII), the public's watchdog on the
nuclear industry, has taken the unusual step of publishing the safety
requirements it is asking of Nuclear Electric, the company that will operate
Sizewell B, before the utility can expect a licence to go ahead with the power
station.

This is an attempt to calm the mounting anxieties of specialists in
"safety-critical" programs such as the protection system at Sizewell - where
lifes are at risk should the software fail.

Two senior inspectors describe the watchdog's requirements in a paper in the
latest issue of the trade journal Nuclear Engineering International.  The paper
is unusual because the NII traditionally keeps its options open when decided
whether to grant a nuclear power station a licence.  The onus is on the
operator to prove that its system is safe.  Publishing this description of its
requirements gives a clear idea of what the NII expects of Nuclear Electric.

Independent experts in safety-critical software are not happy with the NII's
safety requirements.  They say the paper shows the inspectorate is not asking
Nuclear Electric to use the most stringent testing procedures currently
available to prove that the software will work as specified.

They also criticise the inspectorate for not insisting on the most up to date
mathematical analysis that could give an indication of the software's
reliability.

These critics what Nuclear Electric to publish the results of its own internal
assessments and those of independent consultants whose date would give the rest
of the industry a chance to see just to how reliable the protection software is
meant to be.

The British Computer Society says it would welcome the chance to comment of the
safety case for the software.  It is concerned about what it sees as "the
secrecy which surrounds the safety-critical software in the Sizewell B control
and production systems".

David Parnas, an advisor on a similar project at a nuclear reactor in
Darlington in Canada, agrees.  "If somebody is introducing a technology with as
bad a reputation as software, then they are obliged to show that they have done
a really thorough analysis."  Given the public interest, he says, the results
should be published.

The main worry is that the software, being produced by Westinghouse, an
American company, is so large and complex that it is impossible to verify that
is would react as it should if the reactor behaved dangerously.

A simpler system would be easier to verify and to maintain.  But it would be
difficult, though not impossible, to find a politically acceptable route
whereby the existing system could be scrapped and development started again
from scratch.

The software is thought to have reached its size because it has many extra
features which, although desirable, have complicated its structure and blurred
the distinction between the software which controls the nuclear reactor and
that which protects it.

David Hunns, superintending inspector at the NII, and one of the authors of the
recent paper, says that this distinction is "fundamental".  Nuclear Electric
insists that independence between the two systems "is fully maintained".

Mr Hunns adds: "We don't think it [the system] should be scrapped.  We believe
a safety case can be made.  But it has to be proven.  We have made a judgement.
It's an honest and it's heart-searched judgement.  We've looked hard at the
technology - all the aspects that we possibly can.  We've developed a rationale
that is expressed [in the paper].  Providing the elements of that frame work
are fulfilled that [Nuclear Electric] will make it.  If they are not fulfilled,
they won't.  Then maybe they will wish they had started in  another direction."

  [Subsequent to the above, the following appeared -- giving what I thought to
  be a very good introduction to the complex problem of testing, and how it
  relates to the Sizewell B Nuclear Power Station.]

>From the London Independent on Sunday, October 13, 1991

A Complex Problem of Tests, by Susan Watts, Technology Correspondent

Software engineers agree that it is not safe to assume that software will
always operate correctly - especially in systems such as nuclear power stations
where people's lives are at risk if the software fails.

But computer programs are notoriously difficult to verify.  The chief anxiety
over the protection system for Sizewell B is that it is very large and complex.
This bucks the trend in safety critical software, which is to built small,
simple systems.

It is usually impossible to test software for all combinations of the inputs it
may receive - say from the reactor in a nuclear power station.  It would take
thousands of years to test the most simple system.

An alternative approach is to rely on statistical analysis of the system to
give probabilities of its reliability.  This would involve running the system
for long enough to get an idea of the chances of it failing.  But even this
could take many years for a large system.

A more recent technique is to use so-called "formal methods".  This involves
converting the specification for a piece of software into a precise
mathematical model of the requirements to "prove" that it works.  But formal
methods are a new idea;  the software for Sizewell B began life some eight
years ago, when the concept of formal methods was embryonic.

David Hunns, superintending inspector at the Nuclear Installations Inspectorate
and co-author of a recent paper on protection systems for Sizewell B, says the
case for insisting on formal methods is not clear cut.  In his paper Mr Hunns
says "the NII has accepted that is not reasonably practicable to achieve their
incorporation for Sizewell B".

He adds that the viability of a second best approach, which involves
"backfitting" the mathematical techniques of formal methods to a completed
piece of software, also "remains unresolved".

Independent software engineers disagree.  This method of backfitting is being
used to verify software at a nuclear reactor in Darlington in Canada with a
similar, although far smaller, computer-based protection system.

Nuclear Electric has designed its system such that is should fail no more that
once in every 10,000 "demands" (a demand is when the reactor is in a state
where the software should shut it down).

It is just about feasible that a system with such a failure rate could be
simulated by building another  computer program to replicate the reactor,
hooking that up to the protection software and physically testing whether the
protection system behaves as it is designed to.  This approach would be
time-consuming, and expensive.

Nuclear Electric is building a scaled down version of such a test system, and
expects results next summer.  The test involves a prototype of one channel of
the protection software, but will not run for long enough to be described as
exhaustive.  Mt Hunns insists that if the simulation shows the software
contains an unacceptable number of errors, NII will ask Nuclear Electric to
extend the test.

  [Antony Upward, KPMG Management Consulting, Software Development Group,
  8 Salisbury Square, London, UK EC4Y 8BB  Phone:  +44 71 236 8000]


Computer Error by Policeman

KPMG - Antony Upward,IVC
14 Oct 91 11:03 GMT

>From the London Guardian Friday October 11, 1991
Computer Error by Superintendent

A police superintendent discovered his former wife had a new man after checking
a car through the Police National Computer, Bow Street Magistrates court,
London, heard yesterday.  Leslie Bennett, aged 44, based at Chelsea, west
London, asked another officer to access the computer about the car which his
former wife said belonged to a friend of their daughter's But the vehicle's
details related to the wife's firm.  His daughter, Jane, then told him the man
was her mother's "new friend".  Mr Bennett was found guilty of an offence under
the Computer Misuse Act 1990 and fined =L150 with =L250 costs.  The case
followed a complaint by Mrs Bennett.

  [ADDED NOTE: We learned on 13 Jul 2009 that Mr. Bennett's case was later
  appealed and upheld; Mr Bennett received his fine and costs back.  PGN]


Thermostat failure mode

<bukys@cs.rochester.edu>
Mon, 14 Oct 91 12:05:48 EDT

I have a typical electronic setback thermostat installed.  A couple of nights
ago it failed "on", causing my furnace to run and run, until my three-year-old
woke up and came to tell me that she was hot.  The temperature had reached 92
degrees(F).

The thermostat itself had decided that it was still 68 degrees(F).  Rebooting
the thermostat by removing and re-inserting the batteries made it get back in
touch with reality.  I replaced the batteries too, but, considering it had
enough power to run the LCDs, that's probably not it.

There are electronic setback thermostats that mount over existing mechanical
(mercury switch) thermostats.  I always thought it was silly to have a little
motor move the arm up and down to cycle the furnace.  But now I have to wonder
about what temperature my electronically-controlled furnace would be driven my
house to before either reaching thermal equilibrium or igniting or melting
something, especially if left to itself over a vacation.  (I don't know yet
whether the furnace itself has its own thermal shutdown, but I doubt it.)  At
least with a mercury switch in there, the most extreme setting is probably
still below where my good old all-electronic device would have taken my house.

Liudvikas Bukys  <bukys@cs.rochester.edu>


I really like banks -- world wide!

Boyd Roberts <boyd@prl.dec.com>
Mon, 14 Oct 91 11:09:33 +0100

I'm Australian, but I now live in France.  In the past few months I have had
some really gnarly problems with banks in Sydney, Paris, London and the West
Coast.

My `new' bank refuses to give me a cheque book, but they don't bother to tell
me why.  I got this missive from one of the administrators here today:

    Hi Boyd,

    My telephone discussion with your bank this morning was another
    piece of surprise !

    Your checkbook request was refused because records showed Mr.
    Roberts was under national bank interdict !!!  I certainly
    refused such a statement and discussed your personal data
    with them.  What actually happened is that a confusion
    was made between two different Mr.Roberts (I do not understand
    why, as nationality, birth date, address, are totally different ...)

    Mrs.[deleted] is requesting today interdict removal on your name
    and getting a checkbook for you.  Checkbook should be ready
    by next Monday.  I'll keep you posted, as usual.

Boyd Roberts			boyd@prl.dec.com


I'm sorry, the computer says your credit is bad

David Bremner
Mon, 14 Oct 91 14:29:03 PDT

>From an article on financial software in DEC Professional:

"Expert systems is [sic] another rapidly growing area in financial software,
notably in real-time applications found in banking, insurance and accounting
venues.  Inference, for example, has developed expert systems credit approval
software Dun & Bradstreet and Swiss Bank"

Ah yes, I can see it being a major competitive advantage to be able to
propagate data-entry errors in "real-time" :-)

Reference: p. 54, Oct. 1991 DEC Professional  	          ubc-cs!fornax!bremner


"Who Flies the Plane?"

<ken@minster.york.ac.uk>
14 Oct 1991 16:23:59 GMT

I recently caught a television programme on Channel 4 in the UK called "Who
flies the plane?" (part of a consumer affairs series of programmes). The
programme dealt with some of the issues of software in fly-by-wire commercial
aircraft, particularly the `human factors' problem. A number of notable people
in the aviation field were interviewed.

I wrote to Channel 4 and asked if I could have a transcript of the programme,
and permission to post excerpts to RISKS ("RISKS is a highly regarded
international forum on software safety .. blah blah .. welcome comment .. blah
blah .. "). The terse reply was along the lines of "The transcripts are too
costly, and Channel 4 owns the copyright, so no". And the RISK of this story?
I believe these programmes to be sensationalist, interested in viewing figures
rather than rational discussion of the issues.

Ken Tindell, Computer Science Dept., York University, YO1 5DD UK
+44-904-433244


Risks of Enterprise-Wide Phone Systems

David Fiedler <david@infopro.UUCP>
Sat, 12 Oct 91 15:19:14 PDT

The other day, my wife called our local bank to discuss refinancing a
loan. Her call was transferred to a loan officer, and she made an
appointment for us to meet at the bank to discuss matters. When we got
there, nobody at the bank had ever heard of the loan officer. It finally
developed that the loan officer was based at another branch 35 miles away.

When transferring phone calls within a company's phone system is so easy,
customers have no way of knowing that "the bank" they were talking to was
somewhere else. Perhaps phone systems could be designed for the office
personnel to be notified when a call has been transferred from another
location, by a special ring or tone on the line.

David Fiedler       UUCP:{ames,bytepb,mrspoc}!infopro!david
USMail:InfoPro Systems, PO Box 220 Rescue CA 95672 Phone:916/677-5870


AT&T Outage

Jerry Schwarz <jss%summit@lucid.com>
Mon, 14 Oct 91 16:06:01 PDT

It is easy to focus on the proximate causes of accidents and even on the
general system level causes.  But the recent AT&T outage in NY is a perfect
opportunity to ask a question about very high level causes that has bothered
me for a while. Can the recent rash of failures in the phone system be traced
to divestiture and price competition in the phone business?  When new
technology (such as software of fiber optic cables) fails it is hard to
address this question.  But here we had a failure in one of the oldest
technologies in the business.  So we can ask the specific questions: Have
procedures or staffing levels in "power" changed since divestiture?  Would
pre-divestiture procedures or staffing levels have prevented the recent
outage?

I have no particular knowledge of this area.  Perhaps someone who does would
care to address my question.
                                        Jerry Schwarz   jss@lucid.com


Re: "AT&T `Deeply Distressed' (Niland, RISKS-12.48)

Flint Pellett
14 Oct 91 19:30:31 GMT

Why have an all-or-nothing user interface?  If you're at 105% of design
tolerance for G force, it isn't the same thing as exceeding the design by
200%.  Tell the pilot by how far the design is being pushed, and let them
decide if the risk is warranted by the situation.  The machine's job is to
make sure all the information necessary to produce a good decision is
available to the pilot, not to suddenly at some arbitrary cut-off point start
making the decisions for them.

Flint Pellett, Global Information Systems Technology, Inc.,
1800 Woodfield Drive, Savoy, IL 61874 (217) 352-1165 uunet!gistdev!flint


Re: External risks to computer systems (Rose, RISKS-12.47)

p mellor <pm@cs.city.ac.uk>
Fri, 11 Oct 91 17:41:49 BST

> If you don't what catastrophic failures, you need to arrange things so that
> the inevitable failures aren't catastrophic. (P.Rose in RISKS-12.47)

Despite the possible consequences of system design (including software) faults,
it is still the case that fire, flood, and other "natural" disasters can be
far more disastrous for a computer system, particularly a centralised one.

Earlier this week, the London Evening Standard carried a very small paragraph
reporting a fire at Hitchin College of Further Education. This drew my
attention, since it is a few miles from my home, and my ex-wife works there as
a secretary.

The fire started at 2 a.m., and was not discovered for an hour. By then the
building complex housing the total computing facilities of the college, both
educational and administrative, had been gutted. The back-up tapes, needless to
say, were stored on-site, close to the computers, in a non-fireproof cupboard.

Total damage to computer equipment was quoted in the paper as 10 million
pounds. This seemed a bit high to me (for a moderately-sized college with
no large mainframe), and the figure probably includes other damage (library,
dance studios, video equipment, etc., etc.) and a finger-in-the-air guess at
consequential loss. The last item seems to be incalculable, however, given
that the college's financial and student records have all been wiped out.
It is on the cards that Hitchin College will cease to exist as a separate
institution.

My ex-wife was surprised that the college had not been regularly exchanging
back-up media with its associated institutions in Letchworth and elsewhere.
I replied that this is exactly what I recalled ICL doing: 3-level back-up
cycled every day between sites, and all tapes stored in locked fire-proof
safes. I then recalled why ICL adopted this admirable policy. In the early
70s they lost an entire installation, including all back-up, when the night
operator dropped a cigarette into a waste-paper basket.

Do we always have to learn the hard way?

Peter Mellor, Centre for Software Reliability, City University, Northampton Sq.,
London EC1V 0HB +44(0)71-253-4399 Ext. 4162/3/1 p.mellor@uk.ac.city (JANET)


Re: Keeping people in the loop (Re: Thomas, RISKS-12.48)

George W. Leach
Mon, 14 Oct 1991 13:50:12 GMT

>We could make the humans the prime operators, and use the computers as a
>back-up...

               This is exactly how the monorail system at Walt Disney World is
operated.  There is a human driver who controls the speed of the train.  There
are speed zones along the lines where, if the operator fails to keep the speed
under a prescribed value, the train will be shut off automatically.  However,
the human is the primary operator.

George W. Leach, AT&T Paradyne, Largo, FL 34649-2826 USA 1-813-530-2376


Re: ``Friendly'' (?) viruses (Smee, RISKS-12.45)

Brandon S. Allbery KF8NH
Sat, 12 Oct 91 18:41:43 -0400

> Most people reading this use a machine running Unix. Somewhere in its file
> system (usually /usr/spool/cron or /var/spool/cron) there is a directory
> `crontabs' containing files which describe actions to be executed regularly
> without explicit user action.

Granted --- but the system administrator should check these actions both when
installing the system and periodically thereafter.  (I do so once a month; a
more secure environment would require more frequent checks.)

That many novice Unix sysadmins do not check their crontabs (by the way, you
neglected to mention the V7 and BSD /usr/lib/crontab) is a RISK, but is in
essence their fault.  That modern "plug-and-play" Unixes do not provide an easy
means to do this for the novice sysadmin is a contributory factor; this can be
likened to the off-and-on "fly by wire" discussion on this list ("system
administration by wire"?).  Modern Unix systems do not take this into
consideration, mainly because the systems of the past did not --- but those
systems required competent system administrators anyway, so it was not a
problem then.

Brandon S. Allbery allbery@NCoast.ORG uunet!usenet.ins.cwru.edu!ncoast!allbery


Friendly (?) viruses (Meyer, RISKS-12.47)

Paul Smee <P.Smee@bris.ac.uk>
Sun, 13 Oct 1991 16:08:56 GMT

I think you are taking me a bit TOO literally.  True, there are things which
run on 'my' systems which I don't explicitly know about each instance of.
There are also things (user programs) which I generally don't know about
at all, but which are necessary to our service.

The point is that (as an admin), for every 'proper' thing running on the
system I can, if necessary, point a finger at who is responsible for it,
who DOES know about it (if I don't), and who is responsible for making
sure it behaves sensibly (if that's not me).  It's not that I know everything
that is happening, but I can, without needing to use heroic or unusual
procedures, find out what I need to know about it, and where I can get
help if required, in order to ensure continued service.

>It also seems that an automatic or semi-automatic bug correction service,
>working somewhat in the style of mail and news (that is to say, updating remote
>files in controlled conditions) wouldn't be such an absurdity as he suggests.

The salient points here are your 'in controlled conditions', and (in a bit
I've cut) providing that the machine owner/operator/administrator has
`subscribed' to such a service.  (And that the group providing the service can
control it to ensure that only those sites which want it get it.  We don't
even install official manufacturer-provided upgrades without first evaluating
them under test conditions, to make sure they don't interact unfortunately
with other things we run.  It's surprising how often something can't be put up
exactly as supplied, without requiring other work.  I've long been in favor of
automagic DISTRIBUTION of bugfixes, rather than having to wait for the
semi-annual release tape.  But, with the present state of the art, I want to
look at them before I put them in.)


Re: buggy software (Parnas, RISKS-12.47)

<jbs@watson.ibm.com>
Sat, 12 Oct 91 19:49:36 EDT

         So far as I know no one is required by law to buy the products of Mr.
Mitchell's company.  If mature adults wish to buy buggy software I do not see
why this should be any concern of Mr. Parnas.
         A real risk is that laws will be passed requiring people to use
certain crackpot programming methodologies which purport to be better than
existing practice but which for some strange reason people refuse to adopt
voluntarily.
                          James B. Shearer


Re: Security Criteria, Evaluation ... (Spencer, RISKS-12.48)

David States
Sun, 13 Oct 91 20:53:58 GMT

> What, exactly, is the fundamental difference between a
> time-sharing system and (say) a heterogeneous network?
> Answer: there isn't one...

In a closed, time-shared system, if the kernel is secure, all of the kernel
mode communications are secure.  In a network environment, kernel messages
must travel over communication paths which are not guaranteed to be secure.
How long would a time-shared OS remain secure if user programs could monitor
queries made by kernel mode system calls to kernel databases?

By this line of reasoning, public key encryption is essential to the
development of reliable network based computing systems.  If you can't rely on
secure communications to distribute a key, there is not an alternative.
Distributed networks are almost by definition insecure.
                                                             David States


Re: Software migration at Johnson Space Center

Richard H. Miller
Sun, 13 Oct 1991 20:18:07 CDT

This is starting to get off the the strict area and into computer religion was
but I do need to make the following points: [We run both 2200 and VAX systems.]

1) The high end VAX does not even begin to compare to the high end Unisys
systems. A 2200/644 is a large mainframe system and the new 2200/900 is even
more powerful. A large VAX 9000 will not provide the same level of performance
as a fully configured 2200/600 or 2200/900.

2) A Unisys machine will run absolutes created 15 years ago. There is no
requirement for recompiling or relinking.The same application code will run
and most of the same system control software across all processors.

Richard H. Miller, Asst. Dir. for Technical Support, Baylor College of Medicine
One Baylor Plaza, 302H                Houston, Texas 77030 Voice: (713)798-3532


Re: Software Migration at Johnson Space Center (Sherr, RISKS-12.48)

Tim Parker
14 Oct 91 21:01:08 GMT

Where are the Silicon Graphics advocates???  From my recent research, SGI has
a line of equipment much better suited to simulation than anything in the
VAX line (more horsepower, better graphics)- at a small fraction of the price.

Tim Parker - Independent Consultant


Informatik journal available

Duane
Fri, 11 Oct 91 10:23:52 CDT

Announcing the first issue of 'Informatik,' a journal of free information.
Currently available by FTP from: uunet.uu.net       /tmp/inform1.Z
                                 ftp.cs.widener.edu /pub/cud/misc/inform-1.1.Z

Here is an excerpt from the introduction:

/* Introduction */     By the Informatik staff

      Welcome to the inaugural issue of Informatik, an electronic periodical
devoted to the distribution of information not readily available to the public,
with a particular emphasis on technology and the computing world.  First and
foremost, this publication is dedicated to the freedom of information.  This
journal is made possible by The First Amendment of the U.S. Constitution which
states:

      Congress shall make no law respecting an establishment of religion,
      or prohibiting the free exercise thereof; OR ABRIDGING THE FREEDOM
      OF SPEECH OR OF THE PRESS; or the right of the people peaceably to
      assemble, and to petition the Government for redress of grievances.

In this and coming issues, we plan to exercise our First Amendment rights to
the best of our ability.  We will print feature articles on hacking, phreaking,
and various other illicit activities.  We also plan on bringing you recent news
and gossip from the underground, anything news of interest to hackers,
phreakers, grifters, cyber-punks, and the like.  Informatik will also provide a
plethora of information on the inner workings of corporate America and the U.S.
Government.

DO distribute this freely! Remember this is not illegal, this is information.

*Please send submissions and comments to duane@shake.tamu.edu. (for now)*

                                            Mack Hammer & Sterling [Editors]

Please report problems with the web pages to the maintainer

Top