Introduction to Software Engineering

Lecture 3 - Important characteristics of large-scale systems

Jim Briggs, 8 October 1998

Software evolution (Sommerville, 5th edition, chap 32)

Large software systems are not static objects. Their environment changes, and software must adapt or become progressively less useful (then discarded). Process termed software evolution by Manny Lehman (Imperial College).

Maintenance accomplished by series of releases.

If system installed in several different places, unlikely that all installations will include all changes in each release. Furthermore, each installation may make local changes. (Each installation may have different priorities.) Consequently, versions of the system drift apart. This divergence makes future maintenance more difficult.

Lehman's laws (hypotheses derived from observation) of program evolution:

  1. Continuing change: a program must change or become less useful. "Inevitability". Introducing a system into an environment changes the environment. Users in the modified environment then modify their expectations of the system.
  2. Increasing complexity - degraded structure: as a program changes its structure becomes more complex (pure structure corrupted) unless actively avoided. Need to spend money at some point to make system (by restructuring) more adaptable in future.
  3. Program evolution: large systems act like an inertial mass (size inhibits changes because changes introduce new faults, new requirements, etc.). They have a dynamic of their own established early in the development process. Maintenance management cannot override it (inhibits desired major changes because cost too great, perhaps beyond local budget, take time to make, results of organisational decision-making process). Result of inertia in human organisations.
  4. Conservation of organisational stability: the rate of development of a program (over its lifetime) is approximately constant and independent of the resources (e.g. staff) devoted to it. Most large programming projects work in a "saturated" state - change of resources has imperceptible effect on long-term evolution. Large projects are unproductive because of communication overheads.
  5. Conservation of familiarity: the incremental change in each release is approximately constant. Release with large number of changes will probably be followed quickly by another release with small number of changes to fix bugs introduced. Suggests need to interleave enhancement releases with repair releases. There is a limit to the rate at which new functionality can be introduced.

"Laws" based on observation but not (yet) validated properly. Do seem sensible and probably not best to try and buck them. 3-5 may not apply in smaller organisations.

Implications of above:

Reliability v. efficiency (Sommerville, 5th edition, chap 18)

Reliability more important than efficiency because:

Measures of reliability such as "mean time between failures" do not take into account subjective importance of fault. An unimportant fault occurring frequently might not give a system as poor a reputation for reliability as a rare fault that killed someone.

Therefore, important consideration is perceived reliability.

System reliability (in general) is dependent on correctness of design, correctness of mapping design to implementation and reliability of components. Car components wear out, therefore car cannot be 100% reliable. Software does not wear out, therefore reliability depends completely on design and implementation correctness.

Hardware reliability can be improved by duplication - can take similar approach with software. However, no point in executing identical copy of same procedure. Randell suggests that software reliability may be improved by executing a different but functionally equivalent procedure (recovery blocks).

Example of space shuttle system: four processors executing identical software; fifth (different) processor executing independently developed software.

Thus key to software reliability lies in system specification. Frequent problems with reliability requirements:

Result: lack of perceived reliability may be result of misunderstanding between developer and customer.

Even more difficult when using exploratory programming since no specification exists. Verification techniques of no use. For this reason, exploratory programming in safety-critical environments is not recommended.

Default: work to "fail-safe" criteria. Program should:

Do use of formal methods necessarily lead to more reliable systems? Formal specification less likely to contain anomalies but opaqueness of notations may make it more difficult for customer to establish whether system is what is required; therefore, less likely that reliability criteria will be met. If cause of unreliability is environmental then even a formal proof will not ensure reliability.

High reliability inevitably involves extra, otherwise redundant code to perform checking. Hence efficiency is reduced and development costs increase.

Difficult to quantify reliability, so some organisations place constraints on development process on the assumption that adherence to standards will lead to systems whose reliability is acceptable. Danger is that process becomes more important than costs or functionality, and there is no firm evidence concerning relationship between product and process reliability.

Summary

  1. software engineering involves technical and non-technical issues such as human factors and management;
  2. well engineered software provides services required PLUS is maintainable, reliable, efficient and has appropriate user interface;
  3. the waterfall model of software development suffers from inadequacies but will continue to be used;
  4. exploratory programming is not suited to development of large, long-lifetime or safety-critical systems;
  5. software systems have to be maintained;
  6. most important characteristic of software is reliability since cost of failure often exceeds cost of system.