Subscribe Now Subscribe Today
Research Article
 

Personal Construct Psychology (PCP) Expert Systems



B. Venkata Raju and S.P. Rajagopalan
 
Facebook Twitter Digg Reddit Linkedin StumbleUpon E-mail
ABSTRACT

The study on expert systems led to the identification of the ‘knowledge acquisition bottleneck,’ that it was generally extremely difficult to make overt the presumed knowledge of human experts in order to program it for computers. The history and reasons for the adoption of repertory grid methodologies and tools to overcome the knowledge acquisition bottleneck are described. Then a more fundamental analysis is made of why expert systems to date have had only limited success and merits of a personal construct approach to emulating human expertise in greater depth than has been achieved with existing cognitive science models are presented. In conclusion, it is noted that the techniques developed to emulate human expertise are essentially ones for modeling and emulating any person’s psychological processes, not just those of people valued by others as experts. PCP-based expert systems methods and technology have wide relevance, for example, in clinical and educational research and applications. The role of personal construct psychology in computer research and applications concerned with the development of expert systems and their beginnings in artificial intelligence and cognitive science are covered in this study.

Services
Related Articles in ASCI
Search in Google Scholar
View Citation
Report Citation

 
  How to cite this article:

B. Venkata Raju and S.P. Rajagopalan, 2007. Personal Construct Psychology (PCP) Expert Systems. Information Technology Journal, 6: 232-236.

DOI: 10.3923/itj.2007.232.236

URL: https://scialert.net/abstract/?doi=itj.2007.232.236

INTRODUCTION

Research on programming computers to think: The arrival of the first commercial digital computers in the 1950’s led to widespread interest in the potential applications of computing. The use of the term ‘giant brains’ became common in the press although it was clear that the precise, logical operations of computers had little in common with the human brain. However, interest in simulating human thought processes was common among the early computer pioneers. Alan Turing, a brilliant Cambridge logician who had helped develop computers to break enemy message encryption during the second world war, wrote a paper on Computing machinery and intelligence for the journal Mind in which he considers the question Can machines think? (Turing, 1950). He answers it in behavioral terms, proposing what has come to be knows as the Turing test, that if a person communicating with the computer and with another person through the same medium (such as communicating Tele_ printers) cannot distinguish them correctly then the machine, for all practical purposes, can be said to think.

Research on programming computers to think became widespread and McCarthy et al. (1955) proposed to the Rockefeller Foundation that it fund a study of artificial intelligence to be carried out during the summer of 1956 at Dartmouth College, Hanover in the USA. The year of this proposal also saw the publication of Kelly’s seminal work on personal construct psychology but, the pioneers of artificial intelligence and cognitive science never became aware of this work. The next decade was also the era of the development of ‘computer science’ as a new field of study when computers were very expensive and university and funding agency budgets were hard-pressed to supply the demand for computer facilities. In Britain the competition between those wishing to undertake research in computer science and in artificial intelligence was so intense that the UK Science Research Council commissioned Sir James Lighthill to report on the state of the art in machine intelligence. His report (Lighthill, 1973) was damning about both the achievements and the prospects for such research and had a strong negative influence world-wide on funding for research to program computers to think (Fleck, 1982).

Embattled AI researchers focused on specific goals to develop programs that emulated human expert performance in fields of obvious practical value such as mineral exploration and medical diagnosis and in the mid-1970’s announced a number of breakthroughs in what came to be called expert systems (Michie, 1979). The first successful expert systems were DENDRAL (Feigenbaum et al., 1971) for reconstructing molecular structures from mass spectrometer data and MYCIN (Shortliffe, 1976) for diagnosing microbial infections from medical data. The systems were programmed as collections of production rules that expressed a relationship between a premise and a conclusion such that if the conditions of the premise were satisfied then those of the conclusion could be drawn. For example, a rule from MYCIN is:

If

The infection is primary-bacteremia and
The site of the culture is one of the sterile sites and
The suspected portal of entry of the organism is the gastro-intestinal tract

Then there is suggestive evidence that the identity of the organism is bacteroides.

Such rules are obtained from specialists in microbial infections and their application to particular data is fairly simple data processing. The rules are validated through their application to many cases and revised when they fail to give the correct diagnosis. MYCIN was designed to interact with a clinician in order to make a diagnosis and suggest therapy for a particular patient with a suspected microbial infection. It first gathers data about the patient and then uses this to make inferences about the infections and their treatment.

The success of the early expert systems attracted industrial and research attention and a major industry developed in the early 1980’s. The objectives of the study were then defined by one of the commercial AI pioneers, Hayes-Roth (1984), in a workshop on AI Applications for Business in May 1983. He enumerated some situations appropriate to expert systems, such as: the organization requires more skilled people than it can recruit or retain; job excellence requires a scope of knowledge exceeding reasonable demands on human training and continuing education.

As a modern example of the success of expert systems technology, the April and July 2000 issues of InTech Magazine published by the Instrumentation, Systems and Automation Society, has a two-part paper from Eli Lilly on the use of an expert system in its fermentation plant (Alford et al., 2000). The evaluation in 2000 corresponds well to Hayes-Roth’s predictions in 1983. Within a few weeks, the expert was satisfied that the expert system reliably came to the same conclusions he would have by looking at the same data. The expert system then took over this part of the expert’s job, freeing up 40 h per month of his time for other work.

The knowledge acquisition bottleneck: Expert systems appeared at first to be a major validation of the possibility of digital computers being able to emulate human thinking and there is continuing evidence of some successful applications. However, the industry has not grown to the extent predicted, largely because programming such systems has been very much more difficult than expected. Feigenbaum et al. (1970), one of the pioneers of expert system, termed this the knowledge acquisition bottleneck. Hayes-Roth et al. (1983) in their book on Building Expert Systems noted that, since the programmer has far less knowledge of the domain than the expert, communication problems impede the process of transferring expertise in a computer program. The vocabulary initially used by the expert to talk about the domain with a novice is often inadequate for problem-solving; so that the programmer and the expert must work together to extend and refine it. One of the most difficult aspects of the programmer’s task is helping the expert to structure the specialist knowledge, to identify and formalize the expert’s concepts.

From a personal construct perspective, the task of the expert system programmer is to reconstruct the conceptual and operational framework that an expert in a domain uses to solve problems in that domain, noting that the terminology used may be highly idiosyncratic, that is, personal to the expert. However, the expert is, by definition, someone who is effective at problem-solving in the domain and hence, her or his knowledge is valid in some practical sense. The expert’s knowledge has been acquired by some mix of processes, such as trial and error, mimicking others and so on. That corresponds to Kelly’s notion of an individual as a personal scientist (Shaw, 1980) about which he asks:

Might not the individual man, each in his own personal way, assume more of the stature of a scientist, ever seeking to predict and control the course of events with which he is involve? Would he not have his theories, test his hypotheses and weigh his experimental evidence? (Kelly, 1955).

Kelly merges the notions of prediction and control into the unitary notion of anticipation and hence his fundamental postulate:

A person’s processes are psychologically channelized by the ways in which he anticipates events.

Thus, from a personal construct perspective, the task of the expert system programmer is to model the personal construct system of the expert in operational form as a computer program such that the program is able to anticipate events in the same way as the expert. It was suggested in the early years of expert systems (Gaines and Shaw, 1980) that new methods for rule extraction made Kelly’s repertory grid a suitable tool for repertory grids were rapidly modified to support knowledge acquisition for expert systems (Shaw and Gaines, 1983; Boose, 1984). The approach proved successful in industrial applications (Boose, 1986) and a framework based on personal construct psychology became accepted as the foundation for developing knowledge acquisition techniques and tools (Ford et al., 1993; Gaines and Shaw, 1993).

A personal construct alternative to rule-based cognitive models: While repertory grids were widely used as knowledge acquisition tools in the 1980’s and 1990’s expert systems themselves failed to achieve as much as had been expected and a large-scale artificial intelligence industry did not materialize. Various writers have speculated on the reasons for that failure, the deepest analysis being that of Dreyfus and Dreyful (1986). They see the problem as a manifestation of Wittgenstein ‘s (1953) argument that the notion of human behavior following a rule is paradoxical because, as he showed, by a suitable interpretation every course of action could be make to accord with the rule.

The pioneers of cognitive science had modeled the human mind as a repository of so-called production rules (Anderson, 1983) and the designers of expert systems had followed this model in their knowledge representation schemes. Dreyfus and Dreyfus (1986) argue that such representation is a major weakness and that systems based on it could never fully emulate human expert behaviour. In the artificial intelligence literature.

Repretory grid-based knowledge acquisition tools had of necessity delivered knowledge in the form of rules so that it could be utilized by existing expert system knowledge representation tools. However, the analysis leading to the rules is not part of the construction process and may be regarded as an artifact of the need to use rule-based expert system technology. Kelly (1955) developed personal construct psychology from a perspective that was necessary to account for human behaviour and anticipation was a by-product of construction. That is, construction intrinsically supported anticipatory processes without the storage of anticipatory rules but, at a particular stage in the construction of experience, these anticipations might have a regularity that an observer could ascribe to rules of behaviour. In research on the philosophy of mind the term supervenient on (Kim, 1993) is used to describe a phenomenon which is a by-product of another phenomenon but not essential to it; anticipations supervenient anticipations might change and the observer could construe this in terms of the person learning new rules. Kelly (1955) also emphasized that personal construct psychology does not need a notion of ‘learning’ on the part of the personal scientist. Construction alone is sufficient to account for the person’s mental processes and behaviour and it could also account for the models being produced by observers or psychologists. The Wittgenstein paradox presents no problems to personal construct psychologists because there is no assumption that human behaviour is rule-governed.

It is unfortunate that the development of cognitive science in the mid-1950’s became dominated by those whose background was in mathematical logic. Kelly published his major work on personal construct psychology in 1955 and it could easily have become adopted as the foundation for what became called cognitive science’ and provided foundations for artificial intelligence and expert systems. In the few years until his death in 1967 he made a number of presentations to wider audiences that might have triggered recognition of the far-ranging implications of his work. In April 1961 he presented personal construct psychology to Luria and other members of the Moscow Psychological Society in Moscow as a mathematical approach to psychology (Kelly, 1969) paralleling the development of mathematical psychology in the USA by, Miller, Mosteller and others (Hirst, 1988). In June 1962 he was an invited commentator at a conference on the computer simulation of personality held at Princeton University and stated:

In this connection I would like to make a plug for the psychology of personal constructs. Not only is it a system built upon the notion that scientists and human beings, alike, approach truth by erecting simulation devices-called constructs-but is a theory deliberately formulated in a language system which is based on binary elements and which does not accept the so-called subject-predicate error of the Indo-European language system (Kelly, 1963).

However, Kelly’s work was not recognized in the 1950’s by computer and cognitive scientists.

Personal construct psychology as a foundation for modeling human expertise: The models of human thought processes derived from personal construct psychology and from mathematical logic can be contrasted through a simple example. Suppose a student has three constellations of experience:

Student is well-behaved; Teacher is attentive; Teacher smiles;
Student is naughty; Teacher is attentive; Teacher frowns;
Student is passive; Teacher is inattentive.

A machine learning program might derive the rules:

Student is well-behaved implies Teacher smiles;
Student is naughty implies Teachers frowns;
Student is passive implies Teacher is inattentive;
Student is well-behaved or naughty implies Teacher is attentive.

So that a student who is well-behaved might infer that her Teacher will smile, but how does the Student know when she is well-behaved?

A PCP model would be that the Student construes her three sets of experience in terms of the constructs: well-behave-naughty, attentive-inattentive and smiles-frowns. Supervenient on the construing of the three constellations of experience are all the compatible anticipations, that is, those plus:

Teacher smiles implies Student is well-behaved
Teacher frowns implies Student is naughty
Plus others.

These reverse implications will be used to give meaning to the construct well-behave-naughty in novel situations. To act to make the teacher smile the student will chose situations where the student is well-behaved and the teacher smiles. If the Student wants the teacher’s attention then the student may chose situations where the Student is naughty, the teacher frowns but also pays attention to the Student. There are no rules of behaviour but there is the choice of situations in a rather more flexible way than would be entailed through sets of rules. In addition there is an increasing repertoire of behaviour as the Student construes new situations in terms of her behaviour and the teacher’s smile or frown. One might say the student is learning but there is no reinforcement, only construction and choice. Kelly’s (1955) view is that construction provides a complete account of human behaviour and can also model the constructs of different schools of psychology.

Now apply that model to human expertise. It models the expert as a construing agent not as a knowledge base; of rules. The model automatically updates as more experience is construed, that is, as the expert system attempts to solve more problems. The experience can be used in a variety of ways to solve problems and to give meaning to new situations, for example, the availability of a new drug or treatment. Knowledge acquisition is intrinsic to personal construct-based expert systems and does not need to be treated as a separate phase. Expert knowledge can be transferred to the system not only through exemplary problem solving but also by commenting on the system’s problem solving and by choosing problems for the system which are at the limits of the system’s current capabilities. That is, experts can make their behaviour available to be mimicked and can also act as coaches commenting on performance and setting tasks, all major strategies in supporting human development.

An example repertory grid-based expert system development and application tool is WebGrid which is freely available on the World Wide Web (http://repgrid. com/WebGrid/). To use WebGrid and expert enters exemplary situations and, once some have been entered, can enter test cases to see how the system performs (Gaines and Shaw, 1997). If the system is incorrect, the expert can change the result and enter the corrected test case as an additional example until the system is generally correct. The system retains only the repertory grid of constructions as its knowledge base. WebGrid can produce sets of rules at any stage characterize and explain its model of expertise at that stage, but these are not stored just produced on request and are truly supervenient on the expert’s construction.

CONCLUSIONS

Expert systems were recognized as a breakthrough in artificial intelligence, in programming computers to emulate human thinking. However, they were based on a form of cognitive science that took mathematical logic as its foundations and was not well-suited to modeling the full richness of human behaviour. Personal construct psychology developed over the same time period but was not recognized by those working on artificial intelligence and cognitive science as a complete psychological system providing more effective foundations for cognitive science and expert systems. Repertory grid elicitation was recognized as a valuable knowledge acquisition technique with which to develop rules for expert systems, but the knowledge transferred in the form of rules was static and brittle and did not lead to the systems being open to experience. It would be timely to adopt personal construct psychology as the foundations of cognitive science and use it to build expert systems that fully emulated the capabilities of human experts, not only to solve problems but also to be effective in dealing with new problems as they arise.

As a final comment, it is noteworthy that while the expert system community has focused on emulating the capabilities of those with expertise of value to industry, the technology developed is useful for modeling the psychological processes of any person. Kelly noted that all people may be construed as scientists in their processes of modeling their worlds and validating those models. Similarly, we construe everyone as experts in being themselves and living their lives in their own way, whether or not the capabilities involved in doing this are singled out as being of special value by others. In therapy or in education for example, emulation of the person in the computer may provide a cognitive mirror (Shaw, 1980) in which an individual can view their psychological processes and come to understand them better. If there are problems arising from these processes, the increased understanding may help the individual to develop alternative constructions to address them. One by-product of research on the application of personal construct psychology to expert system development is that it has advanced our capabilities to model and emulate a person’s psychological processes in a way that may be useful to that person. The motivation for the research may have been to emulated the expertise of value to industry, but the outcomes have far wider significance.

REFERENCES

1:  Alford, J., C. Cairney, R. Higgs, W. Honsowetz and V. Huynh et al., 2000. Real rewards from artificial intelligence. http://www.highbeam.com/doc/1P3-41346249.html.

2:  Anderson, J.R., 1983. The Architecture of Cognition. Harvard University Press, Cambridge, MA

3:  Boose, J.H., 1984. Personal construct theory and the transfer of human expertise. Proceedings AAAI-84. California: American Association for Artificial Intelligence, pp: 27-33.

4:  Boose, J.H., 1986. Viewing knowledge bases as qualitative models. IEEE Expert, 4: 9-23.

5:  Dreyfus, H.L. and S.E. Dreyfus, 1986. Mind over Machine: The Power of Human Intuition and Expertise in the Era of the Computer. Free Press, New York

6:  Feigenbaum, E.A., B.G. Buchanan and J. Lederberg, 1971. On Generality and Problem Solving: A Case Study Using the DENDRAL Program. In: Machine Intelligence, Meltzer, B. and D. Michie (Eds.). Edinburgh University Press, Edinburgh

7:  Feigenbaum, E.A., B.G. Buchanan and J. Lederberg, 1971. On Generality and Problem Solving: A Case Study Using the DENDRAL Program. In: Machine Intelligence, Meltzer, B. and D. Michie (Eds.). Edinburgh University Press, Edinburgh

8:  Ford, K.M., J.M. Bradshaw, J.R. Adams-Webber and N.M. Agnew, 1993. Knowledge acquisition as a constructive modeling activity. Int. J. Intell. Syst., 8: 9-32.
Direct Link  |  

9:  Gaines, B.R. and M.L.G. Shaw, 1980. Now directions in the analysis and interactive elicitation of personal construct systems. Intl. J. Man-Machine Studies, 13: 81-116.

10:  Gaines, B.R. and M.L.G. Shaw, 1993. Basing knowledge acquisition tools in personal construct psychology. Knowledge Eng. Rev., 8: 49-85.

11:  Gaines, B.R. and M.L.G. Shaw, 1997. Knowledge acquisition, modeling and inference through the world wide web. Int. J. Hum. Comp. Studies, 46: 729-759.
Direct Link  |  

12:  Hayes-Roth, F., 1984. The International of Knowledge Engineering. In: Artificial Intelligence Applications for Business Norwood, Reitman, W. (Ed.). Ablex, New Jersey, pp: 159-177

13:  Hayes-Roth, F., D.A. Waterman and D.B. Lenat, 1983. Building Expert Systems Reading. Addison-Wesley, Massachusetts

14:  Hirst, W., 1988. The Making of Cognitive Science. Cambridge University Press, Cambridge, UK

15:  Kelly, G.A., 1955. The Psychology of Personal Constructs. Norton, New York

16:  Kelly, G.A., 1963. Aldous: The Personable Computer. In: Computer Simulation of Personality, Tomkins, S.S. and S. Messick (Eds.). Wiley, New York, pp: 221-229

17:  Kelly, G.A., 1969. A Mathematical Approach to Psychology. In: Clinical Psychology and Personality: The Selected Papers of George Kelly, Maher, B. (Ed.). Wiley, New York, pp: 91-113

18:  Kim, J., 1993. Supervenience and Mind: Selected Philosophical Essays. Cambridege University Press, Cambridge

19:  Lighthill, J., 1973. Artificial intelligence: A general survey. Artificial Intelligence: A Paper Symposium, Science Research Council, UK.

20:  McCarthy, J., M.L. Minsky, N. Rochester and C.E. Shannon, 1955. A proposal for the dartmouth summer research project on artificial intelligence. http://www-formal.stanford.edu/jmc/history/dartmouth/dartmouth.html.

21:  Michie, D., 1979. Expert Systems in the Micro Electronic Age. Ediinburgh University Press, Edinburgh

22:  Shaw, M.L.G., 1980. On Becoming a Personal Scientist: Interactive Computer Eliciatation of Personal Models of the World. Academic Press, London

23:  Shaw, M.L.G. and B.R. Gaines, 1983. A computer aid to knowledge engineering. Proceedings of the British Computer Society Conference on Expert Systems, December 1983, Cambridge, pp: 263-271

24:  Shortliffe, E.H., 1976. Computer-Based Medical Consultations: MYCIN. Elsevier, New York

25:  Turing, A.M., 1950. Computing machinery and intelligence. Mind, 59: 433-460.
Direct Link  |  

26:  Wittgenstein, L., 1953. Philosophical Investigations. Blackwell, Oxford

©  2021 Science Alert. All Rights Reserved