From its humble origins labeling a technical annoyance for a particular AI formalism, the term frame problem has grown to cover issues confronting broader research programs in AI. In philosophy, the term has come to encompass allegedly fundamental, but merely superficially related, objections to computational models of mind in AI and beyond.
The original frame problem appears within the SITUATION CALCULUS for representing a changing world. In such systems there are "axioms" about changes conditional on prior occurrences -- that pressing a switch changes the illumination of a lamp, that selling the lamp changes who owns it, and so on. Unfortunately, because inferences are to be made solely by deduction, axioms are needed for purported nonchanges -- that pressing the switch does not change the owner, that selling the lamp does not change its illumination, and so on. Without such "frame axioms," a system is unable strictly to deduce that any states persist. The resulting problem is to do without huge numbers of frame axioms potentially relating each representable occurrence to each representable nonchange.
A common response is to handle nonchanges implicitly by allowing the system to assume by default that a state persists, unless there is an axiom specifying that it is changed by an occurrence, given surrounding conditions. Because such assumptions are not deducible from the axioms of change (even given surrounding conditions), and because the licensed conclusions are not cumulative as evidence is added, the frame problem helps motivate the development of special NONMONOTONIC LOGICS intended to minimize the assumptions that must be retracted given further evidence. This is related to discussions of defeasibility and ceteris paribus reasoning in epistemology and philosophy of science (e.g., Harman 1986).
A related challenge is to determine which assumptions to retract when necessary, as in the "Yale Shooting Problem" (Hanks and McDermott 1986). Let a system assume by default (1) that live creatures remain alive, and (2) that loaded guns remain loaded. Confront it with this information: Fred is alive, then a gun is loaded, then, after a delay, the gun is fired at Fred. If assumption (2) is in force through the delay, Fred probably violates (1). But equally, if assumption (1) is in force after the shooting, the gun probably violates (2). Why is (2) the more natural assumption to enforce? Some favor (2) because the delay occurs before the shooting (e.g., Shoham 1988). Others favor (2) because there is no represented reason to believe it violated, although the shooting provides some reason for believing (1) violated (e.g., Morgenstern 1996; cf. philosophical discussions of inference to the best EXPLANATION, e.g., Thagard 1988). Work continues in this vein, seeking to formalize the relevant temporal and rational notions, and to insure that the strategies apply more broadly than the situation calculus.
Another approach to the frame problem seeks to remain within the strictures of classical (monotonic) logic (Reiter 1991). In most circumstances, it avoids the use of huge numbers of axioms about nonchanges, but at the cost of using hugely and implausibly bold axioms about nonchanges. For example, it is assumed that all the possible causes of a certain kind of effect are known, or that all the actual events or actions operating on a given situation are known.
Some philosophers of mind maintain that the original frame problem portends deeper problems for traditional AI, or at least for cognitive science more broadly. (Unless otherwise mentioned, the relevant papers of the authors cited may be found in Pylyshyn 1987.) Daniel Dennett wonders how to ignore information obviously irrelevant to one's goals, as one ignores many obvious nonchanges. John Haugeland wonders how to keep track of salient side effects without constantly checking for them. This includes the "ramification" and "qualification" problems of AI; see Morgenstern (1996) for a survey. Jerry Fodor wonders how to avoid the use of "kooky" concepts that render intuitive nonchanges as changes -- for instance, "fridgeon" which applies to physical particles if and only if Fodor's fridge is on, so that Fodor can "change" the entire universe simply by unplugging his fridge. AI researchers, including Drew McDermott and Pat Hayes, protest that these further issues are unconnected to the original frame problem.
Nevertheless, the philosophers' challenges must be met somehow if human cognition is to be understood in computational terms (see CAUSAL REASONING). Exotic suggestions involve mental IMAGERY as opposed to a LANGUAGE OF THOUGHT (Haugeland, cf. Janlert in AI), nonrepresentational practical skills (Dreyfus and Dreyfus), and emotion-induced temporary modularity (de Sousa 1987, chap. 7). The authors of the Yale Shooting Problem argue, as well, against the hegemony of logical deduction -- whether classical or nonmonotonic -- in AI simulations of commonsense reasoning. More conservative proposed solutions appeal to HEURISTIC SEARCH techniques and ideas about MEMORY long familiar in AI and cognitive psychology (Lormand, in Ford and Pylyshyn 1996; Morgenstern 1996 provides an especially keen survey of AI proposals).
de Sousa, R. (1987). The Rationality of Emotion. Cambridge, MA: MIT Press.
Ford, K., and Z. Pylyshyn, Eds. (1996). The Robot's Dilemma Revisited. Norwood, NJ: Ablex.
Hanks, S., and D. McDermott. (1986). Default reasoning, nonmonotonic logic, and the frame problem. Proceedings of the American Association for Artificial Intelligence 328-333.
Harman, G. (1986). Change in View. Cambridge, MA: MIT Press.
Morgenstern, L. (1996). The problem with solutions to the frame problem. In K. Ford and Z. Pylyshyn, Eds., (1996), pp. 99-133.
Pylyshyn, Z., Ed. (1987). The Robot's Dilemma. Norwood, NJ: Ablex.
Reiter, R. (1991). The frame problem in the situation calculus: a simple solution (sometimes) and a completeness result for goal regression. In V. Lifschitz, Ed., Artificial Intelligence and Mathematical Theory of Computation: Papers in Honor of John McCarthy. Boston: Academic Press, pp. 359-380.
Shoham, Y. (1988). Reasoning about Change. Cambridge, MA: MIT Press.
Thagard, P. (1988). Computational Philosophy of Science. Cambridge, MA: MIT Press .