Skip to main content

Krister Segerberg’s Philosophy of Action

  • Chapter
  • First Online:
Krister Segerberg on Logic of Actions

Part of the book series: Outstanding Contributions to Logic ((OCTR,volume 1))

Abstract

In his logic of action, Krister Segerberg has provided many insights about how to formalize actions. In this chapter I consider these insights critically, concluding that any formalization of action needs to be thoroughly connected to the relevant reasoning, and in particular to temporal reasoning and planning in realistic contexts. This consideration reveals that Segerberg’s ideas are limited in several fundamental ways. To a large extent, these limitations have been overcome by research that has been carried out for many years in Artificial Intelligence.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

eBook
USD 16.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Hardcover Book
USD 109.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Segerberg’s work on action is presented in articles dating from around 1980. See the articles by Segerberg cited in the bibliography of this chapter.

  2. 2.

    See [9].

  3. 3.

    Now we are working with Turing machines that run pseudocode, and whose states consist of assignments of values to an infinite set of variables. This assumption is legitimate, and loses no generality.

  4. 4.

    Corresponding to any state-tree, there is the set of its branches. Conversely, however, not every set of sequences can be pieced together into a tree.

  5. 5.

    Further examples along these lines, and a classification of the cases, can be found in [2]. I believe that an adequate theory of action must do justice to Austin’s distinctions.

  6. 6.

    See, for instance, [7].

  7. 7.

    Such routines might be helpful in designing a robot that could learn to throw darts, but issues like this are controversial in robotics itself. See [6] and other references on “situated robotics.”

  8. 8.

    See [13].

  9. 9.

    There are problems with such a termination rule in cases where the agent can exercise knowledge acquisition routines—routines that can expand the routines available to the agent. But this problem is secondary, and we need not worry about it.

  10. 10.

    In more complex cases, and to do justice to the way humans often plan, we might want to associate various levels of abstraction with a domain, and allow the primitive actions at higher levels of abstraction to be decomposed into complex lower-level actions. This idea has been explored in the AI literature; see, for instance, [15, 37]. One way to look at what Segerberg seems to be doing is this: he is confining means-end reasoning to realization and ignoring causality. He discusses cases in which the reasoning moves from more abstract goals to more concrete goals that realize them. But he ignores cases where, at the same level of abstraction, the reasoning moves from a temporal goal to an action that will bring the goal about if performed.

  11. 11.

    In this section, we use italics for predicates and SmallCaps for actions.

  12. 12.

    We can, of course, think of states as propositions of a special, very informative sort.

  13. 13.

    An action is trivial in \(s\) if it leaves \(s\) unchanged.

  14. 14.

    Reasoning in this direction is cumbersome; it is easier to find opportunities to act in a state \(s\) than to find ways in which \(s\) might have come about. Evidently, Segerberg’s method is not the most natural way to approach this planning problem.

  15. 15.

    See, for instance, [10, 14, 34, 39]. In this chapter I don’t follow any of the usual formalisms precisely, but have invented one that, I hope, will seem more familiar to readers who know some modal logic.

  16. 16.

    I hope the notation is clear. \([a]\) is a modal operator indicating what holds after performing action denoted by \(a\).

  17. 17.

    For more about methods for testing knowledge-based programs, see [12, 36]. For a discussion of the relation of logic to agent architectures, see ([40], Chap. 9).

  18. 18.

    These include [18, 2024, 30, 31].

  19. 19.

    For more about intention-belief inconsistency, see ([5], pp. 37–38).

  20. 20.

    If this example fails to convince you, consider the following one. I’m a terrible typist. When I began to prepare this chapter, I believed I would make many typographical errors in writing it. But when I intended to write the chapter, I didn’t intend to make typographical errors.

  21. 21.

    The theory presented in [31] is by no means the only logic of intention to suffer from this problem. See, for instance, ([40], Chap. 4).

  22. 22.

    For mixed models of this kind, see, for instance, [16].

References

  1. Anscombe, G. (1958). Intention. Oxford: Blackwell Publishers.

    Google Scholar 

  2. Austin, J. L. (1956–57). A plea for excuses. Proceedings of the Aristotelian Society, 57, 1–30.

    Google Scholar 

  3. Brachman, R. J., & Levesque, H. (2004). Knowledge representation and reasoning. Amsterdam: Elsevier.

    Google Scholar 

  4. Bratman, M. E. (1987). Intentions, plans and practical reason. Cambridge: Harvard University Press.

    Google Scholar 

  5. Bratman, M. E. (1990). What is intention? In P. R. Cohen, J. Morgan, & M. Pollack (Eds.), Intentions in communication (pp. 15–32). Cambridge, MA: MIT Press.

    Google Scholar 

  6. Brooks, R. A. (1990). Elephants don’t play chess. Robotics and Autonomous Systems, 6(1–2), 3–15.

    Article  Google Scholar 

  7. Clarke, E. M., Grumberg, O., & Peled, D. A. (1999). Model checking. Cambridge, MA: MIT Press.

    Google Scholar 

  8. Fritz, C., & McIlraith, S. A. (2008). Planning in the face of frequent exogenous events. In Online Poster Proceedings of the 18th International Conference on Automated Planning and Scheduling (ICAPS), Sydney, Australia. http://www.cs.toronto.edu/kr/publications/fri-mci-icaps08.pdf.

  9. Harel, D., Kozen, D., & Tiuryn, J. (2000). Dynamic logic. Cambridge, MA: MIT Press.

    Google Scholar 

  10. Lifschitz, V. (1987). Formal theories of action. In M. L. Ginsberg (Ed.), Readings in nonmonotonic reasoning (pp. 410–432). Los Altos, CA: Morgan Kaufmann.

    Google Scholar 

  11. Lorini, E., & Herzig, A. (2008). A logic of intention and attempt. Synthése, 163(1), 45–77.

    Article  Google Scholar 

  12. McGuinness, D. L., Fikes, R., Rice, J., & Wilder, S. (2000). An environment for merging and testing large ontologies. In A. G. Cohn, F. Giunchiglia, & B. Selman (Eds.), KR2000: Principles of knowledge representation and reasoning (pp. 483–493). San Francisco: Morgan Kaufmann.

    Google Scholar 

  13. Newell, A. (1992). Unified theories of cognition. Cambridge, MA: Harvard University Press.

    Google Scholar 

  14. Reiter, R. (2001). Knowledge in action: Logical foundations for specifying and implementing dynamical systems. Cambridge, MA: MIT Press.

    Google Scholar 

  15. Sacerdoti, E. D. (1974). Planning in a hierarchy of abstraction spaces. Artificial Intelligence, 5(2), 115–135.

    Article  Google Scholar 

  16. Sandewall, E. (1989). Combining logic and differential equations for describing real-world systems. In R. J. Brachman, H. J. Levesque, & R. Reiter (Eds.), KR’89: Principles of knowledge representation and reasoning (pp. 412–420). San Mateo, CA: Morgan Kaufmann.

    Google Scholar 

  17. Segerberg, K. (1980). Applying modal logic. Studia Logica, 39(2–3), 275–295.

    Article  Google Scholar 

  18. Segerberg, K. (1981). Action-games. Acta Philosophica Fennica, 32, 220–231.

    Google Scholar 

  19. Segerberg, K. (1982a). Getting started: Beginnings in the logic of action. Studia Logica, 51(3–4), 437–478.

    Google Scholar 

  20. Segerberg, K. (1982b). The logic of deliberate action. Journal of Philosophical Logic, 11(2), 233–254.

    Article  Google Scholar 

  21. Segerberg, K. (1984). Towards an exact philosophy of action. Topoi, 3(1), 75–83.

    Article  Google Scholar 

  22. Segerberg, K. (1985a). Models for action. In B. K. Matilal & J. L. Shaw (Eds.), Analytical philosophy in contemporary perspective (pp. 161–171). Dordrecht: D. Reidel Publishing Co.

    Chapter  Google Scholar 

  23. Segerberg, K. (1985b). Routines. Synthése, 65(2), 185–210.

    Article  Google Scholar 

  24. Segerberg, K. (1988). Talking about actions. Studia Logica, 47(4), 347–352.

    Article  Google Scholar 

  25. Segerberg, K. (1989). Bringing it about. Journal of Philosophical Logic, 18(4), 327–347.

    Article  Google Scholar 

  26. Segerberg, K. (1992). Representing facts. In C. Bicchieri & M. L. D. Chiara (Eds.), Knowledge, belief, and strategic interaction (pp. 239–256). Cambridge, UK: Cambridge University Press.

    Chapter  Google Scholar 

  27. Segerberg, K. (1994). A festival of facts. Logic and Logical Philosophy, 2, 7–22.

    Google Scholar 

  28. Segerberg, K. (1995). Conditional action. In G. Crocco, L. F. nas del Cerro, & A. Herzig (Eds.), Conditionals: From philosophy to computer science (pp. 241–265). Oxford: Oxford University Press.

    Google Scholar 

  29. Segerberg, K. (1996). To do and not to do. In J. Copeland (Ed.), Logic and reality: Essays on the legacy of Arthur Prior (pp. 301–313). Oxford: Oxford University Press.

    Google Scholar 

  30. Segerberg, K. (1999). Results, consequences, intentions. In G. Meggle (Ed.), Actions, norms, values: Discussion with Georg Henrik von Wright (pp. 147–157). Berlin: Walter de Gruyter.

    Google Scholar 

  31. Segerberg, K. (2005). Intension, intention. In R. Kahle (Ed.), Intensionality (pp. 174–186). Wellesley, MA: A.K. Peters, Ltd.

    Google Scholar 

  32. Segerberg, K., Meyer, J. J., & Kracht, M. (2009). The logic of action. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy, summer (2009th ed.). Stanford: Stanford University.

    Google Scholar 

  33. Setiya, K. (2011). Intention. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy, spring (2011th ed.). Stanford: Stanford University.

    Google Scholar 

  34. Shanahan, M. (1997). Solving the frame problem. Cambridge, MA: MIT Press.

    Google Scholar 

  35. Simon, H. A., & Schaeffer, J. (1992). The game of chess. In R. J. Aumann & S. Hart (Eds.), Handbook of game theory with economic applications (Vol. 1, pp. 1–17). Amsterdam: North-Holland.

    Chapter  Google Scholar 

  36. Stefik, M. J. (1995). An introduction to knowledge systems. San Francisco: Morgan Kaufmann.

    Google Scholar 

  37. Sutton, R. S., Precup, D., & Singh, S. (1999). Between MDPs and semi-MDPs: A framework for temporal abstraction in reinforcement learning. Artificial Intelligence, 112(1–2), 181–211.

    Article  Google Scholar 

  38. Turing, A. M. (1950). Computing machinery and intelligence. Mind, 59(236), 433–460.

    Article  Google Scholar 

  39. Turner, H. (1999). A logic of universal causation. Artificial Intelligence, 113(1–2), 87–123.

    Article  Google Scholar 

  40. Wooldridge, M. J. (2000). Reasoning about Rational Agents. Cambridge, UK: Cambridge University Press.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Richmond H. Thomason .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer Science+Business Media Dordrecht

About this chapter

Cite this chapter

Thomason, R.H. (2014). Krister Segerberg’s Philosophy of Action. In: Trypuz, R. (eds) Krister Segerberg on Logic of Actions. Outstanding Contributions to Logic, vol 1. Springer, Dordrecht. https://doi.org/10.1007/978-94-007-7046-1_1

Download citation

Publish with us

Policies and ethics