Abstract
Both logic and philosophy of science have developed independent pluralistic views. Nevertheless, a general notion of pluralism is quite elusive. This paper will draw on the theory of information, including Shannon’s version and its fundamental concept of coding, to set up a notion of pluralism that may contribute to an updated idea of the scientific method, such as to interact with computational models. I will discuss Patrick Allo’s notion of “informational pluralism” and connect it with Wesley Salmon’s treatment of “statistical relevance” and James Woodward’s analysis of the “data-phenomena” relationship. Based on the concept of information, I will argue that there is a possible convergence of the three views.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Beall and Restall’s formal account of logical pluralism provides the best opportunities for an information-theoretic approach, directed to support scientific methodology. Consequently, this paper will focus on semantic and model-theoretic issues and will be less concerned with modal topics such as necessity, possibility, and consistency. In this line of thought, modal logic will be reduced to a few mentions, mostly concerning the accessibility relation. For an alternative view and a more intense connection between pluralism and modal logic, see Bueno and Shalkowski (2009).
- 2.
If objective information consists of potential data, they plausibly become actual when encoded in the formulae of a language. See Sect. 6.4, below.
- 3.
Russell distinguishes (1) a substitutional account of LC, where non-logical expressions are uniformly substituted throughout the argument by “syntactically appropriate alternative expressions”; (2) an interpretational account, where different meanings replace the meanings of certain non-logical expressions; (3) a representational account, where non-logical expressions are represented by various situations or states of the world; (4) a universal account, where the non-logical vocabulary is uniformly substituted by different assignments of values to non-logical variables. See Russell (2019), 337–339.
- 4.
The concept of informational content has been used by Dretske in his classical analysis. He distinguishes between meaning and information and defines information as “a commodity capable of yielding knowledge” (Dretske 1981, 44). Allo accepts this distinction (Allo 2007, 687), which involves the problem of defining “information.” I am not going to pursue such a thorny question in this paper.
- 5.
- 6.
Any theory of channel cannot but be connected to the information theory that Claude Shannon established in 1948.
- 7.
In the second part of this paper, I intend to show that informational pluralism may also impact the philosophy of science.
- 8.
See also Mares’ relation of “accessibility” in Mares (2010), 117, and 125.
- 9.
- 10.
Oron Shagrir has written that “Computation is information processing.”
- 11.
Data are recorded (and I think encoded!) observations. Phenomena are defined this way: “facts about phenomena are natural candidates for systematic scientific explanations in a way in which facts about data are not.” (Bogen and Woodward 1988, 326)
- 12.
Bogen and Woodward do not consider models and simulations.
- 13.
More than this, I maintain that the pluralistic features of DPR also lead to the autonomy of computational models. I may even suggest that models, insofar as consisting of interpretive patterns applied to data collections, may be equated to the structures that connect and organize what Woodward calls “phenomena”.
- 14.
For an epistemic synthesis of Woodward’s ideas about causality, see Gonzalez 2018.
- 15.
This is based on the well-known equivalence – asserted by Claude Shannon (1998, 40) – between a discrete information source and a stochastic process.
- 16.
Besides the equivalence between situations and classes, we may recall other important computational equivalences, such as that between propositions and types, which is the basis of the fundamental Curry-Howard Isomorphism.
- 17.
- 18.
I am slightly departing from Salmon’s account at this point.
- 19.
Salmon was aware that computability theory was needed for the crucial problem of the construction of objectively homogeneous reference classes (see Salmon 1984, 58–60, 67–68, 81). As for information theory, see Salmon 1984, in particular, 97–101, 125–126, 139–154. S-R approach has been recently updated in informational terms, especially via the notion of “causal power”. This represents the degree to which changes in a cause C produce changes in an effect E. Causal power uses relevance logic to cover not only conditional probabilities and statistical datasets but also Bayesian Networks used in Artificial Intelligence. Causal power is measured using Shannon’s entropy concept (Korb et al. 2009).
- 20.
The completeness of partitions provides to Salmon’s S-R models the character of objectivity that distinguishes them from Hempel’s I-S models (Salmon 1984, 41). Completeness is granted by the maximal homogeneity of both explanans and explanandum partitions (Salmon 1984, 37). It is also to be mentioned that Salmon’s objective homogeneity is the same thing as Shannon’s “ergodicity,” roughly defined as “statistical homogeneity” (Shannon 1998, 45), and reputed to be essential for the final Theorem 11 of Information theory (cf. Kinchin 1957).
- 21.
Salmon’s and Greeno’s visible source is Kullback’s classical treatise Information Theory and Statistics. Kullback supplements Shannon’s concepts with the works of Fisher on information (cf. Kullback 1968). However, apart from Salmon’s theory, it is crucial to shift from Fisher’s to Shannon’s view. In doing so, we gain an intensional perspective on statistical types that is the premise for being able to build computational models. Sayre also misses this point.
References
Abramson, N. (1963). Information theory and coding. New York: McGraw Hill.
Allo, P. (2007). Logical pluralism and semantic information. Journal of Philosophical Logic, 36, 659–694.
Allo, P., & Mares, E. (2012). Informational semantics as a third alternative? Erkenntnis, 77, 167–185.
Barwise, J. (1993). Constraints, channels, and the flow of information. In P. Aczel, D. Israel, S. Peters, & Y. Katagiri (Eds.), Situation theory and its applications (pp. 3–27). Stanford: CSLI.
Barwise, J., & Seligman, J. (1997). Information flow: The logic of distributed systems. Cambridge: Cambridge University Press.
Beall, J. C., & Restall, G. (2000). Logical pluralism. Australasian Journal of Philosophy, 78, 475–493.
Beall, J. C., & Restall, G. (2006). Logical pluralism. Oxford: Oxford University Press.
Berstel, J., & Perrin, D. (2010). Codes and automata. Cambridge: Cambridge University Press.
Bishop, R. (2012). Fluid convection, constraint and causation. Interface Focus, 2, 4–12.
Bogen, J., & Woodward, J. (1988). Saving the phenomena. The Philosophical Review, 97, 303–352.
Bogen, J., & Woodward, J. (2003). Evading the IRS. Poznan Studies in the Philosophy of Science and Humanities, 20, 223–256.
Bueno, O., & Shalkowski, S. (2009). Modalism and logical pluralism. Mind, 118, 295–321.
Camardi, G. (2012). Computational models and information theory. Journal of Experimental and Theoretical Artificial Intelligence, 24, 401–417.
Caret, C. (2019). Why logical pluralism? Synthese. https://doi.org/10.1007/s11229-019-02132-w. Accessed on 06 Sept 2019.
Cook, R. (2019). Pluralism about pluralisms. In J. Wyatt, N. J. L. L. Pedersen, & N. Kellen (Eds.), Pluralism in truth and logic (pp. 365–386). Cham: Palgrave Macmillan.
Cover, T., & Thomas, J. (2006). Elements of information theory. Hoboken: Wiley.
Dretske, F. (1981). Knowledge and the flux of information. Oxford: Blackwell.
Fitting, M. (2002). Fixpoint semantics for logic programming. Theoretical Computer Science, 278, 25–51.
Floridi, L. (2011). The philosophy of information. Oxford: Oxford University Press.
Godfrey-Smith, P. (2010). Causal pluralism. In H. Beebee, C. Hitchcock, & P. Menzies (Eds.), Oxford handbook of causation (pp. 326–337). Oxford: Oxford University Press.
Gonzalez, W. J. (2018). Configuration of causality and philosophy of psychology: An analysis of causality as intervention and its repercussion for psychology. In W. J. Gonzalez (Ed.), Philosophy of psychology: Causality and psychological subject. New reflections on James Woodward’s contribution (pp. 21–70). Boston/Berlin: de Gruyter.
Hall, E. (2004). Two concepts of causation. In L. Paul, E. Hall, & J. Collins (Eds.), Causation and counterfactuals (pp. 225–276). Cambridge, MA: MIT Press.
Hamming, R. W. (1986). Coding and information theory. Englewood Cliffs: Prentice Hall.
Hankerson, D., Harris, G., & Johnson, P. (2003). Introduction to information theory and data compression. Boca Raton: Chapman and Hall/CRC.
Illari, P. (2011). Why theories of causality need production: An information transmission account. Philosophy and Technology, 24, 95–114.
Kinchin. (1957). Mathematical foundations of information theory. New York: Dover Publications.
Korb, K., Hope, L., & Nyberg, E. (2009). Information-theoretic causal power. In F. Emmert-Streiss & M. Dehmer (Eds.), Information theory and statistical learning. Berlin/New York: Springer.
Kullback, S. (1968). Information theory and statistics. Mineola: Dover Publications.
Leonelli, S. (2008). Performing abstraction. Two ways of modeling Arabidopsis Thaliana. Biology and Philosophy, 23, 509–528.
Levy, A., & Bechtel, W. (2013). Abstraction and the organization of mechanisms. Philosophy of Science, 80, 241–261.
Mares, E. (1997). Relevant logics and the theory of information. Synthese, 109, 345–360.
Mares, E. (2006). Relevant logics, probabilistic information and conditionals. Logique et Analyse, 196, 399–411.
Mares, E. (2010). The nature of information: A relevant approach. Synthese, 175, 111–132.
Mares, E., & Meyer, R. K. (2001). Relevant logics. In L. Goble (Ed.), The Blackwell guide to philosophical logic (pp. 280–308). Oxford: Blackwell.
Morrison, M. (2015). Reconstructing reality: Models, mathematics, and simulations. Oxford: Oxford University Press.
O’ Donnell, M. (1998). Introduction: Logic and logic programming languages. In D. Gabbay, C. Hogger, & J. Robinson (Eds.), Handbook of logic in artificial intelligence and logic programming (pp. 1–67). Oxford: Oxford University Press.
Padmanabhan, K., et al. (2013). In situ exploratory data analysis for scientific discovery. In T. Critchlow & K. K. van Dam (Eds.), Data intensive science (pp. 301–350). Boca Raton: Taylor and Francis.
Piccinini, G. (2015). Physical computation: A mechanistic account. Oxford: Oxford University Press.
Restall, G. (1995). Information flow and relevant logic. In J. Seligman & D. Westerstähl (Eds.), Logic, language and computation: The 1994 Moraga proceedings (pp. 463–477). Stanford: CSLI Press.
Russell, G. (2019). Varieties of logical consequence by their resistance to logical nihilism. In J. Wyatt, N. J. L. L. Pedersen, & N. Kellen (Eds.), Pluralism in truth and logic (pp. 331–361). Cham: Palgrave Macmillan.
Salmon, W. (1967). Foundations of scientific inference. Pittsburgh: University of Pittsburgh Press.
Salmon, W. (1970). Bayes’s theorem and the history of science. In R. Stuewer (Ed.), Historical and philosophical perspectives of science (pp. 68–86). Minneapolis: University of Minnesota Press.
Salmon, W. (1977). An “at-at” theory of causal influence. Philosophy of Science, 44, 215–224.
Salmon, W. (1984). Scientific explanation and the causal structure of the world. Princeton: Princeton University Press.
Salmon, W. (1994). Causality without counterfactual. Philosophy of Science, 61, 297–312.
Sayre, K. (1977). Statistical models of causal relations. Philosophy of Science, 44, 203–214.
Shannon, C. (1998). The mathematical theory of communication [1948]. Repr. In C. Shannon & W. Weaver (eds) The mathematical theory of communication (pp. 29–125). Urbana: University of Illinois Press.
Shapere, D. (1982). The concept of observation in science and in philosophy. Philosophy of Science, 49, 485–525.
Strevens, M. (2008). Depth. An account of scientific explanation. Cambridge, MA: Harvard University Press.
Turner, R. (2009). Computable models. London: Springer.
van Fraassen, B. (2008). Scientific representation. Oxford: Oxford University Press.
Van Emden, M., & Kowalski, R. (1976). The semantics of predicate logic programming language. Journal of the ACM, 23, 733–742.
Von Mises, R. (1981). Probability, statistics and truth. Mineola, New York: Dover Publications.
Woodward, J. (2003). Making things happen. In A theory of causal explanation. Oxford: Oxford University Press.
Woodward, J. (2011). Data and phenomena. A restatement and defense. Synthese, 182, 165–179.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Camardi, G. (2020). Information and Pluralism. Consequences for Scientific Representation and Methods. In: Gonzalez, W.J. (eds) Methodological Prospects for Scientific Research. Synthese Library, vol 430. Springer, Cham. https://doi.org/10.1007/978-3-030-52500-2_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-52500-2_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-52499-9
Online ISBN: 978-3-030-52500-2
eBook Packages: Religion and PhilosophyPhilosophy and Religion (R0)