BRANDOM AND POSTHUMAN AGENCY: AN ANTI-NORMATIVIST RESPONSE TO BOUNDED POSTHUMANISM
David Roden, Open University
Introduction: Bounded Posthumanism
Posthumanism can be critical or speculative in orientation. Both kinds are critical of human-centered (anthropocentric) thinking. However, their rejection of anthropocentricism applies to different areas: Critical Posthumanism (CP) rejects the anthropocentrism of modern philosophy and intellectual life; Speculative Posthumanism (SP) opposes human-centric thinking about the long-run implications of modern technology.
Whereas critical posthumanists are interested in the posthuman as a cultural and political condition, speculative posthumanists are interested in a possibility of certain technologically created nonhuman agents. They claim that there could be posthumans – where posthumans would be “wide human descendants” of current humans that have become nonhuman in virtue of some process of technical alteration (Roden 2012; 2014, Chapter 5).
Despite differences in concern and methodology, however, CP and SP have convergent interests. CP requires that there are no transcendental conditions for agenthood derivable from parochial facts about human agency. If this is right, it must true of possible nonhuman agents as it is of actual nonhuman agents.
For this reason, I distinguish two claims regarding technological successors to current humans: an anthropologically bounded posthumanism (ABP); and an anthropologically unbounded posthumanism (AUP).
- There are unique constraints C on cognition and agency which any being qualifying as a posthuman successor to humans must satisfy.
- Agents satisfying C can know that they are agents and can deduce a priori that they satisfy C (they are transcendental constraints)
- Humans typically satisfy C.
ABP’s import becomes clearer if we consider the collection of histories whereby posthuman wide descendants of humans could feasibly emerge. I refer to this set as Posthuman Possibility Space (PPS – See Roden 2014: 53).
Given that posthumans would be agents of some kind (See Chapter 6) and given ABP, members of PPS would have to satisfy the same transcendental conditions (C) on agency as humans.
Daryl Wennemann assumes something along these lines in his book Posthuman Personhood. He adopts the Kantian idea that agency consists in the capacity to justify one’s actions according to reasons and shared norms. For Wennemann, a person is a being able to “reflect on himself and his world from the perspective of a being sharing in a certain community.” (Punzo 1969, cited in Wennemann 2013: 47). This is a condition of posthuman agency as much as of human agency
This implies that, whatever the future throws up, posthuman agents will be social and, arguably linguistic beings, even if they are robots or computers, have strange bodies, or even stranger habits. If so, PPS cannot contain non-anthropomorphic entities whose agency is significantly nonhuman in nature.
ABP implies that there are a priori limits on posthuman weirdness.
AUP, by contrast, leaves the nature of posthuman agency to be settled empirically (or technologically). Posthumans might be social, discursive creatures; or they might be different from us in ways that we cannot envisage short of making some posthumans or becoming posthuman ourselves.
AUP thus extends the critical posthumanist rejection of anthropocentrism to the deep time of the technological future. In Posthuman Life I defended it via a critique of Donald Davidson’s work on intentionality; coupling this with a “naturalistic deconstruction” of transcendental phenomenology in its Husserlian and Heideggerian forms (See also Roden 2013).
Some of these arguments, I believe, carry over to the more overtly normativist philosophy of Robert Brandom – a philosopher whose work I did not address in detail there (for reasons of space and incompetence). The account of the relationship between normativity, social practice, intentionality that Brandom provides in Making It Explicit, and in other writings, is one of most impressively detailed, systematic and historically self-aware attempts to explain subjectivity, agency and intentionality in terms of social practices and statuses. It thus merits the appraisal of all philosophical posthumanists, whether they are of a critical or a speculative bent.
First and Second-Class Agents
I will begin with a thumbnail sketch of how Brandom derives a priori conditions of possibility for agency and meaning from a theory of social practices. Then I will consider whether its foundations are capable of supporting this transcendental superstructure.
Brandom is a philosophical pragmatist. Like other pragmatists, he is committed to the claim that our conceptual and intellectual powers are grounded in our practical abilities rather than in relations between mental entities and what they represent (Brandom 2006).
His pragmatism implies a species of interpretationism with regard to intentional content. Interpretationists, like Daniel Dennett, claim that intentional notions such as “belief” do not track inner vehicles of content but help us assess patterns of rational activity on the part of other “intentional systems” (Wanderer 2008). Belief-desire talk is not a folk psychological “theory” about internal states, but a social “craft” for evaluating and predicting other rational agents.
For Dennett, an entity qualifies as an agent with reasons if predicting its behaviour requires interpreters to attribute it the beliefs and desires it ought to have given its nature and environment. A being whose behaviour is voluminously predictable under this “intentional stance” is called an “intentional system” (IS). In IS theory, there is no gap between predictability under the intentional stance and having real intentionality.
Brandom endorses Dennett’s claim that intentional concepts are fundamentally about rendering agency intelligible in the light of reasons. However, he argues that IS theory furnishes an incomplete account or intentionality. Interpretation is an intentional act; thus interpretationists need to elucidate the relationship between attributed intentionality and attributing intentionality. If we do not understand what kind of being could count as a prospective interpreter, we cannot claim to have understood what it is to attribute intentionality in the first place (Brandom 1994: 59).
Brandom goes one further. The intentionality attributed to intrinsically meaningless events or linguistic inscriptions seems entirely derived from interpreters. Similarly with relatively simple IS’s. Maze-running robots or fly-catching frogs can properly be understood from the intentional stance – making them true-believers by Dennett’s lights. But their intentionality seems likewise observer-relative; derived from attitudes of interpreting IS’s (60). To hold otherwise, he argues, is to risk a disabling regress. For if intentionality is derivative all the way up, there can be no real intentional attributions and thus no derivative (non-observer relative) intentionality (60, 276).
Brandom claims that his theory can be read as an account of the conditions an organism must satisfy to qualify an interpreting intentional system; that is to warrant attributions of non-derived intentionality rather than the as-if intentionality we can attribute to simpler organisms or complex devices.
Whatever else the capacity for original or “first class” intentionality includes, it must involve the ability to evaluate the cognizance and rationality of similar beings and thus to be answerable to reasons (61). Entities with first-class intentionality and thus the capacity to assess and answer to reasons in this way are referred to by Brandom as sapient. Entities with only derived intentionality may exhibit the sentient capacity to react in discriminating and optimizing ways to their environment, but the conceptual content of these responses is attributed and observer-relative.
The claim that intentionality or the capacity for objective thought implies the capacity to evaluate other thinkers obviously has a rich post-Kantian lineage. However, one of the clearest arguments for connecting intentionality and the capacity for other-evaluation is provided by Donald Davidson in his essay “Thought and Talk” (Davidson 1984: 155-170).
Davidson begins with the assumption that belief is an attitude of “holding” true some proposition: for example, that there is a cat behind that wall. If belief is holding true it entails a grasp of truth and the possibility of being mistaken; and thus a concept of belief itself. Thus we cannot believe anything without the capacity to attribute true or false beliefs about the same topic to our fellow creatures (Davidson 1984: 170; 2001b: 104).
This capacity presupposes linguistic abilities, according to Davidson, because attributing contents to fellow creatures requires a common idiom of expression. Absent this, the possession of a concept of belief and, thus, the very having of beliefs, is impossible.
Brandom agrees! We need language to have and attribute beliefs, and, by extension, practical attitudes corresponding to desires and intentions (231-2). However, his official account avoids talk of beliefs or intentions in order to steer clear of the picture of beliefs, etc. as inner vehicles of content (sentences in the head, say) rather than social statuses available to discursive creatures like ourselves.
For Brandom, the primary bearers of propositional content are public assertions. Thus he bases his elaborate theory of intentionality not on a theory of mental representations or sub-propositional concepts, but on a pragmatic account of the place of assertions within the social game of “giving and asking for reasons”.
Correlatively, Brandom’s semantics begins with an explanation of how assertions – and their syntactical proxies, sentences – acquire propositional content. Like Wilfred Sellars’ brand of functional semantics, it is framed in terms of the normative role of utterances within social practices which determine how a speaker can move from one position in the language-game to another.
In the case of assertions, the language-transition rules correspond to materially correct inferences such as that x is colored from x is red. Language entry-rules include observation statements which allow us to make claims like “There is snow on the grass” on the basis of our reliable dispositions to differentially respond (RDRDs) to recurrent states of our environment. Finally, “language exit rules” correspond to practical commitments to forms of non-linguistic action.
Thus Brandom agrees with other post-Wittgensteinian pragmatists that linguistic practices are governed by public norms. However, he follows Davidson in rejecting the “I-we” conception of social structure. (39-40; Davidson 1986). If meanings are inferential roles (as Dummett and Sellars also claim), then the content attributable to expressions will dance in line with the doxastic commitments of individual speakers.
Suppose one observes a masked figure in a red costume clambering up a skyscraper. The language entry rules ambient within your community of English speakers may entitle you (by default) to claim that Spiderman is climbing the building. However, you are unaware that Spiderman is none other than Peter Parker. So you are not yet entitled to infer that Peter Parker is climbing the building – although the substitutional rules of English would commit you to that further inference if (say) some reliable authority informed you of this fact.
This simple example shows that the inferential roles – thus meanings – of expressions like “Spiderman” are not fixed communally but have to vary with the auxiliary assumptions, sensitivities and dispositions of individual speakers. Understanding or interpreting the utterances and beliefs of others is thus a matter of deontic scorekeeping – that is keeping track of the way social statuses alter as speakers update their inferential commitments (Brandom 1994: 142).
Thus semantic and intentional content are co-extensive with the normative-functional roles of states and actions. It follows that what a belief or claim “represents” or is “about” is fixed by its status it can be ascribed from the perspective of various deontic scorekeepers (including the believer or claimant).
The second consequence – which I flagged earlier – is that a serious agent or thinker must, as Davidson held, be a language user. The inferential relations attributed by scorekeepers to pragmatically defined occurrences can only be expressed by a structured language with components such as predicates, singular terms and pronouns. Inferential roles are only learnable and projectable on this basis (Brandom 1994: Chapter 6). Thus Brandom’s account provides a pragmatic-semantic story with which to transcendentally partition PPS.
If posthumans are to be intentional agents in thrall to concepts, they will be subjects of discourse assessing one another according to public inferential proprieties.
The Norm-Grounding Problem
However, we only have reason to adopt this a priori portioning of PPS if normativism can contend with some difficult foundational issues. I will refer to the most pressing of these as “the norm-grounding problem”.
Brandom’s pragmatics implies that the rules which furnish deontic statuses are implicit in what we do, in our linguistic and non-linguistic performances, rather than in some explicit set of semantic rules. But what does it mean for a norm to be implicit in a practice? (Brandom 1994: 29-30; Hattiangadi 2003: 420; Rosen 1997).
Are norms a special kind of fact, to which our practices conform or fail to conform? If there were normative facts that transcended our actions, this could at least explain how our inferences can be held to account by them.
Brandom rejects factualism regarding norms. They are not, he claims, “part of the intrinsic nature of things, which is entirely indifferent to them” (48: Rosen 1997: 163-4).
This seems wise. If there were Platonic norms, it is far from clear how animals like us, or our evolutionary forebears, could come to be aware of them (see next section).
Brandom thus adopts a nonfactualist or “phenomenalist” position regarding norms. Non-normative reality is “clothed” in a web of normative statuses when speakers treat public actions as correct or incorrect, permitted or entitled (Brandom 1994: 48).
However, before considering Brandom’s nonfactualist account of norms in greater detail, it is instructive to consider a superficially appealing position that he rejects: regularism. Regularism is the claim that norms are regularities. To act according to a norm (or follow a rule) is simply to behave in conformity with a regularity (27).
Regularism is consonant with pragmatism because one can obey a regularity without having explicit knowledge of it – thereby avoiding the vicious regress that ensues if we require that semantical rules need to be explicitly grasped by speakers (Brandom 1994: 24-5). Regularism is also appealing to philosophical naturalists since it explains how norms depend (or supervene) on facts about the physical state and structure of individual speakers.
However, Brandom rejects this attempt to ground normative claims in factual claims. Here he follows Kripke’s reading of Wittgenstein’s discussion of rule-following: pointing out that any finite sequence of actions will conform to many or even an infinite number of regularities. Thus there is no such thing as the regularity that a finite performance conforms to. For any continuation of that performance “there is some regularity with respect to which it counts as ‘going on in the same way’” (MIE, 28). There are just too many ways of gerrymandering regularities for any given continuation of a performance and the simple regularity view provides no basis for selecting between them. So the simple regularity account fails to explain how a determinate norm can be implicit in practice.
The standard response to the failure of the simple regularity view is to shift attention from finite stretches of performance “to the sets of performances (for instance, applications of a concept) the individual is disposed to produce” (ibid: my emphasis).
The appeal of unpacking grasping a rule in terms of dispositions is that one can be disposed to do an infinite number of things which one does not actually do because of the absence of triggering input (Martin and Heil 1998: 284).
Thus it might seem that we can avoid the gerrymandering objection by saying that different agents A and B grasp the same rules where they are disposed to perform identically given the same triggering inputs.
However dispositionalism seems unable to account for misapplications of a rule.
A might be disposed to behave in the same ways under the same triggering conditions as B but whereas A is correctly following a rule (say plus) B is incorrectly following a different rule (normative behaviour is compatible with recalcitrance [Brandom 1994: 31]). Thus even though A and B exactly coincide in both their actual and their counterfactual performances, they can be following different rules (Martin and Heil 1998: 284-5). Thus if we unpack dispositions counterfactually we will be unable to account for mistakes in application or reasoning. Thus this version of dispositionalism, at least, is unable to explain how norms repose in practices.
So dispositions (if counterfactually conceived) do not help us solve the norm-grounding problem.
Deontic Statuses and Deontic Attitudes
As advertised, Brandom’s favoured account of norms is nonfactualist. We “clothe” a nonnormative world in deontic statuses by taking certain actions or utterances to be correct or incorrect (Brandom 1994: 161).
So normative statuses arise only insofar as there are creatures who can treat one another as committed or entitled to do this or that. In Brandom’s terminology: deontic statuses as assigned when creatures adopt deontic attitudes towards one another.
But what are deontic attitudes?
If they are necessarily intentional – like propositional attitudes – Brandom is stuck in a regress. The philosophical attraction of normative functionalism is that it promises to reduce intention-talk to norm-talk. If deontic attitudes are necessarily intentional, however, he has made little progress in explaining interpreting intentionality via social practices.
Moreover, his account would fail to accord with a minimal Darwinian naturalism. Norm instituting powers cannot have appeared fully formed but must have emerged gradually from the scum of sentience (Rosen 1997). Thus Brandom’s account must be consistent with the claim that merely sentient creatures capable only of reliable discriminatory dispositions to differentially respond to their environments (RDRD’s) – could non-magically acquire a sapient responsiveness to reasons.
Brandom is sensitive to these requirements. He argues that deontic attitudes can occur in “prelinguistic communities” which lack full noetic and agential powers (161). The simplest model of deontic attribution that he provides is one in which performances are assessed as something the performer is authorized by the withholding of sanctions – where sanctioning behaviour, here, is a manifestation of differentially responsive dispositions and not florid interpretative powers.
For example, the deontic status of being entitled to pass through a door might be instituted by a ticketing system in which “the ticket-taker is the attributer of authority, the one who recognizes or acknowledges it and who by taking the ticket as authorizing, makes it authorizing, so instituting the entitlement” (161) This account can be complicated if we introduce deontic attitudes that institute responsibilities on the part of agents.
For example, taking the Queen’s shilling makes one liable to court martial if certain military duties are not undertaken (163). According to Brandom these cases illustrate how social actors can partition “the space of possible performances into those that have been authorized and those that have not, by being disposed to respond differently in the two cases” (161-2: emphasis added).
Does this model show that Brandom’s account can satisfy the minimal naturalist constraints that he recognizes? A number of commentators – including Daniel Dennett and Anandi Hattiangadi – have pointed out that that it succumbs to the gerrymandering objections that Brandom cites against regularism (Dennett 2010; Hattiangadi 2003). Any performative regularities (actual or counterfactual) exhibited by actors and sanctioners in this simple model will be consistent with multiple normative readings of either behaviours – including interpretations which render the “deontic attitudes” mistaken. If the gerrymandering argument refutes regularist theories of rule-following, it refutes dispositionalist accounts of deontic attitudes.
As Hattiangadi points out, beefing up the noetic powers of instituters will avail little. If we furnish sanctioners with the power to make contentful judgements (about whether an agent is entitled to pass through the door, for example) we are already in the realm of the intentional (Hattiangadi 2003: 428).
It follows that a naturalistically constrained normativism does not appear able to explain how social beings can institute norms, thus normative statuses, thus determinate inferential semantic contents, without a vitiating appeal to florid intentional powers.
The Interpretationist Defense
Can Brandom’s account be repaired in a way that meets his minimal naturalist commitment?
Well, one defense that seems consistent with Brandom’s avowals elsewhere is to follow Davidson and Dennett by claiming that the certain kinds of social behaviour are norm-governed if a) members of our speech community would properly interpret them as normative or b) if an ideally rational interpreter privy to all the relevant behavioral facts would read them as normative. This response has something to recommend it. When interpreting alien social practices we are liable to appeal to our own background assumptions about what performances belong to the sortal “social practice”. Moreover, appealing to notion of an ideal interpreter can be of value when trying to understand the theoretical and empirical constraints on attributions of semantic or normative content.
However, as Hattiangadi remarks, this response misses the point of the dispositional analysis of deontic attitudes. This was to explain how a non-sapient community could bootstrap itself into sapience by setting up a basic deontic scorekeeping system. Appealing to actual or ideal interpreters simply replicates the problem with Dennett’s intentional stance approach since it tells us nothing about the conditions under which a being qualifies as a potential interpreter and thus little about the conditions for meaning, understanding or agency.
Similar problem afflicts Joseph Heath’s (2001) proposal that Brandomian norms emerge from reciprocal expectations supported by sanctions. The idea is that a first person acts in a certain way while expecting a sanctioning response from a second person. The second person, meanwhile, is disposed to respond to certain performances with sanctioning behaviour while the first person recognizes this. Where this minimal intersubjective couple converges towards a single pattern of behaviour over time, Heath argues, we are entitled to treat their activity as implying a norm.
Heath’s proposal may be fine if we assume that certain intentional powers are already in place – e.g. that each individual both expects and sanctions the activity of the other. However, as Hattiangadi’s appeal to the gerrymandering argument shows, this structure presupposes beings capable of intentional states such as expecting and sanctioning. This is presumably what distinguishes it from simpler cases of dynamical coupling where two physical systems converge towards a single pattern of behaviour. But if the normativist is serious about explaining the intentional in normative terms, they are not entitled to these assumptions.
If Brandom is right about the defects of Dennett-style or Davidson-style interpretationism, the tendency for his own account to regress to those positions is telling. It suggests that interpretationist accounts cannot explain the semantic or the intentional without regressing to assumptions about ideal interpreters or background practices whose scope they are incapable of delimiting.
The point is not that interpretationism is false but that it is ultimately unilluminating. It is empirically unproblematic that we interpret other speakers, texts, cultural artifacts, etc. However, if in-principle interpretation according to the intentional stance fixes the content of intentional discourse, but nature of such interpretation is ill-defined we have merely satisfied our curiosity about the nature of mindedness by appealing to our local mind-reading techniques. We do not yet know what the invariants (if any) of intentional interpretation are. Another way of putting this is that our practices of interpretation and deontic assessment are phenomenologically “dark”. The fact that we have them and have a little empirical knowledge of them leaves us ignorant both of their underlying nature and (by extension) of the space of interpretative and psychological possibility. Normativist ABP and its interpretationist variants thus provide no future-proof constraints on the space of possible minds or possible agents (See also Bakker 2014).
If so, then they provide no warrant for the claim that any serious agent must be a “subject of discourse” able to measure its own performances against public standards. Presumably, humans are agents of this kind, but the phenomenological darkness surrounding normativity implies that we should not presume that we understand what normativity must involve.
It follows that Anthropologically Unbounded Posthumanism is not seriously challenged by the argument that mind and meaning are constituted by social practices. AUP implies that we can infer no claims about the denizens of Posthuman Possibility Space a priori, by reflecting on the pragmatic transcendental conditions for semantic content. We thus have no reason to suppose that posthuman agents would have to be subjects of discourse of members of communities.
Nor (given our lack of any transcendental grasp of agency) are we entitled to reflect on the ethical status of very strange posthumans. We have no future-proof grasp of how strange posthumans might be, so we lack any basis for adjudicating the moral status of such beings. We may buy into a parochial humanism which accords humans subjects a level of moral consideration that is greater than the nonhuman creatures we know about. But this does not entail that there are not morally considerable states of being in PPS of which we are currently unaware which have little in common with the modes of being accessible to current humans. If posthuman politics is anthropologically unbounded, in this way, then any ethical assessment of the posthuman must follow on its historical emergence. If we want to do serious posthuman ethics, we need to make posthumans or become posthuman.
Bakker, Scott. 2014. The Blind Mechanic II: Reza Negarestani and the Labor of Ghosts | Three Pound Brain. Retrieved April 30, 2014, from https://rsbakker.wordpress.com/2014/04/13/the-blind-mechanic-ii-reza-negarestani-and-the-labour-of-ghosts
Brandom, R. 1994. Making it Explicit: Reasoning, representing, and discursive commitment. Harvard university press.
Brandom, R. 2001. Articulating Reasons: An Introduction to Inferentialism. Cambridge Mass.: Harvard University Press.
Brandom, R. 2002. Tales of the Mighty Dead: Historical Essays in the Metaphysics of Intentionality. Cambridge: Cambridge University Press.
Brandom, R. 2006. “Kantian Lessons about Mind, Meaning, and Rationality.” Southern Journal of Philosophy 44: 49–71.
Brandom, R. 2007. “Inferentialism and Some of Its Challenges.” Philosophy and Phenomenological Research 74 (3): 651–676.
Brassier, R. 2011. “The View from Nowhere.” Identities: Journal for Politics, Gender and Culture (17): 7–23.
Davidson, D. 1986. “A Nice Derangement of Epitaphs.” In Truth and Interpretation, E. LePore (ed), 433-46. Oxford: Blackwell.
Davidson, D. 1984. Inquiries into Truth and Interpretation. Oxford: Clarendon Press.
Dennett, D. C. (1989). The intentional stance. MIT press.
Dennett, D. C. (2010). The evolution of “why?”: An essay on Robert Brandom’s Making it explicit.
Hattiangadi, A. (2003). Making it implicit: Brandom on rule following. Philosophy and Phenomenological Research, 66(2), 419-431.
Heath, J. (2001). Brandom et les sources de la normativité. Philosophiques, 28(1), 27-46.
Heil, John & Martin, C. B. (1998). Rules and powers. Philosophical Perspectives 12 (S12):283-312.
Hohwy, J. (2006). Internalized meaning factualism. Philosophia, 34(3), 325-336.
Kraut, Robert (2010). Universals, metaphysical explanations, and pragmatism. Journal of Philosophy 107 (11):590-609.
Lewis, Kevin. 2013. ”Carnap, Quine and Sellars on Abstract Entities”, https://www.academia.edu/2364977/Carnap_Quine_and_Sellars_on_Abstract_Entities (Accessed 12-7-14)
Roden, David. 2012. “The Disconnection Thesis”. In The Singularity Hypothesis: A Scienti?c and Philosophical Assessment, A. Eden, J. Søraker, J. Moor & E. Steinhart (eds), 281–98. London: Springer.
Roden, David. 2013. “Nature’s Dark Domain: An Argument for a Naturalised Phenom- enology”. Royal Institute of Philosophy Supplements 72: 169–88.
Roden, David. 2014. Posthuman Life: Philosophy at the Edge of the Human. Routledge.
Rosen, Gideon. 1997. “Who Makes the Rules Around Here?”, Philosophy and Phenomenological Research, Vol. 57, No. 1 (Mar., 1997), pp. 163-171
Wanderer, Jeremy (2008). Robert Brandom. Acumen/McGill-Queens University Press.
Wennemann, D. J. 2013. Posthuman Personhood. New York: University Press of America
This formulation allows that posthumans could be descended from technological assemblages which are existentially dependent on servicing “narrow” human goals. Becoming nonhuman in this sense is not a matter of losing a human essence but of ceasing to belong to a human-oriented socio-technical system: the Wide Human (Roden 2012; 2014). I refer to the claim that becoming posthuman consists in becoming independent of the Wide Human as “the Disconnection Thesis”.
Brandom also follows Kant in trying to understand semantic notions like reference and truth in terms of their roles in articulating judgement rather than as semantic or representational primitives.
Intentional systems are unlikely to contain sawdust or stuffing, but IS theory is agnostic regarding their internal machinery or phenomenology. Thus IS theory undercuts both eliminativist and reductionist accounts of intentionality while providing a workable methodology investigations into the mechanisms that actuate intentional systems.
“The key to the account is that an interpretation of this sort must interpret community members as taking or treating each other in practice as adopting intentionally contentful commitments and other normative statuses” (Brandom 1994: 61)
 I can express the belief that there is a cat behind that wall with a sentence in some natural language but I am also able to use the same sentence to attribute this belief to others.
 His subsequent, very detailed, analysis of subsentential expressions is necessarily decompositional rather than compositional – analyzing down rather than building up from simpler semantic components.
The point of attributions of belief or desire, for example, is to determine what an agent is committed entitled “to say or do”. Likewise, the point of affixing truth values to beliefs or statements is to assess or endorse their propriety within the game of giving and asking for reasons. Is the claimant entitled to assert that p? Are the inferential consequences of p that they acknowledge the actual consequences? (17, 542).
So for a rule with infinite application, it is not necessary for the rule user to have all the triggering instances “before his mind” to have grasped how to perform in any of these instances
Martin and Heil 1998 and Hohwy present a good case for holding that dispositions can avoid Kripkensteinean skeptical conclusions if construed realistically rather than in terms of statements about counterfactual behaviour.
 “Looking at the practices a little more closely involves cashing out the talk of deontic statuses by translating it into talk of deontic attitudes. Practitioners take or treat themselves and others as having various commitments and entitlements. They keep score on deontic statuses by attributing those statuses to others and undertaking them themselves. The significance of a performance is the difference it makes in the deontic score-that is, the way in which it changes what commitments and entitlements the practitioners, including the performer, attribute to each other and acquire, acknowledge, or undertake themselves.” (Brandom 1994: 166).
In a Notre Dame review of Phenomenology and Naturalism: Examining the Relationship between Human Experience and Nature, edited by Havi Carel and Darian Meacham, Evan Thomson criticizes my claim that the existence of dark phenomenology implies that phenomenology must be a naturalistic discipline without transcendental warrant. He is correct about my aims and provides a neat summary of my account of dark phenomenology:
David Roden argues that phenomenology should be retained only as a descriptive, empirical method for providing data about experience. This method must be recognized as limited, because it cannot penetrate “dark phenomena” that are not available to introspection or reflective intuition, such as very fine-grained perceptual discriminations of shades of color that cannot be held in memory, or the deep structure of temporal experience. Roden’s discussion of these dark phenomena is illuminating, but his conclusion about the status of phenomenology does not follow. Although he is right that phenomenology cannot be a completely autonomous investigation, but rather must be informed by experimental investigations, it hardly follows that all that phenomenology can do is provide data about what is available to introspection. On the contrary, as the articles by Zahavi, Ratcliffe, Wheeler, and Morris demonstrate, phenomenology can provide new concepts and models for enriching our understanding of nature.
However, I don’t think Thomson’s objection will do as it stands. The position developed in “Nature’s Dark Domain” is consistent with phenomenology being conceptually productive and revealing about nature. If phenomenology is not completely “dark”, it could not be otherwise. I only argue that phenomenological reflection cannot provide future proof (a priori) grounds for claims about invariants of experience or being because – alone and unaided – it cannot tell us what our phenomenology is.
For this reason, my position differs from Mike Wheeler’s “Science Friction: Phenomenology, Naturalism and Cognitive Science” from the same volume. There Wheeler argues that transcendental phenomenology can unpack the “constitutive” conditions of cognition and agency – which tell us what it is, in general, to be an agent or a cognizer – while cognitive science reveals the causal “enabling” conditions for cognition and agency. For example, he claims that Heidegger’s phenomenology of coping is illuminated by experiments in situated robotics using action-oriented representations – which represent an agent’s world in terms of the way it interacts with its body.
So the transcendental/constitutive conditions for agency may require that contextual relevance and an understanding of affordances is necessary for agency, while action-oriented representations reveal one way in which contextual relevance is enabled in representational mechanisms (Wheeler 2013: 143, 152; 2005 197).
According to Wheeler, this model furnishes a minimal naturalism which “domesticates” the transcendental: constitutive conditions are subject to empirically-motivated revision.
However, the kind of revision that Wheeler envisages in his essay seems modest. For example, Heidegger’s account of temporality as thrownness implies that the human agent always encounters the world “embedded within a pre-structured field of intelligibility into which she has been enculturated.” (Wheeler 2013: 158) Wheeler allows that both the mechanisms and the cultural forms of this field can be revealed scientifically (e.g. via cognitive science or ethology):
A consequence of this temporality-driven cultural conditioning of the transcendental is that although there will be specific factors that are transcendentally presupposed by any particular act of sense-making there is no expectation that those factors will be permanently fixed for all human psychological phenomena across space and time (160)
Earlier in his essay, Wheeler provides a succinct account of the epistemological commitments of naturalism: namely that for the naturalist, science and philosophy are continuous. If so, there is no point in this continuum that can be immune from revision in principle – even transcendental claims about the structure of temporality in human agents. It follows that all constitutive claims are empirically defeasible. There is no interesting epistemological boundary to be called between the transcendentally constitutive structure and the various “fillers” for that structure revealed by science Now, this is just what we would expect if – as I argue – the deep structures posited by phenomenology give only limited insight to bare reflection or phenomenological interpretation.
Thus if the deep structure of lived time is not given to us we have a limited first-person grasp of its nature and scope. A deconstructive reading of Heideggerian temporality, for example, implies that the differential or “ecstatic” model of temporality generalizes well beyond transcendental subjects to structures of “generalized writing” found at all levels of biological and technological existence (Stiegler 1998; Hägglund 2008, 2011). The point being not that deconstruction provides a wider-ranging transcendental warrant but that it reveals an indeterminacy in the more narrowly phenomenological ones. If we do not know what temporality is or what must “have it”, we cannot claim to know that all serious agents must have a culturally pre-structured field, for we have produced only a loose, holistic model of a process whose underlying nature is not reflectively available to us, and which may not even be holistic in the phenomenological sense. If the depth-structure of temporality is dark, the constitutive features of all the phenomena where it is supposedly involved as are also occluded. Thus claims about constitutive conditions of cognition and agency are fodder for empirical defeat even where they yield passing insight into nature.
Hägglund, M. 2008. Radical Atheism: Derrida and the Time of Life. Stanford, CA: Stanford University Press.
____2011. “The Trace of Time and the Death of Life: Bergson, Heidegger, Derrida”, https://www.youtube.com/watch?v=9qqaHGUiew4 (accessed November 2011).
Roden, D. 2013. Nature’s Dark Domain: an Argument for a Naturalised Phenomenology. Royal Institute of Philosophy Supplement, 72, 169-188.
Stiegler, B. 1998. Technics and Time, 1: The Fault of Epimetheus, Vol. 1. Stanford, CA: Stanford University Press.
Wheeler, M. 2005. Reconstructing the Cognitive World: The Next Step. Cambridge, MA: MIT Press.
____ 2013. Science Friction: Phenomenology, Naturalism and Cognitive Science. Royal Institute of Philosophy Supplement, 72, 135-167.
Catherine Malabou has an intriguing piece on the vexed question of the relationship between the “humanities” and science in the journal Transeuropeennes here.
It is dominated by a clear and subtle reading of Kant, Foucault and Derrida’s discussion of the meaning of Enlightenment and modernity. Malabou argues that the latter thinkers attempt to escape Kantian assumptions about human invariance by identifying the humanities with “plasticity itself”. The Humanities need not style themselves in terms of some invariant essence of humanity. They can be understood as a site of transformation and “deconstruction” as such. Thus for Derrida in “University Without Condition”, the task of the humanities is:
the deconstruction of « what is proper to man » or to humanism. The transgression of the transcendental implies that the very notion of limit or frontier will proceed from a contingent, that is, historical, mutable, and changing deconstruction of the frontier of the « proper ».
Where, as for Foucault, the deconstruction of the human involves exhibiting its historical conditions of possibility and experimenting with these by, for example, thinking about “our ways of being, thinking, the relation to authority, relations between the sexes, the way in which we perceive insanity or illness “.
This analysis might suggest that the Humanities have little to fear from technological and scientific transformations of humans bodies or minds; they are just the setting in which the implications of these alterations are hammered out.
This line of thought reminds me of a revealingly bad argument produced by Andy Clark in his Natural Born Cyborgs:
The promise, or perhaps threatened, transition to a world of wired humans and semi-intelligent gadgets is just one more move in an ancient game . . . We are already masters at incorporating nonbiological stuff and structure deep into our physical and cognitive routines. To appreciate this is to cease to believe in any post-human future and to resist the temptation to define ourselves in brutal opposition to the very worlds in which so many of us now live, love and work (Clark 2003, 142).
This is obviously broken-backed: that earlier bootstrapping didn’t produce posthumans doesn’t entail that future ones won’t. Even if humans are essentially self-modifying it doesn’t follow that any prospective self-modifying entity is human.
The same problem afflicts Foucault and Derrida’s attempts to hollow out a reservation for humanities scholars by identifying them with the promulgation of transgression or deconstruction. Identifying the humanities with plasticity as such throws the portals of possibility so wide that it can only refer to an abstract possibility space whose contents and topology remains closed to us. If, with Malabou, we allow that some of these transgressions will operate on the material substrate of life, then we cannot assume that its future configurations will resemble human communities or human thinkers – thinkers concerned with topics like sex, work and death for example.
Malabou concludes with the suggestion that Foucault and Derrida fail to confront a quite different problem. They do not provide a historical explanation of the possibility of transformations of life and mind to which they refer:
They both speak of historical transformations of criticism without specifying them. I think that the event that made the plastic change of plasticity possible was for a major part the discovery of a still unheard of plasticity in the middle of the XXth century, and that has become visible and obvious only recently, i.e. the plasticity of the brain that worked in a way behind continental philosophy’s back. The transformation of the transcendental into a plastic material did not come from within the Humanities. It came precisely from the outside of the Humanities, with again, the notion of neural plasticity. I am not saying that the plasticity of the human as to be reduced to a series of neural patterns, nor that the future of the humanities consists in their becoming scientific, even if neuroscience tends to overpower the fields of human sciences (let’s think of neurolinguistics, neuropsychoanalysis, neuroaesthetics, or of neurophilosophy), I only say that the Humanities had not for the moment taken into account the fact that the brain is the only organ that grows, develops and maintains itself in changing itself, in transforming constantly its own structure and shape. We may evoke on that point a book by Norman Doidge, The Brain that changes itself. Doidge shows that this changing, self-fashioning organ is compelling us to elaborate new paradigms of transformation.
I’m happy to concede that the brain is a special case of biological plasticity, but, as Eileen Joy notes elsewhere, the suggestion that the humanities have been out of touch with scientific work on the brain is unmotivated. The engagement between the humanities (or philosophy, at least) and neuroscience already includes work as diverse as Paul and Patricia Churchland’s work on neurophilosophy and Derrida’s early writings on Freud’s Scientific Project.
I’m also puzzled by the suggestion that we need to preserve a place for transcendental thinking at all here. Our posthuman predicament consists in the realization that we are alterable configurations of matter and that our powers of self-alteration are changing in ways that put the future of human thought and communal life in doubt. This is not a transcendental claim. It’s a truistic generalisation which tells us little about the cosmic fate of an ill-assorted grab bag of academic disciplines.
Clark, A. 2003. Natural-born Cyborgs: Minds, Technologies, and the Future of Human Intelligence. New York: Oxford University Press.
CALL FOR PAPERS
The Society for European Philosophy and Forum for European Philosophy
Joint Annual Conference
Philosophy After Nature
3-5 September 2014
The Joint Annual Conference of The Society for European Philosophy and Forum for European Philosophy in 2014 will be hosted by the Centre for the Humanities, the Faculty of Humanities and the Descartes Institute, Utrecht University, the Netherlands.
Professor Michel Serres, Stanford University, Académie française
Information and Thinking/l’information et la pensée
respondent: Professor Françoise Balibar, Université Paris-Diderot
Professor Rahel Jaeggi, Humboldt-Universität zu Berlin
Critique of Forms of Life
Professor Mark B.N. Hansen, Duke University
Entangled in Media, Towards a Speculative Phenomenology of Microtemporal Operations
The SEP/FEP conference is the largest annual event in Europe that aims to bring together researchers, teachers and others, from different disciplines, interested in all areas of contemporary European philosophy. Submissions are therefore invited for individual papers and panel sessions in all areas of contemporary European philosophy. For 2014, submissions that address the conference’s plenary theme – Philosophy After Nature – are particularly encouraged. This would include papers and panels that are after nature in the sense of being in pursuit of nature’s consequences. We invite perspectives on critique, science, ecology, technology and subjectivity as bound up with conceptions of nature and experiment with various positions in contemporary thought.
Abstracts of 500 words for individual paper submissions and proposals for panels should be sent to Rick Dolphijn (firstname.lastname@example.org) by 17 May 2014. Proposals for panels should include a 500-word abstract for each paper within the panel. Proposals from academics, graduate students and independent scholars are welcome.
Conference committee: Rosi Braidotti, Bert van den Brink, Rick Dolphijn, Iris van der Tuin and Paul Ziche.
Enquiries: Rick Dolphijn (email@example.com)
People and cultures have some non-overlapping beliefs. Some folk believe that there is a God, some that there is no God, some that there are many gods. Some people believe that personal autonomy is a paramount value, while others feel that virtues like honour and courage take precedence over personal freedom. These core beliefs are serious, in that they make a difference to whether people live or die, or are able to live the kinds of life that they wish. People fight and die for the sake of autonomy. People fight, die or (institute gang rapes) in the interests of personal honour.
Some folk – the self-styled pluralists – believe that respect for otherness is a paramount political value. Respecting otherness, they say, is so paramount that it should regulate our ontological commitments – our assumptions about what exists. I must admit that I find this hard to credit ontologically or ethically. But it is also unclear how we should spell the principle out. So I’ll consider two versions that have circulated in the blogosphere recently. The first, I will argue, teeters on incoherence or, where not incoherent, is hard to justify in ethical or political terms. The second – which demands that we build a common world – may also be incoherent, but I will argue that we have no reason to think that its ultimate goal is realisable.
According to Philip at Circling Squares Isabel Stengers and Bruno Latour think that this position should enjoin us to avoid ridiculing or undermining others’ values or ontologies. Further, that we should:
grant that all entities exist and, second, that to say that someone’s cherished idol (or whatever disputed entity they hold dear) is non-existent is a ‘declaration of war’ – ‘this means war,’ as Stengers often says.
I’ll admit that I find first part of this principle this damn puzzling. Even if we assume – for now – that it is wrong to attempt to undermine another person’s central beliefs this principle seems to require a) that people actually embrace ontological commitments that are contrary to the one’s they adhere to; b) pretend not to have one’s core beliefs; c) adopt some position of public neutrality vis a vis all core beliefs.
The first interpretation (a) results in the principle that one should embrace the contrary of every core belief; or, in effect, that no one should believe everything. So (in the interests of charity) we should pass on.
b) allows us to have beliefs so long as they are unexpressed. Depending on your view of beliefs, this is either incoherent (because there are no inexpressible beliefs) or burdens believers that no one is likely to find it acceptable.
So I take Philip to embrace c). His clarification suggests something along these lines. For example. He claims that it is consistent with respecting otherness to say what we believe about other’s idols but not to publicly undermine their reasons for believing in them. Thus:
Their basic claim seems to be that ‘respect for otherness,’ i.e. political pluralism, can only come from granting the entities that others hold dear an ontology, even if you don’t ‘believe’ in them. You are thus permitted to say ‘I do not follow that god, he has no hold over me’ but you are not permitted to say ‘your god is an inane, infantile, non-existent fantasy, grow up.’ And it’s not just a question of politeness (although there’s that too). The point is to grant others’ idols and deities an existence – one needn’t agree over what that existence entails, over what capacities that entity has or what obligations it impresses upon you as someone in its partial presence but to deny it existence entirely is to ‘declare war’ – to deny the possibility of civil discourse, of pluralistic co-existence.
I must admit that I find this principle of respect puzzling as well. After all, some of my reasons for being an atheist are also reasons against being a theist. So unless this is just an innocuous plea for good manners (which I’m happy to sign up to on condition that notional others show me and mine the same forbearance) it seems to require that all believers keep their reasons for their belief to themselves. This, again, seems to demand an impossible or repugnant quietism.
So, thus far, ontological pluralism seems to be either incoherent or to impose such burdens on all believers that nobody should be required to observe it. There is, of course, a philosophical precedent for restricted ontological quietism in Rawls’ political liberalism. Rawls’ proposes that reasonable public deliberation recognize the “burdens of judgement” by omitted any justification that hinges on “comprehensive” ethical or religious doctrines over which there can be reasonable disagreement (Rawls 2005, 54). Deliberations about justice under Political Liberalism are thus constrained to be neutral towards “conflicting worldviews” so long as they are tolerant and reasonable (Habermas 1995, 119, 124-5).
However, there is an important difference between the political motivations behind Rawlsian public reason and the position of “ontological charity” Philip attributes to Stengers and Latour. Rawls’ is motivated by the need to preserve stability within plural democratic societies. Public reason does not apply outside the domain of political discourse in which reasonable citizens hash out basic principles of justice and constitutional essentials. It is also extremely problematic in itself. Habermas argues that Rawls exclusion of plural ethical or religious beliefs from the public court is self-vitiating because comprehensive perspectives are sources of disagreement about shared principles (for example, the legitimacy of abortion or same-sex marriage) and these must accordingly be addressed through dialogue rather than circumvented if a politically stable consensus is to be achieved (126).
Finally, apart from being incoherent, the principle of ontological charity seems unnecessary. As Levi Bryant points out in his realist retort to the pluralist, people are not the sum of their beliefs. Beliefs can be revised without effacing the believer. Thus an attack on core beliefs is not an attack on the person holding those beliefs.
So it is hard to interpret the claim that we should grant the existence of others’ “idols” as much more than the principle that it is wrong to humiliate, ridicule or insult people because of what their beliefs are. This seems like a good rule of thumb, but it is hard to justify the claim that it is an overriding principle. For example, even if Rushdie’s Satanic Verses “insults Islam” having an open society in which aesthetic experimentation and the critical evaluation of ideas is possible is just more important than saving certain sections of it from cognitive dissonance or intellectual discomfort. Too many people have suffered death, terror and agony because others had aberrant and false core beliefs to make it plausible that these should be immune from criticism or ridicule. A little personal dissonance is a small price to pay for not going to the oven.
So what of the principle that we should build a “common world”. This is set out by Jeremy Trombley in his Struggle Forever blog under the rubric of “cosmopolitics”. Jeremy regards this project as an infinite task that requires us to seek a kind of fusion between different word views, phenomenologies and ontologies:
The project, as Latour, Stengers, James, and others have described it, is to compose a common world. What pluralism recognizes is that, in this project, we all start from different places – Latour’s relativity rather than relativism. The goal, then, (and it has to be recognized that this project is always contingent and prone to failure) is to make these different positions converge, but in a way that doesn’t impose one upon the other as the Modern Nature/Culture dichotomy tends to do. Why should we avoid imposing one on the other? In part because it’s the right thing to do – by imposing we remove or reduce the agency of the other. The claim to unmediated access to reality makes us invulnerable – no other claim has that grounding, and therefore we can never be wrong. But we are wrong – the science of the Enlightenment gave us climate change, environmental destruction, imperialism in the name of rationality (indigenous peoples removed from their land and taken to reeducation facilities where they were taught “rational” economic activities such as farming), and so on. It removed us from the world and placed us above it – the God’s eye view.
I think there a number of things wrong with cosmopolitics as Jeremy describes it here.
Firstly, seeking to alter beliefs or values does not necessarily reduce agency because people are not their beliefs.
Secondly, some worldviews – like the racist belief-systems that supported the European slave trade – just need to be imposed upon because they are bound up with violent and corrupting socio-political systems.
Thirdly, I know of no Enlightenment thinker, or realist, for whom “unmediated access to reality” is a sine qua non for knowledge. Let’s assume that “realism” is the contrary of pluralism here. It’s not clear what unmediated access would be like, but all realists are committed to the view that we we don’t have it since if we believe that reality has a mind-independent existence and nature, it can presumably vary independently of our beliefs about it. In its place, we have various doctrines of evidence and argument that are themselves susceptible to revision. Some analyses of realism suppose that realists are committed to the claim that there is a one true account of the world (the God’s Eye View) but – as pointed out in an earlier post – this commitment is debatable. In any case, supposing the the existence of a uniquely true theory is very different from claiming to have it.
Finally, much hinges on what we mean by a common world here. I take it that it is not the largely mind-independent reality assumed by the realist since – being largely mind-independent – it exists quite independently of any political project. So I take it that Jeremy is adverting something like a shared phenomenology or experience: a kind of fusion of horizons at the end of time. If we inflect “world” in this sense, then there is no reason for believing that such an aim is possible, let alone coherent. This possibility depends on there being structures of worldhood that are common to all beings that can be said to have one (Daseins, say). I’ve argued that there are no reasons for holding that we have access to such a priori knowledge because – like Scott Bakker - I hold that phenomenology gives us very limited insight into its nature. Thus we have no a priori grasp of what a world is and no reason to believe that Daseins (human or nonhuman) could ever participate in the same one. The argument for this is lengthy so I refer the reader to my paper “Nature’s Dark Domain” and my forthcoming book Posthuman Life.
Habermas, Jurgen. 1995. “Reconciliation through the Public Use of Reason: Remarks on John Rawls’s Political Liberalism.” The Journal of Philosophy 92 (3): 109–131.
Rawls, John. 2005. Political Liberalism. Columbia University Press.