Robert Brandom and Posthumanism

On August 23, 2014, in Uncategorized, by enemyin1

Here’s a first draft of my paper for the Philosophy After Nature Conference in Utrecht, Sept 3-5. All feedback gratefully received! (For users a pdf is available here)


David Roden, Open University

Introduction: Bounded Posthumanism

Posthumanism can be critical or speculative in orientation. Both kinds are critical of human-centered (anthropocentric) thinking. However, their rejection of anthropocentricism applies to different areas: Critical Posthumanism (CP) rejects the anthropocentrism of modern philosophy and intellectual life; Speculative Posthumanism (SP) opposes human-centric thinking about the long-run implications of modern technology.

Whereas critical posthumanists are interested in the posthuman as a cultural and political condition, speculative posthumanists are interested in a possibility of certain technologically created nonhuman agents. They claim that there could be posthumans - where posthumans would be “wide human descendants” of current humans that have become nonhuman in virtue of some process of technical alteration (Roden 2012; 2014, Chapter 5).[1]

Despite differences in concern and methodology, however, CP and SP have convergent interests. CP requires that there are no transcendental conditions for agenthood derivable from parochial facts about human agency. If this is right, it must true of possible nonhuman agents as it is of actual nonhuman agents.

For this reason, I distinguish two claims regarding technological successors to current humans: an anthropologically bounded posthumanism (ABP); and an anthropologically unbounded posthumanism (AUP).

ABP holds:

  1. There are unique constraints C on cognition and agency which any being qualifying as a posthuman successor to humans must satisfy.
  2. Agents satisfying C can know that they are agents and can deduce a priori that they satisfy C (they are transcendental constraints)
  3. Humans typically satisfy C.

ABP’s import becomes clearer if we consider the collection of histories whereby posthuman wide descendants of humans could feasibly emerge. I refer to this set as Posthuman Possibility Space (PPS – See Roden 2014: 53).

Given that posthumans would be agents of some kind (See Chapter 6) and given ABP, members of PPS would have to satisfy the same transcendental conditions (C) on agency as humans.

Daryl Wennemann assumes something along these lines in his book Posthuman Personhood. He adopts the Kantian idea that agency consists in the capacity to justify one’s actions according to reasons and shared norms. For Wennemann, a person is a being able to “reflect on himself and his world from the perspective of a being sharing in a certain community.” (Punzo 1969, cited in Wennemann 2013: 47). This is a condition of posthuman agency as much as of human agency

This implies that, whatever the future throws up, posthuman agents will be social and, arguably linguistic beings, even if they are robots or computers, have strange bodies, or even stranger habits. If so, PPS cannot contain non-anthropomorphic entities whose agency is significantly nonhuman in nature.

ABP implies that there are a priori limits on posthuman weirdness.

AUP, by contrast, leaves the nature of posthuman agency to be settled empirically (or technologically). Posthumans might be social, discursive creatures; or they might be different from us in ways that we cannot envisage short of making some posthumans or becoming posthuman ourselves.

AUP thus extends the critical posthumanist rejection of anthropocentrism to the deep time of the technological future. In Posthuman Life I defended it via a critique of Donald Davidson’s work on intentionality; coupling this with a “naturalistic deconstruction” of transcendental phenomenology in its Husserlian and Heideggerian forms (See also Roden 2013).

Some of these arguments, I believe, carry over to the more overtly normativist philosophy of Robert Brandom – a philosopher whose work I did not address in detail there (for reasons of space and incompetence). The account of the relationship between normativity, social practice, intentionality that Brandom provides in Making It Explicit, and in other writings, is one of most impressively detailed, systematic and historically self-aware attempts to explain subjectivity, agency and intentionality in terms of social practices and statuses. It thus merits the appraisal of all philosophical posthumanists, whether they are of a critical or a speculative bent.

First and Second-Class Agents

I will begin with a thumbnail sketch of how Brandom derives a priori conditions of possibility for agency and meaning from a theory of social practices. Then I will consider whether its foundations are capable of supporting this transcendental superstructure.

Brandom is a philosophical pragmatist. Like other pragmatists, he is committed to the claim that our conceptual and intellectual powers are grounded in our practical abilities rather than in relations between mental entities and what they represent (Brandom 2006).[2]

His pragmatism implies a species of interpretationism with regard to intentional content. Interpretationists, like Daniel Dennett, claim that intentional notions such as “belief” do not track inner vehicles of content but help us assess patterns of rational activity on the part of other “intentional systems” (Wanderer 2008). Belief-desire talk is not a folk psychological “theory” about internal states, but a social “craft” for evaluating and predicting other rational agents.

For Dennett, an entity qualifies as an agent with reasons if predicting its behaviour requires interpreters to attribute it the beliefs and desires it ought to have given its nature and environment. A being whose behaviour is voluminously predictable under this “intentional stance” is called an “intentional system” (IS). In IS theory, there is no gap between predictability under the intentional stance and having real intentionality.[3]

Brandom endorses Dennett’s claim that intentional concepts are fundamentally about rendering agency intelligible in the light of reasons. However, he argues that IS theory furnishes an incomplete account or intentionality. Interpretation is an intentional act; thus interpretationists need to elucidate the relationship between attributed intentionality and attributing intentionality. If we do not understand what kind of being could count as a prospective interpreter, we cannot claim to have understood what it is to attribute intentionality in the first place (Brandom 1994: 59).

Brandom goes one further. The intentionality attributed to intrinsically meaningless events or linguistic inscriptions seems entirely derived from interpreters. Similarly with relatively simple IS’s. Maze-running robots or fly-catching frogs can properly be understood from the intentional stance – making them true-believers by Dennett’s lights. But their intentionality seems likewise observer-relative; derived from attitudes of interpreting IS’s (60). To hold otherwise, he argues, is to risk a disabling regress. For if intentionality is derivative all the way up, there can be no real intentional attributions and thus no derivative (non-observer relative) intentionality (60, 276).

Brandom claims that his theory can be read as an account of the conditions an organism must satisfy to qualify an interpreting intentional system; that is to warrant attributions of non-derived intentionality rather than the as-if intentionality we can attribute to simpler organisms or complex devices.

Whatever else the capacity for original or “first class” intentionality includes, it must involve the ability to evaluate the cognizance and rationality of similar beings and thus to be answerable to reasons (61).[4] Entities with first-class intentionality and thus the capacity to assess and answer to reasons in this way are referred to by Brandom as sapient. Entities with only derived intentionality may exhibit the sentient capacity to react in discriminating and optimizing ways to their environment, but the conceptual content of these responses is attributed and observer-relative.

The claim that intentionality or the capacity for objective thought implies the capacity to evaluate other thinkers obviously has a rich post-Kantian lineage. However, one of the clearest arguments for connecting intentionality and the capacity for other-evaluation is provided by Donald Davidson in his essay “Thought and Talk” (Davidson 1984: 155-170).

Davidson begins with the assumption that belief is an attitude of “holding” true some proposition: for example, that there is a cat behind that wall. If belief is holding true it entails a grasp of truth and the possibility of being mistaken; and thus a concept of belief itself. Thus we cannot believe anything without the capacity to attribute true or false beliefs about the same topic to our fellow creatures (Davidson 1984: 170; 2001b: 104).

This capacity presupposes linguistic abilities, according to Davidson, because attributing contents to fellow creatures requires a common idiom of expression.[5] Absent this, the possession of a concept of belief and, thus, the very having of beliefs, is impossible.

Brandom agrees! We need language to have and attribute beliefs, and, by extension, practical attitudes corresponding to desires and intentions (231-2). However, his official account avoids talk of beliefs or intentions in order to steer clear of the picture of beliefs, etc. as inner vehicles of content (sentences in the head, say) rather than social statuses available to discursive creatures like ourselves.

For Brandom, the primary bearers of propositional content are public assertions. Thus he bases his elaborate theory of intentionality not on a theory of mental representations or sub-propositional concepts, but on a pragmatic account of the place of assertions within the social game of “giving and asking for reasons”.

Correlatively, Brandom’s semantics begins with an explanation of how assertions – and their syntactical proxies, sentences – acquire propositional content.[6] Like Wilfred Sellars’ brand of functional semantics, it is framed in terms of the normative role of utterances within social practices which determine how a speaker can move from one position in the language-game to another.

In the case of assertions, the language-transition rules correspond to materially correct inferences such as that x is colored from x is red. Language entry-rules include observation statements which allow us to make claims like “There is snow on the grass” on the basis of our reliable dispositions to differentially respond (RDRDs) to recurrent states of our environment. Finally, “language exit rules” correspond to practical commitments to forms of non-linguistic action.

Thus Brandom agrees with other post-Wittgensteinian pragmatists that linguistic practices are governed by public norms. However, he follows Davidson in rejecting the “I-we” conception of social structure. (39-40; Davidson 1986). If meanings are inferential roles (as Dummett and Sellars also claim), then the content attributable to expressions will dance in line with the doxastic commitments of individual speakers.

Suppose one observes a masked figure in a red costume clambering up a skyscraper. The language entry rules ambient within your community of English speakers may entitle you (by default) to claim that Spiderman is climbing the building. However, you are unaware that Spiderman is none other than Peter Parker. So you are not yet entitled to infer that Peter Parker is climbing the building – although the substitutional rules of English would commit you to that further inference if (say) some reliable authority informed you of this fact.

This simple example shows that the inferential roles – thus meanings – of expressions like “Spiderman” are not fixed communally but have to vary with the auxiliary assumptions, sensitivities and dispositions of individual speakers. Understanding or interpreting the utterances and beliefs of others is thus a matter of deontic scorekeeping – that is keeping track of the way social statuses alter as speakers update their inferential commitments (Brandom 1994: 142).[7]

Thus semantic and intentional content are co-extensive with the normative-functional roles of states and actions. It follows that what a belief or claim “represents” or is “about” is fixed by its status it can be ascribed from the perspective of various deontic scorekeepers (including the believer or claimant).

The second consequence – which I flagged earlier – is that a serious agent or thinker must, as Davidson held, be a language user. The inferential relations attributed by scorekeepers to pragmatically defined occurrences can only be expressed by a structured language with components such as predicates, singular terms and pronouns. Inferential roles are only learnable and projectable on this basis (Brandom 1994: Chapter 6). Thus Brandom’s account provides a pragmatic-semantic story with which to transcendentally partition PPS.

If posthumans are to be intentional agents in thrall to concepts, they will be subjects of discourse assessing one another according to public inferential proprieties.

The Norm-Grounding Problem

However, we only have reason to adopt this a priori portioning of PPS if normativism can contend with some difficult foundational issues. I will refer to the most pressing of these as “the norm-grounding problem”.

Brandom’s pragmatics implies that the rules which furnish deontic statuses are implicit in what we do, in our linguistic and non-linguistic performances, rather than in some explicit set of semantic rules.  But what does it mean for a norm to be implicit in a practice? (Brandom 1994: 29-30; Hattiangadi 2003: 420; Rosen 1997).

Are norms a special kind of fact, to which our practices conform or fail to conform? If there were normative facts that transcended our actions, this could at least explain how our inferences can be held to account by them.

Brandom rejects factualism regarding norms. They are not, he claims, “part of the intrinsic nature of things, which is entirely indifferent to them” (48: Rosen 1997: 163-4).

This seems wise. If there were Platonic norms, it is far from clear how animals like us, or our evolutionary forebears, could come to be aware of them (see next section).

Brandom thus adopts a nonfactualist or “phenomenalist” position regarding norms. Non-normative reality is “clothed” in a web of normative statuses when speakers treat public actions as correct or incorrect, permitted or entitled (Brandom 1994: 48).

However, before considering Brandom’s nonfactualist account of norms in greater detail, it is instructive to consider a superficially appealing position that he rejects: regularism. Regularism is the claim that norms are regularities. To act according to a norm (or follow a rule) is simply to behave in conformity with a regularity (27).

Regularism is consonant with pragmatism because one can obey a regularity without having explicit knowledge of it – thereby avoiding the vicious regress that ensues if we require that semantical rules need to be explicitly grasped by speakers (Brandom 1994: 24-5). Regularism is also appealing to philosophical naturalists since it explains how norms depend (or supervene) on facts about the physical state and structure of individual speakers.

However, Brandom rejects this attempt to ground normative claims in factual claims. Here he follows Kripke’s reading of Wittgenstein’s discussion of rule-following: pointing out that any finite sequence of actions will conform to many or even an infinite number of regularities. Thus there is no such thing as the regularity that a finite performance conforms to. For any continuation of that performance “there is some regularity with respect to which it counts as ‘going on in the same way’” (MIE, 28). There are just too many ways of gerrymandering regularities for any given continuation of a performance and the simple regularity view provides no basis for selecting between them. So the simple regularity account fails to explain how a determinate norm can be implicit in practice.

The standard response to the failure of the simple regularity view is to shift attention from finite stretches of performance “to the sets of performances (for instance, applications of a concept) the individual is disposed to produce” (ibid: my emphasis).

The appeal of unpacking grasping a rule in terms of dispositions is that one can be disposed to do an infinite number of things which one does not actually do because of the absence of triggering input (Martin and Heil 1998: 284).[8]

Thus it might seem that we can avoid the gerrymandering objection by saying that different agents A and B grasp the same rules where they are disposed to perform identically given the same triggering inputs.

However dispositionalism seems unable to account for misapplications of a rule.

A might be disposed to behave in the same ways under the same triggering conditions as B but whereas A is correctly following a rule (say plus) B is incorrectly following a different rule (normative behaviour is compatible with recalcitrance [Brandom 1994: 31]). Thus even though A and B exactly coincide in both their actual and their counterfactual performances, they can be following different rules (Martin and Heil 1998: 284-5). Thus if we unpack dispositions counterfactually we will be unable to account for mistakes in application or reasoning. Thus this version of dispositionalism, at least, is unable to explain how norms repose in practices.

So dispositions (if counterfactually conceived) do not help us solve the norm-grounding problem.[9]

Deontic Statuses and Deontic Attitudes

As advertised, Brandom’s favoured account of norms is nonfactualist. We “clothe” a nonnormative world in deontic statuses by taking certain actions or utterances to be correct or incorrect (Brandom 1994: 161).

So normative statuses arise only insofar as there are creatures who can treat one another as committed or entitled to do this or that. In Brandom’s terminology: deontic statuses as assigned when creatures adopt deontic attitudes towards one another[10].

But what are deontic attitudes?

If they are necessarily intentional – like propositional attitudes – Brandom is stuck in a regress. The philosophical attraction of normative functionalism is that it promises to reduce intention-talk to norm-talk. If deontic attitudes are necessarily intentional, however, he has made little progress in explaining interpreting intentionality via social practices.

Moreover, his account would fail to accord with a minimal Darwinian naturalism. Norm instituting powers cannot have appeared fully formed but must have emerged gradually from the scum of sentience (Rosen 1997). Thus Brandom’s account must be consistent with the claim that merely sentient creatures capable only of reliable discriminatory dispositions to differentially respond to their environments (RDRD’s) – could non-magically acquire a sapient responsiveness to reasons.

Brandom is sensitive to these requirements. He argues that deontic attitudes can occur in “prelinguistic communities” which lack full noetic and agential powers (161). The simplest model of deontic attribution that he provides is one in which performances are assessed as something the performer is authorized by the withholding of sanctions – where sanctioning behaviour, here, is a manifestation of differentially responsive dispositions and not florid interpretative powers.

For example, the deontic status of being entitled to pass through a door might be instituted by a ticketing system in which “the ticket-taker is the attributer of authority, the one who recognizes or acknowledges it and who by taking the ticket as authorizing, makes it authorizing, so instituting the entitlement” (161) This account can be complicated if we introduce deontic attitudes that institute responsibilities on the part of agents.

For example, taking the Queen’s shilling makes one liable to court martial if certain military duties are not undertaken (163). According to Brandom these cases illustrate how social actors can partition “the space of possible performances into those that have been authorized and those that have not, by being disposed to respond differently in the two cases” (161-2: emphasis added).

Does this model show that Brandom’s account can satisfy the minimal naturalist constraints that he recognizes? A number of commentators – including Daniel Dennett and Anandi Hattiangadi – have pointed out that that it succumbs to the gerrymandering objections that Brandom cites against regularism (Dennett 2010; Hattiangadi 2003). Any performative regularities (actual or counterfactual) exhibited by actors and sanctioners in this simple model will be consistent with multiple normative readings of either behaviours – including interpretations which render the “deontic attitudes” mistaken. If the gerrymandering argument refutes regularist theories of rule-following, it refutes dispositionalist accounts of deontic attitudes.

As Hattiangadi points out, beefing up the noetic powers of instituters will avail little. If we furnish sanctioners with the power to make contentful judgements (about whether an agent is entitled to pass through the door, for example) we are already in the realm of the intentional (Hattiangadi 2003: 428).

It follows that a naturalistically constrained normativism does not appear able to explain how social beings can institute norms, thus normative statuses, thus determinate inferential semantic contents, without a vitiating appeal to florid intentional powers.

The Interpretationist Defense

Can Brandom’s account be repaired in a way that meets his minimal naturalist commitment?

Well, one defense that seems consistent with Brandom’s avowals elsewhere is to follow Davidson and Dennett by claiming that the certain kinds of social behaviour are norm-governed if a) members of our speech community would properly interpret them as normative or b) if an ideally rational interpreter privy to all the relevant behavioral facts would read them as normative. This response has something to recommend it. When interpreting alien social practices we are liable to appeal to our own background assumptions about what performances belong to the sortal “social practice”. Moreover, appealing to notion of an ideal interpreter can be of value when trying to understand the theoretical and empirical constraints on attributions of semantic or normative content.

However, as Hattiangadi remarks, this response misses the point of the dispositional analysis of deontic attitudes. This was to explain how a non-sapient community could bootstrap itself into sapience by setting up a basic deontic scorekeeping system. Appealing to actual or ideal interpreters simply replicates the problem with Dennett’s intentional stance approach since it tells us nothing about the conditions under which a being qualifies as a potential interpreter and thus little about the conditions for meaning, understanding or agency.

Similar problem afflicts Joseph Heath’s (2001) proposal that Brandomian norms emerge from reciprocal expectations supported by sanctions. The idea is that a first person acts in a certain way while expecting a sanctioning response from a second person. The second person, meanwhile, is disposed to respond to certain performances with sanctioning behaviour while the first person recognizes this. Where this minimal intersubjective couple converges towards a single pattern of behaviour over time, Heath argues, we are entitled to treat their activity as implying a norm.

Heath’s proposal may be fine if we assume that certain intentional powers are already in place – e.g. that each individual both expects and sanctions the activity of the other. However, as Hattiangadi’s appeal to the gerrymandering argument shows, this structure presupposes beings capable of intentional states such as expecting and sanctioning. This is presumably what distinguishes it from simpler cases of dynamical coupling where two physical systems converge towards a single pattern of behaviour. But if the normativist is serious about explaining the intentional in normative terms, they are not entitled to these assumptions.

Unbounded Posthumanism

If Brandom is right about the defects of Dennett-style or Davidson-style interpretationism, the tendency for his own account to regress to those positions is telling. It suggests that interpretationist accounts cannot explain the semantic or the intentional without regressing to assumptions about ideal interpreters or background practices whose scope they are incapable of delimiting.

The point is not that interpretationism is false but that it is ultimately unilluminating. It is empirically unproblematic that we interpret other speakers, texts, cultural artifacts, etc. However, if in-principle interpretation according to the intentional stance fixes the content of intentional discourse, but nature of such interpretation is ill-defined we have merely satisfied our curiosity about the nature of mindedness by appealing to our local mind-reading techniques. We do not yet know what the invariants (if any) of intentional interpretation are. Another way of putting this is that our practices of interpretation and deontic assessment are phenomenologically “dark”. The fact that we have them and have a little empirical knowledge of them leaves us ignorant both of their underlying nature and (by extension) of the space of interpretative and psychological possibility. Normativist ABP and its interpretationist variants thus provide no future-proof constraints on the space of possible minds or possible agents (See also Bakker 2014).

If so, then they provide no warrant for the claim that any serious agent must be a “subject of discourse” able to measure its own performances against public standards. Presumably, humans are agents of this kind, but the phenomenological darkness surrounding normativity implies that we should not presume that we understand what normativity must involve.

It follows that  Anthropologically Unbounded Posthumanism is not seriously challenged by the argument that mind and meaning are constituted by social practices. AUP implies that we can infer no claims about the denizens of Posthuman Possibility Space a priori, by reflecting on the pragmatic transcendental conditions for semantic content. We thus have no reason to suppose that posthuman agents would have to be subjects of discourse of members of communities.

Nor (given our lack of any transcendental grasp of agency) are we entitled to reflect on the ethical status of very strange posthumans. We have no future-proof grasp of how strange posthumans might be, so we lack any basis for adjudicating the moral status of such beings. We may buy into a parochial humanism which accords humans subjects a level of moral consideration that is greater than the nonhuman creatures we know about. But this does not entail that there are not morally considerable states of being in PPS of which we are currently unaware which have little in common with the modes of being accessible to current humans. If posthuman politics is anthropologically unbounded, in this way, then any ethical assessment of the posthuman must follow on its historical emergence. If we want to do serious posthuman ethics, we need to make posthumans or become posthuman.



Bakker, Scott. 2014. The Blind Mechanic II: Reza Negarestani and the Labor of Ghosts | Three Pound Brain. Retrieved April 30, 2014, from

Brandom, R. 1994. Making it Explicit: Reasoning, representing, and discursive commitment. Harvard university press.

Brandom, R.  2001. Articulating Reasons: An Introduction to Inferentialism. Cambridge Mass.: Harvard University Press.

Brandom, R. 2002. Tales of the Mighty Dead: Historical Essays in the Metaphysics of Intentionality. Cambridge: Cambridge University Press.

Brandom, R. 2006. “Kantian Lessons about Mind, Meaning, and Rationality.” Southern Journal of Philosophy 44: 49–71.

Brandom, R. 2007. “Inferentialism and Some of Its Challenges.” Philosophy and Phenomenological Research 74 (3): 651–676.

Brassier, R. 2011. “The View from Nowhere.” Identities: Journal for Politics, Gender and Culture (17): 7–23.

Davidson, D.  1986. “A Nice Derangement of Epitaphs.” In Truth and Interpretation, E. LePore (ed), 433-46. Oxford: Blackwell.

Davidson, D. 1984. Inquiries into Truth and Interpretation. Oxford: Clarendon Press.

Dennett, D. C. (1989). The intentional stance. MIT press.

Dennett, D. C. (2010). The evolution of “why?”: An essay on Robert Brandom’s Making it explicit.

Hattiangadi, A. (2003). Making it implicit: Brandom on rule following. Philosophy and Phenomenological Research66(2), 419-431.

Heath, J. (2001). Brandom et les sources de la normativité. Philosophiques28(1), 27-46.

Heil, John & Martin, C. B. (1998). Rules and powers. Philosophical Perspectives 12 (S12):283-312.

Hohwy, J. (2006). Internalized meaning factualism. Philosophia34(3), 325-336.

Kraut, Robert (2010). Universals, metaphysical explanations, and pragmatism. Journal of Philosophy 107 (11):590-609.

Lewis, Kevin. 2013. ”Carnap, Quine and Sellars on Abstract Entities”, (Accessed 12-7-14)

Roden, David. 2012. “The Disconnection Thesis”. In The Singularity Hypothesis: A Scienti?c and Philosophical Assessment, A. Eden, J. Søraker, J. Moor & E. Steinhart (eds), 281–98. London: Springer.

Roden, David. 2013. “Nature’s Dark Domain: An Argument for a Naturalised Phenom- enology”. Royal Institute of Philosophy Supplements 72: 169–88.

Roden, David. 2014. Posthuman Life: Philosophy at the Edge of the Human. Routledge.

Rosen, Gideon. 1997. “Who Makes the Rules Around Here?”, Philosophy and Phenomenological Research, Vol. 57, No. 1 (Mar., 1997), pp. 163-171

Wanderer, Jeremy (2008). Robert Brandom. Acumen/McGill-Queens University Press.

Wennemann, D. J. 2013. Posthuman Personhood. New York: University Press of America





[1]This formulation allows that posthumans could be descended from technological assemblages which are existentially dependent on servicing “narrow” human goals. Becoming nonhuman in this sense is not a matter of losing a human essence but of ceasing to belong to a human-oriented socio-technical system: the Wide Human (Roden 2012; 2014). I refer to the claim that becoming posthuman consists in becoming independent of the Wide Human as “the Disconnection Thesis”.

[2]Brandom also follows Kant in trying to understand semantic notions like reference and truth in terms of their roles in articulating judgement rather than as semantic or representational primitives.

[3]Intentional systems are unlikely to contain sawdust or stuffing, but IS theory is agnostic regarding their internal machinery or phenomenology. Thus IS theory undercuts both eliminativist and reductionist accounts of intentionality while providing a workable methodology investigations into the mechanisms that actuate intentional systems.

[4]“The key to the account is that an interpretation of this sort must interpret community members as taking or treating each other in practice as adopting intentionally contentful commitments and other normative statuses” (Brandom 1994: 61)

[5] I can express the belief that there is a cat behind that wall with a sentence in some natural language but I am also able to use the same sentence to attribute this belief to others.

[6] His subsequent, very detailed, analysis of subsentential expressions is necessarily decompositional rather than compositional – analyzing down rather than building up from simpler semantic components.

[7]The point of attributions of belief or desire, for example, is to determine what an agent is committed entitled “to say or do”. Likewise, the point of affixing truth values to beliefs or statements is to assess or endorse their propriety within the game of giving and asking for reasons. Is the claimant entitled to assert that p? Are the inferential consequences of p that they acknowledge the actual consequences?  (17, 542).

[8]So for a rule with infinite application, it is not necessary for the rule user to have all the triggering instances “before his mind” to have grasped how to perform in any of these instances

[9]Martin and Heil 1998 and Hohwy present a good case for holding that dispositions can avoid Kripkensteinean skeptical conclusions if construed realistically rather than in terms of statements about counterfactual behaviour.

[10] “Looking at the practices a little more closely involves cashing out the talk of deontic statuses by translating it into talk of deontic attitudes. Practitioners take or treat themselves and others as having various commitments and entitlements. They keep score on deontic statuses by attributing those statuses to others and undertaking them themselves. The significance of a performance is the difference it makes in the deontic score-that is, the way in which it changes what commitments and entitlements the practitioners, including the performer, attribute to each other and acquire, acknowledge, or undertake themselves.” (Brandom 1994: 166).



Tagged with:

Kevin has provided a typically engaging gloss on the difference between posthumanism and transhumanism over at the IEET site. I don’t fundamentally disagree with his account of transhumanism (though I think he needs to emphasize its fundamentally normative character) but the account of posthumanism he gives here has some shortcomings:

Two significant differences between transhumanism and the posthuman is the posthuman’s focus on information and systems theories (cybernetics), and the posthuman’s consequent, primary relationship to digital technology; and also the posthuman’s emphasis on systems (such as humans) as distributed entities—that is, as systems comprised of, and entangled with, other systems.  Transhumanism does not emphasize either of these things. 

Posthumanism derives from the posthuman because the latter represents the death of the humanist subject: the qualities that make up that subject depend on a privileged position as a special, stand-alone entity that possesses unique characteristics that make it exceptional in the universe—characteristics such as unique and superior intellect to all other creatures, or a natural right to freedoms that do not accrue similarly to other animals.  If the focus is on information as the essence of all intelligent systems, and materials and bodies are merely substrates that carry the all-important information of life, then there is no meaningful difference between humans and intelligent machines—or any other kind of intelligent system, such as animals. 

Now, I realize we can spin definitions to different ends; but even allowing for our different research aims, this won’t do. Posthumanists may, but need not, claim that humans are becoming more intertwined with technology. They may, but need not, claim that functions, relations or systems are more ontologically basic than intrinsic properties. Many arch-humanists are functionalists, holists or relationists (I Kant, R Brandom, D Davidson, G Hegel . . .) and one can agree that human subjectivity is constitutively technological (A Clark) without denying its distinctive moral or epistemological status. Reducing stuff to relations can be a way of emphasizing the transcendentally constitutive status of the human subject, taking anthropocentrism to the max (see below). Emphasizing the externality or contingency of relations can be a way of arguing that things are fundamentally independent of that constitutive activity (as in Harman’s OOO or DeLanda’s assemblage ontology).

So I raise Kevin’s thumbnails with a few of my own.

  • A philosopher is a humanist if she believes that humans are importantly distinct from non-humans and supports this distinctiveness claim with a philosophical anthropology: an account of the central features of human existence and their relations to similarly general aspects of nonhuman existence.
  • A humanist philosophy is anthropocentric if it accords humans a superlative status that all or most nonhumans lack
  • Transhumanists claims that technological enhancement of human capacities is a desirable aim (all other things being equal). So the normative content of transhumanism is largely humanist. Transhumanists just hope to add some new ways of cultivating human values to the old unreliables of education and politics.
  • Posthumanists reject anthropocentrism. So philosophical realists, deconstructionists, new materialists, Cthulhu cultists and naturalists are posthumanists even if they are unlikely to crop up on one another’s Christmas lists.

For more, see my forthcoming book Posthuman Life and my post Humanism, Transhumanism and Posthumanism.


Tagged with:

Conferences in September

On June 21, 2014, in Uncategorized, by enemyin1

I’ll be attending conferences at either end of our continent in September:

Philosophy After Nature, University of Utrecht, 3-5 September

Posthuman Politics, 25th until the 28th of September 2014, University of the Aegean, Department of Cultural Technology and Communication, Geography Building – University Campus.

I’m presenting the same paper at both. Here’s the abstract, though the details of the argument remain to be filled in!

On Reason and Spectral Machines: an anti-normativist response to Bounded Posthumanism

David Roden, The Open University UK

In Posthuman Life I distinguish two speculative claims regarding technological successors to current humans: an anthropologically bounded posthumanism (ABP) and an anthropologically unbounded posthumanism. ABP holds:

1) There are transcendental constrains on cognition and agency that any entity qualifying as a posthuman successor under the Disconnection Thesis (Roden 2012, 2014) would have to obey.

2) These constraints are realized in the structure of human subjectivity and rationality.

One version of ABP is implied by normativist theories of intentionality for which original or “first class” intentionality is only possible for beings that can hold one another publicly to account by ascribing and adopting normative statuses (Brandom 1994). If Normativist ABP is correct, then posthumans – were they to exist – would not be so different from us for they would have to belong to discursive communities and subscribe to inter-subjective norms (See Wennemann 2013).

Normativist ABP thus imposes severe constraints on posthuman “weirdness” and limits the political implications of speculative claims about posthuman possibility such as those in my book. In this paper, I will argue that we should reject Normativist ABP because we should reject normativist theories of intentionality. For normativism to work, it must be shown that the objectivity and “bindingness” of social norms is independent of individual beliefs or endorsements. I will argue that the only way in which this can be achieved is by denying the dependence of normative statuses upon the particular dispositions, states and attitudes of individuals; thus violating plausible naturalistic constraints on normativism.

In response, I will argue for an anthropologically unbounded posthumanism for which all constraints on posthuman possibility must be discovered empirically by making posthumans or becoming posthuman. This implies a similarly unbounded posthuman politics for which there is no universal reason or transhistorical subjectivity.


Bakker, Scott. 2014. The Blind Mechanic II: Reza Negarestani and the Labor of Ghosts | Three Pound Brain. Retrieved April 30, 2014, from

Brandom, R. 1994. Making it Explicit: Reasoning, representing, and discursive commitment. Harvard university press.

Brandom, R.  2001. Articulating Reasons: An Introduction to Inferentialism. Cambridge Mass.: Harvard University Press.

Brandom, R. 2002. Tales of the Mighty Dead: Historical Essays in the Metaphysics of Intentionality. Cambridge: Cambridge University Press.

Brandom, R. 2006. “Kantian Lessons about Mind, Meaning, and Rationality.” Southern Journal of Philosophy 44: 49–71.

Brandom, R. 2007. “Inferentialism and Some of Its Challenges.” Philosophy and Phenomenological Research 74 (3): 651–676.

Brassier, R. 2011. “The View from Nowhere.” Identities: Journal for Politics, Gender and Culture (17): 7–23.

Davidson, D.  1986. “A Nice Derangement of Epitaphs.” In Truth and Interpretation, E. LePore (ed), 433-46. Oxford: Blackwell.

Negarestani, Reza. 2014. The Labor of the Inhuman, Part I: Human | e-flux. Retrieved April 30, 2014, from

Negarestani, Reza. 2014. ‘The Labor of the Inhuman, Part II: The Inhuman’ | e-flux. Retrieved April 30, 2014, from

Roden, D. 2012. “The Disconnection Thesis.” The Singularity Hypothesis: A Scientific and Philosophical Assessment, A. Eden, J. Søraker, J. Moor & E. Steinhart (eds), 281-298. London: Springer.

Roden, David. 2014. Posthuman Life: Philosophy at the Edge of the Human. Routledge.

Turner, S. P. 2010. Explaining the normative. Polity.

Wennemann, D. J. 2013. Posthuman Personhood. New York: University Press of America.


Tagged with:

Computer Music and Posthumanism

On June 9, 2014, in Uncategorized, by enemyin1

A possibly ill-advised idea for a presentation on computer music and posthumanism entitled “Computer Music and Posthumanism”.

I will introduce two flavors of posthumanism: critical posthumanism (CP) and speculative posthumanism (SP) and provide an overview of some of the ways in which they might be explored by thinking through philosophical issues raised by computer music practice.
CP questions the dualist modes of thinking that have traditionally assigned human subjects a privileged place within philosophical thought: for example, the distinction between the formative power of minds and subjects and the inertia of matter.
The use of computers to supplement human performance raises questions about where agency is ascribed. Is it always on the side of the human musician or can it also be ascribed also to the devices or software used to generate sound events? If so, what kind of status can be granted to such artificial agents? Does their agency locally supervene on human agency, for example? I will also argue that the intractability and complexity of some computer generated sound confronts us with the nonhuman, mind-independent reality of sonic events. It thus provides an aesthetic grounding for a posthumanist realism.
SP (by contrast) is a metaphysical possibility claim about technological successors to humans. It can be summed up in the SP Schema: “Descendants of current humans could cease to be human by virtue of a history of technical alteration” CP and SP are conceptually distinct but, I argue, the most radical form of SP converges with the anti-anthropocentrism of CP (Roden 2014). In particular, non-anthropologically bounded SP implies that the only way in which we can acquire substantive knowledge of posthumans is through making posthumans or becoming posthuman. I will argue that computer music development may have a role in this project of engineering a posthuman succession.

Roden, D. 2010b. “Sonic Art and the Nature of Sonic Events.” Review of Philosophy and Psychology 1 (1): 141–156.
Roden, D. 2012. “The Disconnection Thesis.” The Singularity Hypothesis: A Scientific and Philosophical Assessment, A. Eden, J. Søraker, J. Moor & E. Steinhart (eds), 281-298. London: Springer.
Roden, David. 2014. Posthuman Life: Philosophy at the Edge of the Human. Routledge.


Tagged with:

No Future? Catherine Malabou on the Humanities

On February 19, 2014, in Uncategorized, by enemyin1

Catherine Malabou has an intriguing piece on the vexed question of the relationship between the “humanities” and science in the journal Transeuropeennes here.

It is dominated by a clear and subtle reading of Kant, Foucault and Derrida’s discussion of the meaning of Enlightenment and modernity. Malabou argues that the latter thinkers attempt to escape Kantian assumptions about human invariance by identifying the humanities with “plasticity itself”. The Humanities need not style themselves in terms of some invariant essence of humanity. They can be understood as a site of transformation and “deconstruction” as such.  Thus for Derrida in “University Without Condition”, the task of the humanities is:

the deconstruction of « what is proper to man » or to humanism. The transgression of the transcendental implies that the very notion of limit or frontier will proceed from a contingent, that is, historical, mutable, and changing deconstruction of the frontier of the « proper ».

Where, as for Foucault, the deconstruction of the human involves exhibiting its historical conditions of possibility and experimenting with these by, for example, thinking about “our ways of being, thinking, the relation to authority, relations between the sexes, the way in which we perceive insanity or illness “.

This analysis might suggest that the Humanities have little to fear from technological and scientific transformations of humans bodies or minds; they are just the setting in which the implications of these alterations are hammered out.

This line of thought reminds me of a revealingly bad argument produced by Andy Clark in his Natural Born Cyborgs:

The promise, or perhaps threatened, transition to a world of wired humans and semi-intelligent gadgets is just one more move in an ancient game . . . We are already masters at incorporating nonbiological stuff and structure deep into our physical and cognitive routines. To appreciate this is to cease to believe in any post-human future and to resist the temptation to define ourselves in brutal opposition to the very worlds in which so many of us now live, love and work (Clark 2003, 142).

This is obviously broken-backed: that earlier bootstrapping didn’t produce posthumans doesn’t entail  that future ones won’t. Even if humans are essentially self-modifying it doesn’t follow that any prospective self-modifying entity is human.

The same problem afflicts Foucault and Derrida’s attempts to hollow out a reservation for humanities scholars by identifying them with the promulgation of transgression or deconstruction. Identifying the humanities with plasticity as such throws the portals of possibility so wide that it can only refer to an abstract possibility space whose contents and topology remains closed to us. If, with Malabou, we allow that some of these transgressions will operate on the material substrate of life, then we cannot assume that its future configurations will resemble human communities or human thinkers – thinkers concerned with topics like sex, work and death for example.

Malabou concludes with the suggestion that Foucault and Derrida fail to confront a quite different problem. They do not provide a historical explanation of the possibility of transformations of life and mind to which they refer:

They both speak of historical transformations of criticism without specifying them. I think that the event that made the plastic change of plasticity possible was for a major part the discovery of a still unheard of plasticity in the middle of the XXth century, and that has become visible and obvious only recently, i.e. the plasticity of the brain that worked in a way behind continental philosophy’s back. The transformation of the transcendental into a plastic material did not come from within the Humanities. It came precisely from the outside of the Humanities, with again, the notion of neural plasticity. I am not saying that the plasticity of the human as to be reduced to a series of neural patterns, nor that the future of the humanities consists in their becoming scientific, even if neuroscience tends to overpower the fields of human sciences (let’s think of neurolinguistics, neuropsychoanalysis, neuroaesthetics, or of neurophilosophy), I only say that the Humanities had not for the moment taken into account the fact that the brain is the only organ that grows, develops and maintains itself in changing itself, in transforming constantly its own structure and shape. We may evoke on that point a book by Norman Doidge, The Brain that changes itself. Doidge shows that this changing, self-fashioning organ is compelling us to elaborate new paradigms of transformation.

I’m happy to concede that the brain is a special case of biological plasticity, but, as Eileen Joy notes elsewhere, the suggestion that the humanities have been out of touch with scientific work on the brain is unmotivated. The engagement between the humanities (or philosophy, at least) and neuroscience already includes work as diverse as Paul and Patricia Churchland’s work on neurophilosophy and Derrida’s early writings on Freud’s Scientific Project.

I’m also puzzled by the suggestion that we need to preserve a place for transcendental thinking at all here. Our posthuman predicament consists in the realization that we are alterable configurations of matter and that our powers of self-alteration are changing in ways that put the future of human thought and communal life in doubt. This is not a transcendental claim. It’s a truistic generalisation which tells us little about the cosmic fate of an ill-assorted grab bag of  academic disciplines.


Clark, A. 2003. Natural-born Cyborgs: Minds, Technologies, and the Future of Human Intelligence. New York: Oxford University Press.







People and cultures have some non-overlapping beliefs. Some folk believe that there is a God, some that there is no God, some that there are many gods. Some people believe that personal autonomy is a paramount value, while others feel that virtues like honour and courage take precedence over personal freedom. These core beliefs are serious, in that they make a difference to whether people live or die, or are able to live the kinds of life that they wish. People fight and die for the sake of autonomy. People fight, die or (institute gang rapes) in the interests of personal honour.

Some folk – the self-styled pluralists – believe that respect for otherness is a paramount political value. Respecting otherness, they say, is so paramount that it should regulate our ontological commitments – our assumptions about what exists. I must admit that I find this hard to credit ontologically or ethically. But it is also unclear how we should spell the principle out. So I’ll consider two versions that have circulated in the blogosphere recently. The first, I will argue, teeters on incoherence or, where not incoherent, is hard to justify in ethical or political terms. The second – which demands that we build a common world – may also be incoherent, but I will argue that we have no reason to think that its ultimate goal is realisable.

According to Philip at Circling Squares Isabel Stengers and Bruno Latour think that this position should enjoin us to avoid ridiculing or undermining others’ values or ontologies. Further, that we should:

grant that all entities exist and, second, that to say that someone’s cherished idol (or whatever disputed entity they hold dear) is non-existent is a ‘declaration of war’ – ‘this means war,’ as Stengers often says.

I’ll admit that I find first part of this principle this damn puzzling. Even if we assume – for now – that it is wrong to attempt to undermine another person’s central beliefs this principle seems to require a) that people actually embrace ontological commitments that are contrary to the one’s they adhere to; b) pretend not to have one’s core beliefs; c) adopt some position of public neutrality vis a vis all core beliefs.

The first interpretation (a) results in the principle that one should embrace the contrary of every core belief; or, in effect, that no one should believe everything. So (in the interests of charity) we should pass on.

b) allows us to have beliefs so long as they are unexpressed. Depending on your view of beliefs, this is either incoherent (because there are no inexpressible beliefs) or burdens believers that no one is likely to find it acceptable.

So I take Philip to embrace c).  His clarification suggests something along these lines. For example. He claims that it is consistent with respecting otherness to say what we believe about other’s idols but not to publicly undermine their reasons for believing in them. Thus:

Their basic claim seems to be that ‘respect for otherness,’ i.e. political pluralism, can only come from granting the entities that others hold dear an ontology, even if you don’t ‘believe’ in them.  You are thus permitted to say ‘I do not follow that god, he has no hold over me’ but you are not permitted to say ‘your god is an inane, infantile, non-existent fantasy, grow up.’  And it’s not just a question of politeness (although there’s that too).  The point is to grant others’ idols and deities an existence – one needn’t agree over what that existence entails, over what capacities that entity has or what obligations it impresses upon you as someone in its partial presence but to deny it existence entirely is to ‘declare war’ – to deny the possibility of civil discourse, of pluralistic co-existence.

I must admit that I find this principle of respect puzzling as well. After all, some of my reasons for being an atheist are also reasons against being a theist. So unless this is just an innocuous plea for good manners (which I’m happy to sign up to on condition that notional others show me and mine the same forbearance) it seems to require that all believers keep their reasons for their belief to themselves. This, again, seems to demand an impossible or repugnant quietism.

So, thus far, ontological pluralism seems to be either incoherent or to impose such burdens on all believers that nobody should be required to observe it. There is, of course, a philosophical precedent for restricted ontological quietism in Rawls’ political liberalism. Rawls’ proposes that reasonable public deliberation recognize the “burdens of judgement” by omitted any justification that hinges on “comprehensive” ethical or religious doctrines over which there can be reasonable disagreement (Rawls 2005, 54). Deliberations about justice under Political Liberalism are thus constrained to be neutral towards “conflicting worldviews” so long as they are tolerant and reasonable (Habermas 1995, 119, 124-5).

However, there is an important difference between the political motivations behind Rawlsian public reason and the position of “ontological charity” Philip attributes to Stengers and Latour. Rawls’ is motivated by the need to preserve stability within plural democratic societies. Public reason does not apply outside the domain of political discourse in which reasonable citizens hash out basic principles of justice and constitutional essentials. It is also extremely problematic in itself.  Habermas  argues that Rawls exclusion of plural ethical or religious beliefs from the public court is self-vitiating because comprehensive perspectives are sources of disagreement about shared principles (for example, the legitimacy of abortion or same-sex marriage) and these must accordingly be addressed through dialogue rather than circumvented if a politically stable consensus is to be achieved (126).

Finally, apart from being incoherent, the principle of ontological charity seems unnecessary. As Levi Bryant points out in his realist retort to the pluralist, people are not the sum of their beliefs. Beliefs can be revised without effacing the believer. Thus an attack on core beliefs is not an attack on the person holding those beliefs.

So it is hard to interpret the claim that we should grant the existence of others’ “idols” as much more than the principle that it is wrong to humiliate, ridicule or insult people because of what their beliefs are. This seems like a good rule of thumb, but it is hard to justify the claim that it is an overriding principle. For example, even if  Rushdie’s Satanic Verses “insults Islam” having an open society in which aesthetic experimentation and the critical evaluation of ideas is possible is just more important than saving certain sections of it from cognitive dissonance or intellectual discomfort. Too many people have suffered death, terror and agony because others had aberrant and false core beliefs to make it plausible that these should be immune from criticism or ridicule. A little personal dissonance is a small price to pay for not going to the oven.

So what of the principle that we should build a “common world”. This is set out by Jeremy Trombley in his Struggle Forever blog under the rubric of “cosmopolitics”. Jeremy regards this project as an infinite task that requires us to seek a kind of fusion between different word views, phenomenologies and ontologies:

The project, as Latour, Stengers, James, and others have described it, is to compose a common world. What pluralism recognizes is that, in this project, we all start from different places – Latour’s relativity rather than relativism. The goal, then, (and it has to be recognized that this project is always contingent and prone to failure) is to make these different positions converge, but in a way that doesn’t impose one upon the other as the Modern Nature/Culture dichotomy tends to do. Why should we avoid imposing one on the other? In part because it’s the right thing to do – by imposing we remove or reduce the agency of the other. The claim to unmediated access to reality makes us invulnerable – no other claim has that grounding, and therefore we can never be wrong. But we are wrong – the science of the Enlightenment gave us climate change, environmental destruction, imperialism in the name of rationality (indigenous peoples removed from their land and taken to reeducation facilities where they were taught “rational” economic activities such as farming), and so on. It removed us from the world and placed us above it – the God’s eye view.

I think there a number of things wrong with cosmopolitics as Jeremy describes it here.

Firstly, seeking to alter beliefs or values does not necessarily reduce agency because people are not their beliefs.

Secondly, some worldviews – like the racist belief-systems that supported the European slave trade – just need to be imposed upon because they are bound up with violent and corrupting socio-political systems.

Thirdly, I know of no Enlightenment thinker, or realist, for whom “unmediated access to reality” is a sine qua non for knowledge. Let’s assume that “realism” is the contrary of pluralism here. It’s not clear what unmediated access would be like, but all realists are committed to the view that we we don’t have it since if we believe that reality has a mind-independent existence and nature, it can presumably vary independently of our beliefs about it. In its place, we have various doctrines of evidence and argument that are themselves susceptible to revision.  Some analyses of realism suppose that realists are committed to the claim that there is a one true account of the world (the God’s Eye View) but – as pointed out in an earlier post – this commitment is  debatable. In any case, supposing the the existence of a uniquely true theory is very different from claiming to have it.

Finally, much hinges on what we mean by a common world here. I take it that it is not the largely mind-independent reality assumed by the realist since – being largely mind-independent – it exists quite independently of any political project. So I take it that Jeremy is adverting something like a shared phenomenology or experience: a kind of fusion of horizons at the end of time. If we inflect “world” in this sense, then there is no reason for believing that such an aim is possible, let alone coherent. This possibility depends on there being structures of worldhood that are common to all beings that can be said to have one (Daseins, say). I’ve argued that there are no reasons for holding that we have access to such a priori knowledge because – like Scott Bakker - I hold that phenomenology gives us very limited insight into its nature. Thus we have no a priori grasp of what a world is and no reason to believe that Daseins (human or nonhuman) could ever participate in the same one. The argument for this is lengthy so I refer the reader to my paper “Nature’s Dark Domain” and my forthcoming book Posthuman Life.


Habermas, Jurgen. 1995. “Reconciliation through the Public Use of Reason: Remarks on John Rawls’s Political Liberalism.” The Journal of Philosophy 92 (3): 109–131.

Rawls, John. 2005. Political Liberalism. Columbia University Press.




Tagged with:

Putnam and Speculative Realism

On January 16, 2014, in Uncategorized, by enemyin1

Stephen Shakespeare has an interesting post over at An und für sich discussing Hilary Putnam’s argument against Metaphysical Realism and the positions of contemporary speculative realists like Meillassoux and Harman. Putnam (circa Reason Truth and History) treats Metaphysical Realism (MR) a package deal with three components: Independence (there is a fixed totality of mind-independent objects); Correspondence (there are word-world relations between bits of theories and the things to which they refer); Uniqueness (there is one true theory that correctly describes the state of these objects).

He then uses his model-theoretic argument to undermine Uniqueness. Given an epistemologically ideal theory and an interpretation function which maps that theory onto one of some totality of possible worlds, you can always come up with another mapping and hence another theory that is equally true of that world, elegant, simple, well-confirmed, etc. Unless, there is some other property that picks out a single theory as God’s Own other than its epistemic and semantic virtues, Uniqueness fails and with it MR.

Shakespeare argues that speculative realists reject the form of the independence thesis, denying that there is a fixed totality of mind-independent objects:

[Contemporary Realism] need not entail a conviction that objects in the world are a ‘fixed totality’. Objects can change or join to form new, irreducibly real objects. The lists of objects which are part of the rhetorical style of OOO encompass radically diverse things, including physical assemblages, social groups and fictional works. Each of these ‘objects’ consists of other irreducible objects and so on. There is not simply one stratum of object.

For Meillassoux, the picture is different. In one respect, the absolute consists of the fact that anything can be different for no reason: there is no founding ontological or transcendental necessity for the order of things. And this is what we can know. So his realism also does not entail that there is one fixed totality, or one complete and true description of things.

I demur partly from this analysis of where SR diverges from MR – though I’m happy to be persuaded otherwise. By “fixed” Putnam just means determinate. If there are fictional objects or sensa, then these must be part of God’s Own Theory (given MR). If there are assemblages with emergent properties, they too might click into God’s Own Ontology. Moreover, the Harmiverse has to consist of discrete, encodable objects, so it’s quite susceptible to a model-theoretic analysis of the kind that Putnam offers (See my Harman on Patterns and Harms).

Shakespeare may be right about Meillassoux’s ontology. One could argue that hyperchaos is not a thing and thus cannot be part of a model.

If we read Hyperchaos as the absolute contingency of any thinkable possibility then representing hyperchaos might seem pretty easy. Meillassoux is just saying that any non-contradictory event could occur (I will not consider whether he is justified in saying this).

So perhaps his ontology just comes down to the claim that any arbitrary, non-contradictory sentence is true in at least one possible world.

I suspect (but cannot show) that the real problem with reconciling Meillassoux’s SR with MR is in how one interprets this  modality. Saying that any arbitrary, non-contradictory sentence is true in at least possible world, is not what Meillassoux has in mind since this resembles a standard definition of de dicto contingency in possible world semantics. Moreover, Meillassoux (2010) denies we have warrant to believe that the thinkable can be totalized a priori on the grounds that set theory shows that there are always more things than can be contained in any totality. If this is right, then it is precipitate to assume a totality of all objects or a totality of all models under which God’s Own Theory could be interpreted. MR cannot even get started.

However, there are other ways in which contemporary realists (and not just speculative realists) could diverge from MR. For example, Devitt denies that realism is really committed to Uniqueness – the view that there is exactly “one true and complete description of the world” (Devitt 1984: 229). We might also demur from the assumption that the world consists of objects or only objects that enter into semantic relationships with bits of language or mind. Structural realists, for example, argue that reality is structure and that this is precisely what approximately similar theories capture – regardless of their official ontological divergences (Ladyman and Ross 2007: 94-5). Some speculative ontologies deny the Correspondence assumption, holding that the world contains entities that cannot be fully represented in any theory: e.g. powers, Deleuzean intensities.

Perhaps the Correspondence assumption just replicates the Kantian view that entities must conform to our modes of representation – in which case a robust realist should reject it in any case. This, interestingly, is where the issue of realism segues into the issues addressed in my forthcoming book Posthuman Life. For, analogously to Meillassoux’s claim about totalizing the thinkable, one can also reject the claim that we have any advance, future-proof knowledge of the forms in which reality must be “thought” If we have no access to the space of possible minds, then we can have no a priori conception of what a world must be as such.

Devitt, M. 1984. Realism and Truth. Princeton: Princeton University Press.

Meillassoux, Q. 2010. After Finitude: An Essay on the Necessity of Contingency, R. Brassier (trans). London: Continuum.

Putnam, H 1981. Reason, Truth and History, Cambridge University Press, Cambridge.

Ladyman James, Ross Don, (2007), Every Thing Must Go: Metaphysics Naturalized, Oxford: Oxford University Press.

Tagged with:

A highly illuminating discussion of the place of value, meaning and purpose within a naturalistic worldview. H/p synthetic zero

Invisible Clock: semi-algorithmic improvisation

On January 13, 2014, in Uncategorized, by enemyin1

Haven’t done this in a long while. I fired up my antique version of MAX MSP and Reaper and used my java based probabilistic sequencer jDelta (code here) to belt out this short improvisation. The sound is a marimba-like tuned percussion designed on the Native Instruments FM8 synthesizer. jDelta allows you to take a short seed sequence (here a repeated cluster chord) and graphically determine the probability of individual notes playing in the sequence, transpositions, velocity or tempo changes in real time using multislider objects. I then improvised a few ornaments over algorithmic variations induced on the seed phrase.



Relevance to posthuman performance practice: JDelta is just a sequencer that allows a certain global control over event probabilities. It assigns played note values to some arrays, then determines the probability of some output related to those values by imposing conditions on a random number output. A smarter program might (for example) use Bayesian statistics or neural networks rather than random numbers to fix the probabilities of an event relative to a given musical context (a little beyond my programming ability at the moment). While the program is not remotely smart, it mediates performance by allowing one to conceive the distribution of events in a graphical way, delegating how the events fall to the machine.

Tagged with: