Catherine Malabou has an intriguing piece on the vexed question of the relationship between the “humanities” and science in the journal Transeuropeennes here.
It is dominated by a clear and subtle reading of Kant, Foucault and Derrida’s discussion of the meaning of Enlightenment and modernity. Malabou argues that the latter thinkers attempt to escape Kantian assumptions about human invariance by identifying the humanities with “plasticity itself”. The Humanities need not style themselves in terms of some invariant essence of humanity. They can be understood as a site of transformation and “deconstruction” as such. Thus for Derrida in “University Without Condition”, the task of the humanities is:
the deconstruction of « what is proper to man » or to humanism. The transgression of the transcendental implies that the very notion of limit or frontier will proceed from a contingent, that is, historical, mutable, and changing deconstruction of the frontier of the « proper ».
Where, as for Foucault, the deconstruction of the human involves exhibiting its historical conditions of possibility and experimenting with these by, for example, thinking about “our ways of being, thinking, the relation to authority, relations between the sexes, the way in which we perceive insanity or illness “.
This analysis might suggest that the Humanities have little to fear from technological and scientific transformations of humans bodies or minds; they are just the setting in which the implications of these alterations are hammered out.
This line of thought reminds me of a revealingly bad argument produced by Andy Clark in his Natural Born Cyborgs:
The promise, or perhaps threatened, transition to a world of wired humans and semi-intelligent gadgets is just one more move in an ancient game . . . We are already masters at incorporating nonbiological stuff and structure deep into our physical and cognitive routines. To appreciate this is to cease to believe in any post-human future and to resist the temptation to define ourselves in brutal opposition to the very worlds in which so many of us now live, love and work (Clark 2003, 142).
This is obviously broken-backed: that earlier bootstrapping didn’t produce posthumans doesn’t entail that future ones won’t. Even if humans are essentially self-modifying it doesn’t follow that any prospective self-modifying entity is human.
The same problem afflicts Foucault and Derrida’s attempts to hollow out a reservation for humanities scholars by identifying them with the promulgation of transgression or deconstruction. Identifying the humanities with plasticity as such throws the portals of possibility so wide that it can only refer to an abstract possibility space whose contents and topology remains closed to us. If, with Malabou, we allow that some of these transgressions will operate on the material substrate of life, then we cannot assume that its future configurations will resemble human communities or human thinkers – thinkers concerned with topics like sex, work and death for example.
Malabou concludes with the suggestion that Foucault and Derrida fail to confront a quite different problem. They do not provide a historical explanation of the possibility of transformations of life and mind to which they refer:
They both speak of historical transformations of criticism without specifying them. I think that the event that made the plastic change of plasticity possible was for a major part the discovery of a still unheard of plasticity in the middle of the XXth century, and that has become visible and obvious only recently, i.e. the plasticity of the brain that worked in a way behind continental philosophy’s back. The transformation of the transcendental into a plastic material did not come from within the Humanities. It came precisely from the outside of the Humanities, with again, the notion of neural plasticity. I am not saying that the plasticity of the human as to be reduced to a series of neural patterns, nor that the future of the humanities consists in their becoming scientific, even if neuroscience tends to overpower the fields of human sciences (let’s think of neurolinguistics, neuropsychoanalysis, neuroaesthetics, or of neurophilosophy), I only say that the Humanities had not for the moment taken into account the fact that the brain is the only organ that grows, develops and maintains itself in changing itself, in transforming constantly its own structure and shape. We may evoke on that point a book by Norman Doidge, The Brain that changes itself. Doidge shows that this changing, self-fashioning organ is compelling us to elaborate new paradigms of transformation.
I’m happy to concede that the brain is a special case of biological plasticity, but, as Eileen Joy notes elsewhere, the suggestion that the humanities have been out of touch with scientific work on the brain is unmotivated. The engagement between the humanities (or philosophy, at least) and neuroscience already includes work as diverse as Paul and Patricia Churchland’s work on neurophilosophy and Derrida’s early writings on Freud’s Scientific Project.
I’m also puzzled by the suggestion that we need to preserve a place for transcendental thinking at all here. Our posthuman predicament consists in the realization that we are alterable configurations of matter and that our powers of self-alteration are changing in ways that put the future of human thought and communal life in doubt. This is not a transcendental claim. It’s a truistic generalisation which tells us little about the cosmic fate of an ill-assorted grab bag of academic disciplines.
Clark, A. 2003. Natural-born Cyborgs: Minds, Technologies, and the Future of Human Intelligence. New York: Oxford University Press.
People and cultures have some non-overlapping beliefs. Some folk believe that there is a God, some that there is no God, some that there are many gods. Some people believe that personal autonomy is a paramount value, while others feel that virtues like honour and courage take precedence over personal freedom. These core beliefs are serious, in that they make a difference to whether people live or die, or are able to live the kinds of life that they wish. People fight and die for the sake of autonomy. People fight, die or (institute gang rapes) in the interests of personal honour.
Some folk – the self-styled pluralists – believe that respect for otherness is a paramount political value. Respecting otherness, they say, is so paramount that it should regulate our ontological commitments – our assumptions about what exists. I must admit that I find this hard to credit ontologically or ethically. But it is also unclear how we should spell the principle out. So I’ll consider two versions that have circulated in the blogosphere recently. The first, I will argue, teeters on incoherence or, where not incoherent, is hard to justify in ethical or political terms. The second – which demands that we build a common world – may also be incoherent, but I will argue that we have no reason to think that its ultimate goal is realisable.
According to Philip at Circling Squares Isabel Stengers and Bruno Latour think that this position should enjoin us to avoid ridiculing or undermining others’ values or ontologies. Further, that we should:
grant that all entities exist and, second, that to say that someone’s cherished idol (or whatever disputed entity they hold dear) is non-existent is a ‘declaration of war’ – ‘this means war,’ as Stengers often says.
I’ll admit that I find first part of this principle this damn puzzling. Even if we assume – for now – that it is wrong to attempt to undermine another person’s central beliefs this principle seems to require a) that people actually embrace ontological commitments that are contrary to the one’s they adhere to; b) pretend not to have one’s core beliefs; c) adopt some position of public neutrality vis a vis all core beliefs.
The first interpretation (a) results in the principle that one should embrace the contrary of every core belief; or, in effect, that no one should believe everything. So (in the interests of charity) we should pass on.
b) allows us to have beliefs so long as they are unexpressed. Depending on your view of beliefs, this is either incoherent (because there are no inexpressible beliefs) or burdens believers that no one is likely to find it acceptable.
So I take Philip to embrace c). His clarification suggests something along these lines. For example. He claims that it is consistent with respecting otherness to say what we believe about other’s idols but not to publicly undermine their reasons for believing in them. Thus:
Their basic claim seems to be that ‘respect for otherness,’ i.e. political pluralism, can only come from granting the entities that others hold dear an ontology, even if you don’t ‘believe’ in them. You are thus permitted to say ‘I do not follow that god, he has no hold over me’ but you are not permitted to say ‘your god is an inane, infantile, non-existent fantasy, grow up.’ And it’s not just a question of politeness (although there’s that too). The point is to grant others’ idols and deities an existence – one needn’t agree over what that existence entails, over what capacities that entity has or what obligations it impresses upon you as someone in its partial presence but to deny it existence entirely is to ‘declare war’ – to deny the possibility of civil discourse, of pluralistic co-existence.
I must admit that I find this principle of respect puzzling as well. After all, some of my reasons for being an atheist are also reasons against being a theist. So unless this is just an innocuous plea for good manners (which I’m happy to sign up to on condition that notional others show me and mine the same forbearance) it seems to require that all believers keep their reasons for their belief to themselves. This, again, seems to demand an impossible or repugnant quietism.
So, thus far, ontological pluralism seems to be either incoherent or to impose such burdens on all believers that nobody should be required to observe it. There is, of course, a philosophical precedent for restricted ontological quietism in Rawls’ political liberalism. Rawls’ proposes that reasonable public deliberation recognize the “burdens of judgement” by omitted any justification that hinges on “comprehensive” ethical or religious doctrines over which there can be reasonable disagreement (Rawls 2005, 54). Deliberations about justice under Political Liberalism are thus constrained to be neutral towards “conflicting worldviews” so long as they are tolerant and reasonable (Habermas 1995, 119, 124-5).
However, there is an important difference between the political motivations behind Rawlsian public reason and the position of “ontological charity” Philip attributes to Stengers and Latour. Rawls’ is motivated by the need to preserve stability within plural democratic societies. Public reason does not apply outside the domain of political discourse in which reasonable citizens hash out basic principles of justice and constitutional essentials. It is also extremely problematic in itself. Habermas argues that Rawls exclusion of plural ethical or religious beliefs from the public court is self-vitiating because comprehensive perspectives are sources of disagreement about shared principles (for example, the legitimacy of abortion or same-sex marriage) and these must accordingly be addressed through dialogue rather than circumvented if a politically stable consensus is to be achieved (126).
Finally, apart from being incoherent, the principle of ontological charity seems unnecessary. As Levi Bryant points out in his realist retort to the pluralist, people are not the sum of their beliefs. Beliefs can be revised without effacing the believer. Thus an attack on core beliefs is not an attack on the person holding those beliefs.
So it is hard to interpret the claim that we should grant the existence of others’ “idols” as much more than the principle that it is wrong to humiliate, ridicule or insult people because of what their beliefs are. This seems like a good rule of thumb, but it is hard to justify the claim that it is an overriding principle. For example, even if Rushdie’s Satanic Verses “insults Islam” having an open society in which aesthetic experimentation and the critical evaluation of ideas is possible is just more important than saving certain sections of it from cognitive dissonance or intellectual discomfort. Too many people have suffered death, terror and agony because others had aberrant and false core beliefs to make it plausible that these should be immune from criticism or ridicule. A little personal dissonance is a small price to pay for not going to the oven.
So what of the principle that we should build a “common world”. This is set out by Jeremy Trombley in his Struggle Forever blog under the rubric of “cosmopolitics”. Jeremy regards this project as an infinite task that requires us to seek a kind of fusion between different word views, phenomenologies and ontologies:
The project, as Latour, Stengers, James, and others have described it, is to compose a common world. What pluralism recognizes is that, in this project, we all start from different places – Latour’s relativity rather than relativism. The goal, then, (and it has to be recognized that this project is always contingent and prone to failure) is to make these different positions converge, but in a way that doesn’t impose one upon the other as the Modern Nature/Culture dichotomy tends to do. Why should we avoid imposing one on the other? In part because it’s the right thing to do – by imposing we remove or reduce the agency of the other. The claim to unmediated access to reality makes us invulnerable – no other claim has that grounding, and therefore we can never be wrong. But we are wrong – the science of the Enlightenment gave us climate change, environmental destruction, imperialism in the name of rationality (indigenous peoples removed from their land and taken to reeducation facilities where they were taught “rational” economic activities such as farming), and so on. It removed us from the world and placed us above it – the God’s eye view.
I think there a number of things wrong with cosmopolitics as Jeremy describes it here.
Firstly, seeking to alter beliefs or values does not necessarily reduce agency because people are not their beliefs.
Secondly, some worldviews – like the racist belief-systems that supported the European slave trade – just need to be imposed upon because they are bound up with violent and corrupting socio-political systems.
Thirdly, I know of no Enlightenment thinker, or realist, for whom “unmediated access to reality” is a sine qua non for knowledge. Let’s assume that “realism” is the contrary of pluralism here. It’s not clear what unmediated access would be like, but all realists are committed to the view that we we don’t have it since if we believe that reality has a mind-independent existence and nature, it can presumably vary independently of our beliefs about it. In its place, we have various doctrines of evidence and argument that are themselves susceptible to revision. Some analyses of realism suppose that realists are committed to the claim that there is a one true account of the world (the God’s Eye View) but – as pointed out in an earlier post – this commitment is debatable. In any case, supposing the the existence of a uniquely true theory is very different from claiming to have it.
Finally, much hinges on what we mean by a common world here. I take it that it is not the largely mind-independent reality assumed by the realist since – being largely mind-independent – it exists quite independently of any political project. So I take it that Jeremy is adverting something like a shared phenomenology or experience: a kind of fusion of horizons at the end of time. If we inflect “world” in this sense, then there is no reason for believing that such an aim is possible, let alone coherent. This possibility depends on there being structures of worldhood that are common to all beings that can be said to have one (Daseins, say). I’ve argued that there are no reasons for holding that we have access to such a priori knowledge because – like Scott Bakker - I hold that phenomenology gives us very limited insight into its nature. Thus we have no a priori grasp of what a world is and no reason to believe that Daseins (human or nonhuman) could ever participate in the same one. The argument for this is lengthy so I refer the reader to my paper “Nature’s Dark Domain” and my forthcoming book Posthuman Life.
Habermas, Jurgen. 1995. “Reconciliation through the Public Use of Reason: Remarks on John Rawls’s Political Liberalism.” The Journal of Philosophy 92 (3): 109–131.
Rawls, John. 2005. Political Liberalism. Columbia University Press.
Stephen Shakespeare has an interesting post over at An und für sich discussing Hilary Putnam’s argument against Metaphysical Realism and the positions of contemporary speculative realists like Meillassoux and Harman. Putnam (circa Reason Truth and History) treats Metaphysical Realism (MR) a package deal with three components: Independence (there is a fixed totality of mind-independent objects); Correspondence (there are word-world relations between bits of theories and the things to which they refer); Uniqueness (there is one true theory that correctly describes the state of these objects).
He then uses his model-theoretic argument to undermine Uniqueness. Given an epistemologically ideal theory and an interpretation function which maps that theory onto one of some totality of possible worlds, you can always come up with another mapping and hence another theory that is equally true of that world, elegant, simple, well-confirmed, etc. Unless, there is some other property that picks out a single theory as God’s Own other than its epistemic and semantic virtues, Uniqueness fails and with it MR.
Shakespeare argues that speculative realists reject the form of the independence thesis, denying that there is a fixed totality of mind-independent objects:
[Contemporary Realism] need not entail a conviction that objects in the world are a ‘fixed totality’. Objects can change or join to form new, irreducibly real objects. The lists of objects which are part of the rhetorical style of OOO encompass radically diverse things, including physical assemblages, social groups and fictional works. Each of these ‘objects’ consists of other irreducible objects and so on. There is not simply one stratum of object.
For Meillassoux, the picture is different. In one respect, the absolute consists of the fact that anything can be different for no reason: there is no founding ontological or transcendental necessity for the order of things. And this is what we can know. So his realism also does not entail that there is one fixed totality, or one complete and true description of things.
I demur partly from this analysis of where SR diverges from MR – though I’m happy to be persuaded otherwise. By “fixed” Putnam just means determinate. If there are fictional objects or sensa, then these must be part of God’s Own Theory (given MR). If there are assemblages with emergent properties, they too might click into God’s Own Ontology. Moreover, the Harmiverse has to consist of discrete, encodable objects, so it’s quite susceptible to a model-theoretic analysis of the kind that Putnam offers (See my Harman on Patterns and Harms).
Shakespeare may be right about Meillassoux’s ontology. One could argue that hyperchaos is not a thing and thus cannot be part of a model.
If we read Hyperchaos as the absolute contingency of any thinkable possibility then representing hyperchaos might seem pretty easy. Meillassoux is just saying that any non-contradictory event could occur (I will not consider whether he is justified in saying this).
So perhaps his ontology just comes down to the claim that any arbitrary, non-contradictory sentence is true in at least one possible world.
I suspect (but cannot show) that the real problem with reconciling Meillassoux’s SR with MR is in how one interprets this modality. Saying that any arbitrary, non-contradictory sentence is true in at least possible world, is not what Meillassoux has in mind since this resembles a standard definition of de dicto contingency in possible world semantics. Moreover, Meillassoux (2010) denies we have warrant to believe that the thinkable can be totalized a priori on the grounds that set theory shows that there are always more things than can be contained in any totality. If this is right, then it is precipitate to assume a totality of all objects or a totality of all models under which God’s Own Theory could be interpreted. MR cannot even get started.
However, there are other ways in which contemporary realists (and not just speculative realists) could diverge from MR. For example, Devitt denies that realism is really committed to Uniqueness – the view that there is exactly “one true and complete description of the world” (Devitt 1984: 229). We might also demur from the assumption that the world consists of objects or only objects that enter into semantic relationships with bits of language or mind. Structural realists, for example, argue that reality is structure and that this is precisely what approximately similar theories capture – regardless of their official ontological divergences (Ladyman and Ross 2007: 94-5). Some speculative ontologies deny the Correspondence assumption, holding that the world contains entities that cannot be fully represented in any theory: e.g. powers, Deleuzean intensities.
Perhaps the Correspondence assumption just replicates the Kantian view that entities must conform to our modes of representation – in which case a robust realist should reject it in any case. This, interestingly, is where the issue of realism segues into the issues addressed in my forthcoming book Posthuman Life. For, analogously to Meillassoux’s claim about totalizing the thinkable, one can also reject the claim that we have any advance, future-proof knowledge of the forms in which reality must be “thought” If we have no access to the space of possible minds, then we can have no a priori conception of what a world must be as such.
Devitt, M. 1984. Realism and Truth. Princeton: Princeton University Press.
Meillassoux, Q. 2010. After Finitude: An Essay on the Necessity of Contingency, R. Brassier (trans). London: Continuum.
Putnam, H 1981. Reason, Truth and History, Cambridge University Press, Cambridge.
Ladyman James, Ross Don, (2007), Every Thing Must Go: Metaphysics Naturalized, Oxford: Oxford University Press.
Haven’t done this in a long while. I fired up my antique version of MAX MSP and Reaper and used my java based probabilistic sequencer jDelta (code here) to belt out this short improvisation. The sound is a marimba-like tuned percussion designed on the Native Instruments FM8 synthesizer. jDelta allows you to take a short seed sequence (here a repeated cluster chord) and graphically determine the probability of individual notes playing in the sequence, transpositions, velocity or tempo changes in real time using multislider objects. I then improvised a few ornaments over algorithmic variations induced on the seed phrase.
Relevance to posthuman performance practice: JDelta is just a sequencer that allows a certain global control over event probabilities. It assigns played note values to some arrays, then determines the probability of some output related to those values by imposing conditions on a random number output. A smarter program might (for example) use Bayesian statistics or neural networks rather than random numbers to fix the probabilities of an event relative to a given musical context (a little beyond my programming ability at the moment). While the program is not remotely smart, it mediates performance by allowing one to conceive the distribution of events in a graphical way, delegating how the events fall to the machine.
Radical art defies and transforms collective modes of understanding. Wagner’s famous “Tristan chord” segues between classical harmony, late romanticism and twentieth century atonality due to its ambiguous relationship to its tonal context. The aesthetic value of Xenakis’ Concret Ph lies partly in the technological potentials realized subsequently in granular synthesis techniques which employ global statistical parameters to control flocks of auditory events. Such sensations are, in Brian Massumi words, “in excess over experience” – suspending practices and meanings in ways that catalyse deterritorializing movement towards non-actual futures (Massumi 2005: 136). The aesthetics of excess provides a limit case of the reflective creation of value that occurs when we modify existing modes of sense-making or embodiment. It also provides a window upon the posthuman as potentiality shadowing our interactions with technological environments.
This contingency is amplified in another radical art work, J G Ballard’s novel Crash. As I wrote back in 1999:
In Crash the technology of the car has become the adjunct to a violent sexuality. Its erotic focus and ideologue, Vaughan, is an ambulance chasing ex-TV presenter whose career as a glamorous ‘hoodlum scientist’ has been cut short by his disfigurement in a motorcycling accident. Marking the parameters of vehicle collisions and casual sexual encounters with Polaroid and cine camera, Vaughan is a social being of sorts, assembling around him a crew of co-experimenters whose sexuality has been activated by ‘the perverse eroticisms of the car-crash’. The novel’s narrator ‘James Ballard’ recounts his induction into the crashpack; first through a motorway accident, then via a succession of techno-erotic duels and excursions, culminating in Vaughan’s attempted ‘seduction’ of the actress Elizabeth Taylor in the environs of London Airport . .
It is only in so far as Vaughan ‘[mimes] the equations between the styling of a motor-car and the organic elements of his body’ (Ballard 1995: 170), modulating the symbolic requirements of Ballard’s narrative with his histrionic body, that he can remain its primary sexual focus. . . These impersonal ‘equations’ mediate every affective relationship between the characters and Crash’s residual city of multi-storey car parks, airport termini, hermetic suburbs and motorway slip roads. They are expressed in a language of excremental objects – ‘aluminium ribbons’, Gabrielle’s thigh wound, Vaughan’s sectioned nipples, torn fenders, scars, etc. – whose very lack of quotidian function commends them as arbitrary tokens in the symbolic algebra (Roden 2002).
Crash thus construct an internally referential system of desire around sites, surfaces and interstices of late twentieth century technological landscapes (Roden 2002). But despite its contemporary setting, the novel does not describe this world: it potentiates it. Crash exhibits the contingency of human subjectivity and social relationships given its irrevocably technological condition.
A similar claim is made about the work of the Australian performance artist Stelarc in Massumi’s “The Evolutionary Alchemy of Reason”. Massumi argues that the content of Stelarc’s performances – such as his series of body suspensions or his hook-ups with industrial robots, prosthetic hands and compound-eye goggles – is nothing to do with the functional utility of these systems or events. They have no use. Rather their effect is to place bodies and technologies in settings where their incorporation as use-values is interrupted. Of the compound eye goggles that Stelarc created for his work Helmet no. 3: put on and walk 1970 he writes: “They extended no-need into no-utility. And they extended no-utility into ‘art’” (Massumi 2005: 131).
Stelarc’s somewhat elliptical rationale is to “extend intelligence beyond the Earth”. His performances decouple the body from its functions and from the empathic responses of observers – even when dangling from skin hooks over a city street, Stelarc never appears as suffering or abject. They register the body’s potential for “off world” environments rather than its actual functional involvements with our technological landscape. Space colonization is not a current use value or industrial application, but a project for our planned obsolesce:
The terrestrial body will be obsolete from the moment a certain subpopulation feels compelled to launch itself into an impossible, unthinkable future of space colonization. To say that the obsolescence of the body is produced is to say that it is compelled. To say that it is compelled is to say that it is “driven by desire” rather than by need or utility (151-2).
These performances embody a potential that is “unthinkable” because aesthetically disjoined from our phenomenology and world. But, as Claire Colebrook suggests, we have been incipiently “off world” since the dawn of the industrial era:
We have perhaps always lived in a time of divergent, disrupted and diffuse systems of forces, in which the role of human decisions and perceptions is a contributing factor at best. Far from being resolved by returning to the figure of the bounded globe or subject of bios rather than zoe, all those features that one might wish to criticize in the bio-political global era can only be confronted by a non-global temporality and counter-ethics (Colebrook 2012: 38).
The counter-final nature of modern technique means that the conditions under which human ethical judgements are adapted can be overwritten by systems over which we have no ultimate control. Posthumanity would be only the most extreme consequence of this ramifying technics. An ethics bounded by the human world thus ignores its already excessive character (32).
Ballard, J.G. 1995. Crash. London: Vintage.
Massumi, Brian. 2005. “The Evolutionary Alchemy of Reason: Stelarc.” In Stelarc: The Monograph, Marquand Smith (ed). MIT Press: 125-192.
Roden, DAvid 2003. “Cyborgian Subjects and the Auto-destruction of Metaphor.” In Crash Cultures: Modernity, Mediation and the Material, Jane Arthurs and Iain Grant (eds). Intellect Books: 91–102.
Colebrook, Claire 2012. “A Globe of One’s Own: In Praise of the Flat Earth.” Substance: A Review of Theory & Literary Criticism 41 (1): 30–39.
This is a sketch of a partial value theory that I’ve been developing while completing my book Posthuman Life. If there are similar theories out there, I’d be grateful for links to bibdata so that I can properly acknowledge them!
In order to construct an anthropologically unbounded account of posthumans, we need a psychology-free account of value. There may, after all, be many possible posthuman psychologies but we don’t know about any of them to date. However, the theory requires posthumans to be autonomous systems of a special kind: Functionally Autonomous Systems (see below). I understand ”autonomy” here as a biological capacity for active self maintenance. The idea of a system which intervenes in the boundary conditions required for its existence can be used to formulate an Autonomous Systems Account of function which avoids some of the metaphysical problems associated with the more standard etiological theory. The version of ASA developed by Wayne Christensen and Mark Bickhard defines the functions of an entity in terms of its contribution to the persistence of an autonomous system, which they conceive as a group of interdependent processes (Christensen and Bickhard 2002: 3). Functions are process dependence relations within actively self-maintaining systems.
Ecological values are constituted by functions. The conception, in turn, allows us to formulate an account of “enlistment” which then allows us to define what it is to be an FAS.
1) (ASA) Each autonomous system has functions belonging to it at some point in its history. Its functions are the interdependent processes it requires to remain autonomous at that point.
2) (Value) If a process, thing or state is required for a function to occur, then that thing or process is a value for that function. Any entity, state or resource can be a value. For example, the proper functioning of a function can be a value for the functions that require it to work.
3) (Enlistment) When an autonomous system produces a function, then any value of that function is enlisted by that system.
4) (Accrual) An FAS actively accrues functions by producing functions that are also values for other FAS’s.
5) (Functional Autonomy) A functionally autonomous system (FAS) is any autonomous system that can enlist values and accrue functions.
People are presumably FAS’s on this account, but also nonhuman organisms and (perhaps) lineages of organisms. Likewise, social systems (Collier and Hooker 2009) and (conceivably) posthumans. To date, technical entities are not FAS’s because they are non-autonomous. Historical technologies are mechanisms of enlistment, however. For example. Without mining technology, certain ores would not be values for human activities. Social entities, such as corporations, are autonomous in the relevant and sense and thus can have functions (process interdependency relations) and constitute values of their own. However, while not-narrowly human, current social systems are wide humans not posthumans. As per the Disconnection Thesis: Posthumans would be FAS’s no longer belonging to WH (the Wide Human socio-technical assemblage – See Roden 2012).
This is an ecological account in the strict sense of specifying values in terms of environmental relations between functions and their prerequisites (though “environment” should be interpreted broadly to include endogenous and well as exogenous entities or states). It is also an objective rather than subjective account which has no truck with the spirit (meaning, culture or subjectivity, etc.). Value are just things which enter into constitutive relations with functions (Definition 2 could be expanded and qualified by introducing degrees of dependency). Oxygen was an ecological value for aerobic organisms long before Lavoisier. We can be ignorant of our values and mistake non-values for values, etc. It is also arguable that some ecological values are pathological in that they support some functions while hindering others.
The theory is partial because it only provides a sufficient condition for value. Some values – Opera, cigarettes, incest prohibitions and sunsets – are arguably things of the spirit, constituted as values by desires or cultural meanings.
Christensen, W. D., and M. H. Bickhard. 2002. “The Process Dynamics of Normative Function.” The Monist 85 (1): 3–28.
Collier, J. D., & Hooker, C. A. 1999. Complexly organised dynamical systems. Open Systems & Information Dynamics, 6(3): 241-302.
Roden. 2012. “The Disconnection Thesis.” The Singularity Hypothesis: A Scientific and Philosophical Assessment, Edited by Amnon Eden, Johnny Søraker, Jim Moor, and Eric Steinhart.Springer Frontiers Collection.
 An issue I do not have time to consider is that ecological dependency is transitive. If a function depends on a thing whose exist depends on another thing, then it depends on that other thing. Ecological dependencies thus overlap.
 Addictive substances may fall into this class.
Filippo Bertoni and Uli Beisel have written an interesting and intriguing essay on the ethical challenges posed by extending person-status to cetaceans: “More-than-human intelligence: of dolphins, Indian law and the multispecies turn” over at Society and Space. I only wish the terms in which their discussion is framed had been clearer. For example, they provide no working conception of personhood – though there are many philosophical candidates – and their passing allusion to sharing “somatic sensibilities” is similarly vague. This is a pity because the philosophical and ethical problem they raise is genuine. Should we regard human-style personhood as the tip of an cosmic moral hierarchy from which the only way is down? If not, then how do we negotiate relationships between beings with lives that are phenomenologically different but perhaps not less valuable than humans?