{"id":214808,"date":"2017-03-10T07:50:34","date_gmt":"2017-03-10T12:50:34","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/functionalism-philosophy-of-mind-wikipedia.php"},"modified":"2017-03-10T07:50:34","modified_gmt":"2017-03-10T12:50:34","slug":"functionalism-philosophy-of-mind-wikipedia","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/mind-upload\/functionalism-philosophy-of-mind-wikipedia.php","title":{"rendered":"Functionalism (philosophy of mind) &#8211; Wikipedia"},"content":{"rendered":"<p><p>    Functionalism is a theory of the mind in contemporary    philosophy,    developed largely as an alternative to both the identity theory of mind and    behaviorism. Its core idea is that mental    states (beliefs, desires, being in pain, etc.) are constituted    solely by their functional role  that is, they have causal    relations to other mental states, numerous sensory inputs, and    behavioral outputs.[1]    Functionalism is a theoretical level between the physical    implementation and behavioral output.[2] Therefore, it is    different from its predecessors of Cartesian dualism (advocating    independent mental and physical substances) and Skinnerian    behaviorism    and physicalism (declaring only physical    substances) because it is only concerned with the effective    functions of the brain, through its organization or its    \"software programs\".  <\/p>\n<p>    Since mental states are identified by a functional role, they    are said to be realized on multiple levels; in other words,    they are able to be manifested in various systems, even perhaps    computers, so long as the system performs the appropriate    functions. While computers are physical devices with electronic    substrate that perform computations on inputs to give outputs,    so brains are physical devices with neural substrate that    perform computations on inputs which produce behaviors.  <\/p>\n<p>    An important part of some accounts of functionalism is the idea    of multiple realizability. Since,    according to standard functionalist theories, mental states are    the corresponding functional role, mental states can be    sufficiently explained without taking into account the    underlying physical medium (e.g. the brain, neurons, etc.) that    realizes such states; one need only take into account the    higher-level functions in the cognitive system. Since mental    states are not limited to a particular medium, they can be    realized in multiple ways, including, theoretically, within    non-biological systems, such as computers. In other words, a    silicon-based machine could, in principle, have the same sort    of mental life that a human being has, provided that its    cognitive system realized the proper functional roles. Thus,    mental states are individuated much like a valve; a valve can    be made of plastic or metal or whatever material, as long as it    performs the proper function (say, controlling the flow of    liquid through a tube by blocking and unblocking its pathway).  <\/p>\n<p>    However, there have been some functionalist theories that    combine with the identity theory of mind, which deny multiple    realizability. Such Functional Specification Theories    (FSTs) (Levin,  3.4), as they are called, were most notably    developed by David    Lewis[3] and    David Malet Armstrong.[4] According to FSTs,    mental states are the particular \"realizers\" of the functional    role, not the functional role itself. The mental state of    belief, for example, just is whatever brain or neurological    process that realizes the appropriate belief function. Thus,    unlike standard versions of functionalism (often called    Functional State Identity Theories), FSTs do not allow    for the multiple realizability of mental states, because the    fact that mental states are realized by brain states is    essential. What often drives this view is the belief that if we    were to encounter an alien race with a cognitive system    composed of significantly different material from humans'    (e.g., silicon-based) but performed the same functions as human    mental states (e.g., they tend to yell \"Yowzas!\" when poked    with sharp objects, etc.) then we would say that their type of    mental state is perhaps similar to ours, but too different to    say it's the same. For some, this may be a disadvantage to    FSTs. Indeed, one of Hilary Putnam's[5][6] arguments for his    version of functionalism relied on the intuition that such    alien creatures would have the same mental states as humans do,    and that the multiple realizability of standard functionalism    makes it a better theory of mind.  <\/p>\n<p>    The broad position of \"functionalism\" can be articulated in    many different varieties. The first formulation of a    functionalist theory of mind was put forth by Hilary    Putnam.[5][6] This formulation, which    is now called machine-state functionalism, or just    machine functionalism, was inspired by the analogies    which Putnam and others noted between the mind and the theoretical \"machines\" or    computers capable of computing any given algorithm which were    developed by Alan Turing (called Turing    machines).  <\/p>\n<p>    In non-technical terms, a Turing machine can be visualized as    an indefinitely and infinitely long tape divided into    rectangles (the memory) with a box-shaped scanning device that    sits over and scans one component of the memory at a time. Each    unit is either blank (B) or has a 1 written on    it. These are the inputs to the machine. The possible outputs    are:  <\/p>\n<p>    An extremely simple example of a Turing machine which writes    out the sequence '111' after scanning three blank squares and    then stops as specified by the following machine table:  <\/p>\n<p>    This table states that if the machine is in state one and scans    a blank square (B), it will print a 1 and remain    in state one. If it is in state one and reads a 1, it    will move one square to the right and also go into state two.    If it is in state two and reads a B, it will print a    1 and stay in state two. If it is in state two and reads    a 1, it will move one square to the right and go into    state three. If it is in state three and reads a B, it    prints a 1 and remains in state three. Finally, if it is    in state three and reads a 1, then it will stay in state    three.  <\/p>\n<p>    The essential point to consider here is the nature of the    states of the Turing machine. Each state can be defined    exclusively in terms of its relations to the other states as    well as inputs and outputs. State one, for example, is simply    the state in which the machine, if it reads a B, writes    a 1 and stays in that state, and in which, if it reads a    1, it moves one square to the right and goes into a    different state. This is the functional definition of state    one; it is its causal role in the overall system. The details    of how it accomplishes what it accomplishes and of its material    constitution are completely irrelevant.  <\/p>\n<p>    According to machine-state functionalism, the nature of a    mental state is just like the nature of the automaton states    described above. Just as state one simply is the state    in which, given an input B, such and such happens, so    being in pain is the state which disposes one to cry \"ouch\",    become distracted, wonder what the cause is, and so forth.  <\/p>\n<p>    A second form of functionalism is based on the rejection of    behaviorist    theories in psychology and their replacement with empirical    cognitive models of the mind. This view is most closely    associated with Jerry Fodor and Zenon    Pylyshyn and has been labeled psychofunctionalism.  <\/p>\n<p>    The fundamental idea of psychofunctionalism is that psychology    is an irreducibly complex science and that the terms that we    use to describe the entities and properties of the mind in our    best psychological theories cannot be redefined in terms of    simple behavioral dispositions, and further, that such a    redefinition would not be desirable or salient were it    achievable. Psychofunctionalists view psychology as employing    the same sorts of irreducibly teleological or purposive explanations as the    biological sciences. Thus, for example, the function or role of    the heart is to pump blood, that of the kidney is to filter it    and to maintain certain chemical balances and so onthis is    what accounts for the purposes of scientific explanation and    taxonomy. There may be an infinite variety of physical    realizations for all of the mechanisms, but what is important    is only their role in the overall biological theory. In an    analogous manner, the role of mental states, such as belief and    desire, is determined by the functional or causal role that is    designated for them within our best scientific    psychological theory. If some mental state which is postulated    by folk    psychology (e.g. hysteria) is determined not to have any    fundamental role in cognitive psychological explanation, then    that particular state may be considered not to exist . On the    other hand, if it turns out that there are states which    theoretical cognitive psychology posits as necessary for    explanation of human behavior but which are not foreseen by    ordinary folk psychological language, then these entities or    states exist.  <\/p>\n<p>    A third form of functionalism is concerned with the meanings of    theoretical terms in general. This view is most closely    associated with David Lewis and    is often referred to as analytic functionalism or    conceptual functionalism. The basic idea of analytic    functionalism is that theoretical terms are implicitly defined    by the theories in whose formulation they occur and not by    intrinsic properties of the phonemes they comprise. In the case    of ordinary language terms, such as \"belief\", \"desire\", or    \"hunger\", the idea is that such terms get their meanings from    our common-sense \"folk psychological\" theories about them, but    that such conceptualizations are not sufficient to withstand    the rigor imposed by materialistic theories of reality and    causality. Such terms are subject to conceptual analyses which    take something like the following form:  <\/p>\n<p>    For example, the state of pain is caused by    sitting on a tack and causes loud cries, and higher    order mental states of anger and resentment directed at the    careless person who left a tack lying around. These sorts of    functional definitions in terms of causal roles are claimed to    be analytic and a priori truths about the    submental states and the (largely fictitious) propositional    attitudes they describe. Hence, its proponents are known as    analytic or conceptual functionalists. The    essential difference between analytic and psychofunctionalism    is that the latter emphasizes the importance of laboratory    observation and experimentation in the determination of which    mental state terms and concepts are genuine and which    functional identifications may be considered to be genuinely    contingent and a    posteriori identities. The former, on the other hand,    claims that such identities are necessary and not subject to    empirical scientific investigation.  <\/p>\n<p>    Homuncular functionalism was developed largely by    Daniel    Dennett and has been advocated by William Lycan.    It arose in response to the challenges that Ned Block's China Brain (a.k.a. Chinese nation) and    John    Searle's Chinese room thought experiments presented    for the more traditional forms of functionalism (see below    under \"Criticism\"). In attempting to overcome the conceptual    difficulties that arose from the idea of a nation full of    Chinese people wired together, each person working as a single    neuron to produce in the wired-together whole the functional    mental states of an individual mind, many functionalists simply    bit the bullet, so to speak, and argued that such a Chinese    nation would indeed possess all of the qualitative and    intentional properties of a mind; i.e. it would become a sort    of systemic or collective mind with propositional attitudes and    other mental characteristics. Whatever the worth of this latter    hypothesis, it was immediately objected that it entailed an unacceptable sort of    mind-mind supervenience: the systemic mind which somehow    emerged at the higher-level must necessarily supervene on the    individual minds of each individual member of the Chinese    nation, to stick to Block's formulation. But this would seem to    put into serious doubt, if not directly contradict, the    fundamental idea of the supervenience thesis: there can be no    change in the mental realm without some change in the    underlying physical substratum. This can be easily seen if we    label the set of mental facts that occur at the higher-level    M1 and the set of mental facts that occur at the    lower-level M2. Given the transitivity of supervenience,    if M1 supervenes on M2, and M2 supervenes    on P (physical base), then M1 and M2 both    supervene on P, even though they are (allegedly) totally    different sets of mental facts.  <\/p>\n<p>    Since mind-mind supervenience seemed to have become acceptable    in functionalist circles, it seemed to some that the only way    to resolve the puzzle was to postulate the existence of an    entire hierarchical series of mind levels (analogous to    homunculi) which became less and less    sophisticated in terms of functional organization and physical    composition all the way down to the level of the    physico-mechanical neuron or group of neurons. The homunculi at    each level, on this view, have authentic mental properties but    become simpler and less intelligent as one works one's way down    the hierarchy.  <\/p>\n<p>    Mechanistic functionalism, originally formulated and defended    by Gualtiero Piccinini[7] and Carl    Gillett[8][9] independently,    augments previous functionalist accounts of mental states by    maintaining that any psychological explanation must be rendered    in mechanistic terms. That is, instead of mental states    receiving a purely functional explanation in terms of their    relations to other mental states, like those listed above,    functions are seen as playing only a partthe other part being    played by structures of the explanation of a given mental    state.  <\/p>\n<p>    A mechanistic explanation[10] involves    decomposing a given system, in this case a mental system, into    its component physical parts, their activities or functions,    and their combined organizational relations.[7] On this    account the mind remains a functional system, but one that is    understood mechanistically. This account remains a sort of    functionalism because functional relations are still essential    to mental states, but it is mechanistic    because the functional relations are always manifestations of    concrete structuresalbeit structures understood at a certain    level of abstraction. Functions are individuated and explained    either in terms of the contributions they make to the given    system[11] or in teleological terms. If the functions are    understood in teleological terms, then they may be    characterized either etiologically or    non-etiologically.[12]  <\/p>\n<p>    Mechanistic functionalism leads functionalism away from the    traditional functionalist autonomy of psychology from    neuroscience and towards integrating psychology and    neuroscience.[13] By providing an applicable    framework for merging traditional psychological models with    neurological data, mechanistic functionalism may be understood    as reconciling the functionalist theory of mind with    neurological accounts of how the brain actually works. This is    due to the fact that mechanistic explanations of function    attempt to provide an account of how functional states (mental    states) are physically realized through neurological    mechanisms.  <\/p>\n<p>    There is much confusion about the sort of relationship that is    claimed to exist (or not exist) between the general thesis of    functionalism and physicalism. It has often been claimed that    functionalism somehow \"disproves\" or falsifies physicalism    tout court (i.e. without further    explanation or description). On the other hand, most    philosophers of mind who are functionalists claim to be    physicalistsindeed, some of them, such as David Lewis, have    claimed to be strict reductionist-type physicalists.  <\/p>\n<p>    Functionalism is fundamentally what Ned Block has called a    broadly metaphysical thesis as opposed to a narrowly ontological one. That is, functionalism is    not so much concerned with what there is than with what    it is that characterizes a certain type of mental state, e.g.    pain, as the type of state that it is. Previous attempts to    answer the mind-body problem have all tried to resolve it by    answering both questions: dualism says there are two    substances and that mental states are characterized by their    immateriality; behaviorism claimed that there was one substance    and that mental states were behavioral disposition; physicalism    asserted the existence of just one substance and characterized    the mental states as physical states (as in \"pain = C-fiber    firings\").  <\/p>\n<p>    On this understanding, type physicalism can be seen as    incompatible with functionalism, since it claims that what    characterizes mental states (e.g. pain) is that they are    physical in nature, while functionalism says that what    characterizes pain is its functional\/causal role and its    relationship with yelling \"ouch\", etc. However, any weaker sort    of physicalism which makes the simple ontological claim that    everything that exists is made up of physical matter is    perfectly compatible with functionalism. Moreover, most    functionalists who are physicalists require that the properties    that are quantified over in functional definitions be physical    properties. Hence, they are physicalists, even though    the general thesis of functionalism itself does not commit them    to being so.  <\/p>\n<p>    In the case of David Lewis, there is a distinction in the    concepts of \"having pain\" (a rigid designator true of the same    things in all possible worlds) and just \"pain\" (a non-rigid    designator). Pain, for Lewis, stands for something like the    definite description \"the state with the causal role x\". The    referent of the description in humans is a type of brain state    to be determined by science. The referent among silicon-based    life forms is something else. The referent of the description    among angels is some immaterial, non-physical state. For Lewis,    therefore, local type-physical reductions are possible    and compatible with conceptual functionalism. (See also Lewis's    mad pain and Martian pain.)    There seems to be some confusion between types and tokens that    needs to be cleared up in the functionalist analysis.  <\/p>\n<p>    Ned Block[14]    argues against the functionalist proposal of multiple realizability, where    hardware implementation is irrelevant because only the    functional level is important. The \"China brain\" or \"Chinese    nation\" thought experiment involves supposing that the entire    nation of China systematically organizes itself to operate just    like a brain, with each individual acting as a neuron.    According to functionalism, so long as the people are    performing the proper functional roles, with the proper causal    relations between inputs and outputs, the system will be a real    mind, with mental states, consciousness, and so on. However,    Block argues, this is patently absurd, so there must be    something wrong with the thesis of functionalism since it would    allow this to be a legitimate description of a mind.  <\/p>\n<p>    Some functionalists believe China would have qualia but that    due to the size it is impossible to imagine China being    conscious.[15]    Indeed, it may be the case that we are constrained by our    theory of    mind[16] and    will never be able to understand what Chinese-nation    consciousness is like. Therefore, if functionalism is true    either qualia will exist across all hardware or will not exist    at all but are illusory.[17]  <\/p>\n<p>    The Chinese    room argument by John Searle[18] is a direct attack on    the claim that thought can be represented as a set of    functions. The thought experiment asserts that it is possible    to mimic intelligent action without any interpretation or    understanding through the use of a purely functional system. In    short, Searle describes a person who only speaks English who is    in a room with only Chinese symbols in baskets and a rule book    in English for moving the symbols around. The person is then    ordered by people outside of the room to follow the rule book    for sending certain symbols out of the room when given certain    symbols. Further suppose that the people outside of the room    are Chinese speakers and are communicating with the person    inside via the Chinese symbols. According to Searle, it would    be absurd to claim that the English speaker inside knows    Chinese simply based on these syntactic processes. This thought    experiment attempts to show that systems which operate merely    on syntactic processes (inputs and outputs, based on    algorithms) cannot realize any semantics (meaning) or    intentionality (aboutness). Thus, Searle attacks the idea that    thought can be equated with following a set of syntactic rules;    that is, functionalism is an insufficient theory of the mind.  <\/p>\n<p>    As noted above, in connection with Block's Chinese nation, many    functionalists responded to Searle's thought    experiment by suggesting that there was a form of mental    activity going on at a higher level than the man in the Chinese    room could comprehend (the so-called \"system reply\"); that is,    the system does know Chinese. Of course, Searle responds that    there is nothing more than syntax going on at the higher-level    as well, so this reply is subject to the same initial problems.    Furthermore, Searle suggests the man in the room could simply    memorize the rules and symbol relations. Again, though he would    convincingly mimic communication, he would be aware only of the    symbols and rules, not of the meaning behind them.  <\/p>\n<p>    Another main criticism of functionalism is the inverted    spectrum or inverted qualia scenario, most specifically proposed as an    objection to functionalism by Ned Block.[14][19] This thought experiment    involves supposing that there is a person, call her Jane, that    is born with a condition which makes her see the opposite    spectrum of light that is normally perceived. Unlike normal    people, Jane sees the color violet as yellow, orange as blue,    and so forth. So, suppose, for example, that you and Jane are    looking at the same orange. While you perceive the fruit as    colored orange, Jane sees it as colored blue. However, when    asked what color the piece of fruit is, both you and Jane will    report \"orange\". In fact, one can see that all of your    behavioral as well as functional relations to colors will be    the same. Jane will, for example, properly obey traffic signs    just as any other person would, even though this involves the    color perception. Therefore, the argument goes, since there can    be two people who are functionally identical, yet have    different mental states (differing in their qualitative or    phenomenological aspects), functionalism is not robust enough    to explain individual differences in qualia.[20]  <\/p>\n<p>    David    Chalmers tries to show[21] that even though    mental content cannot be fully accounted for in functional    terms, there is nevertheless a nomological correlation    between mental states and functional states in this world. A    silicon-based robot, for example, whose functional profile    matched our own, would have to be fully conscious. His    argument for this claim takes the form of a reductio ad absurdum. The    general idea is that since it would be very unlikely for a    conscious human being to experience a change in its qualia    which it utterly fails to notice, mental content and functional    profile appear to be inextricably bound together, at least in    the human case. If the subject's qualia were to change, we    would expect the subject to notice, and therefore his    functional profile to follow suit. A similar argument is    applied to the notion of absent qualia. In this case,    Chalmers argues that it would be very unlikely for a subject to    experience a fading of his qualia which he fails to notice and    respond to. This, coupled with the independent assertion that a    conscious being's functional profile just could be maintained,    irrespective of its experiential state, leads to the conclusion    that the subject of these experiments would remain fully    conscious. The problem with this argument, however, as Brian G.    Crabb (2005) has observed, is that it begs the central    question: How could Chalmers know that functional    profile can be preserved, for example while the conscious    subject's brain is being supplanted with a silicon substitute,    unless he already assumes that the subject's possibly changing    qualia would not be a determining factor? And while changing or    fading qualia in a conscious subject might force changes in its    functional profile, this tells us nothing about the case of a    permanently inverted or unconscious robot. A subject with    inverted qualia from birth would have nothing to notice or    adjust to. Similarly, an unconscious functional simulacrum of    ourselves (a zombie) would have no experiential changes to    notice or adjust to. Consequently, Crabb argues, Chalmers'    \"fading qualia\" and \"dancing qualia\" arguments fail to    establish that cases of permanently inverted or absent qualia    are nomologically impossible.  <\/p>\n<p>    A related critique of the inverted spectrum argument is that it    assumes that mental states (differing in their qualitative or    phenomenological aspects) can be independent of the functional    relations in the brain. Thus, it begs the question of functional mental    states: its assumption denies the possibility of functionalism    itself, without offering any independent justification for    doing so. (Functionalism says that mental states are produced    by the functional relations in the brain.) This same type of    problemthat there is no argument, just an antithetical    assumption at their basecan also be said of both the Chinese    room and the Chinese nation arguments. Notice, however, that    Crabb's response to Chalmers does not commit this fallacy: His    point is the more restricted observation that even if    inverted or absent qualia turn out to be nomologically    impossible, and it is perfectly possible that we might    subsequently discover this fact by other means, Chalmers'    argument fails to demonstrate that they are impossible.  <\/p>\n<p>    The Twin Earth thought    experiment, introduced by Hilary Putnam,[22] is responsible for    one of the main arguments used against functionalism, although    it was originally intended as an argument against semantic internalism. The    thought experiment is simple and runs as follows. Imagine a    Twin Earth which is identical to Earth in every way but one:    water does not have the chemical structure HO, but rather some    other structure, say XYZ. It is critical, however, to note that    XYZ on Twin Earth is still called \"water\" and exhibits all the    same macro-level properties that HO exhibits on Earth (i.e.,    XYZ is also a clear drinkable liquid that is in lakes, rivers,    and so on). Since these worlds are identical in every way    except in the underlying chemical structure of water, you and    your Twin Earth doppelgnger see exactly the same things,    meet exactly the same people, have exactly the same jobs,    behave exactly the same way, and so on. In other words, since    you share the same inputs, outputs, and relations between other    mental states, you are functional duplicates. So, for example,    you both believe that water is wet. However, the content of    your mental state of believing that water is wet differs from    your duplicate's because your belief is of HO, while your    duplicate's is of XYZ. Therefore, so the argument goes, since    two people can be functionally identical, yet have different    mental states, functionalism cannot sufficiently account for    all mental states.  <\/p>\n<p>    Most defenders of functionalism initially responded to this    argument by attempting to maintain a sharp distinction between    internal and external content. The internal contents of    propositional attitudes, for example, would consist exclusively    in those aspects of them which have no relation with the    external world and which bear the necessary    functional\/causal properties that allow for relations with    other internal mental states. Since no one has yet been able to    formulate a clear basis or justification for the existence of    such a distinction in mental contents, however, this idea has    generally been abandoned in favor of externalist causal    theories of mental contents (also known as informational    semantics). Such a position is represented, for example, by    Jerry    Fodor's account of an \"asymmetric causal theory\" of mental    content. This view simply entails the modification of    functionalism to include within its scope a very broad    interpretation of input and outputs to include the objects that    are the causes of mental representations in the external world.  <\/p>\n<p>    The twin earth argument hinges on the assumption that    experience with an imitation water would cause a different    mental state than experience with natural water. However, since    no one would notice the difference between the two waters, this    assumption is likely false. Further, this basic assumption is    directly antithetical to functionalism; and, thereby, the twin    earth argument does not constitute a genuine argument: as this    assumption entails a flat denial of functionalism itself (which    would say that the two waters would not produce different    mental states, because the functional relationships would    remain unchanged).  <\/p>\n<p>    Another common criticism of functionalism is that it implies a    radical form of semantic holism. Block and Fodor[19] referred to this as the    damn\/darn problem. The difference between saying \"damn\"    or \"darn\" when one smashes one's finger with a hammer can be    mentally significant. But since these outputs are, according to    functionalism, related to many (if not all) internal mental    states, two people who experience the same pain and react with    different outputs must share little (perhaps nothing) in common    in any of their mental states. But this is counter-intuitive;    it seems clear that two people share something significant in    their mental states of being in pain if they both smash their    finger with a hammer, whether or not they utter the same word    when they cry out in pain.  <\/p>\n<p>    Another possible solution to this problem is to adopt a    moderate (or molecularist) form of holism. But even if this    succeeds in the case of pain, in the case of beliefs and    meaning, it faces the difficulty of formulating a distinction    between relevant and non-relevant contents (which can be    difficult to do without invoking an analytic-synthetic    distinction, as many seek to avoid).  <\/p>\n<p>    According to Ned Block, if functionalism is to avoid the    chauvinism of type-physicalism, it becomes overly liberal in    \"ascribing mental properties to things that do not in fact have    them\".[14]    As an example, he proposes that the economy of Bolivia might be    organized such that the economic states, inputs, and outputs    would be isomorphic to a person under some bizarre mapping from    mental to economic variables.[14]  <\/p>\n<p>    Hilary Putnam,[23]John    Searle,[24]    and others[25][26] have offered    further arguments that functionalism is trivial, i.e. that the    internal structures functionalism tries to discuss turn out to    be present everywhere, so that either functionalism turns out    to reduce to behaviorism, or to complete triviality and    therefore a form of panpsychism. These arguments typically use    the assumption that physics leads to a progression of unique    states, and that functionalist realization is present whenever    there is a mapping from the proposed set of mental states to    physical states of the system. Given that the states of a    physical system are always at least slightly unique, such a    mapping will always exist, so any system is a mind.    Formulations of functionalism which stipulate absolute    requirements on interaction with external objects (external to    the functional account, meaning not defined functionally) are    reduced to behaviorism instead of absolute triviality, because    the input-output behavior is still required.  <\/p>\n<p>    Peter Godfrey-Smith has argued    further[27] that such    formulations can still be reduced to triviality if they accept    a somewhat innocent-seeming additional assumption. The    assumption is that adding a transducer layer, that is,    an input-output system, to an object should not change whether    that object has mental states. The transducer layer is    restricted to producing behavior according to a simple mapping,    such as a lookup table, from inputs to actions on the system,    and from the state of the system to outputs. However, since the    system will be in unique states at each moment and at each    possible input, such a mapping will always exist so there will    be a transducer layer which will produce whatever physical    behavior is desired.  <\/p>\n<p>    Godfrey-Smith believes that these problems can be addressed    using causality, but that it may be necessary to    posit a continuum between objects being minds and not being    minds rather than an absolute distinction. Furthermore,    constraining the mappings seems to require either consideration    of the external behavior as in behaviorism, or discussion of    the internal structure of the realization as in identity    theory; and though multiple realizability does not seem to be    lost, the functionalist claim of the autonomy of high-level    functional description becomes questionable.[27]  <\/p>\n<p>    The general theory of adaptive biological systems, named    practopoiesis (meaning creation of actions), has been    used to derive a theory that explains mental operations as an    adaptive process. Much like species adapt through evolution and    an organism adapts through development, the theory of    anapoiesis (meaning re-creation) proposes that a thought    is a process of adaptation to the immediate environment. This    is performed by fast physiological machinery that can operate    within a few 100s of milliseconds and relies on the mechanisms    of neural adaptation. A key difference    between anapoietic approach and the functional approach is that    for anapoietic process much of the information needed for the    mental operations is located outside the organism. If mental    operations are an adaptive process, they do not juggle symbols    internally (like a computer) but make guesses of what changes    should be made to the nervous system and then test them against    the environment.  <\/p>\n<p>    The mechanisms of anapoiesis offer a solution to the problem of    the Chinese Room posed by John Searle.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Read the original:<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/en.wikipedia.org\/wiki\/Functionalism_(philosophy_of_mind)\" title=\"Functionalism (philosophy of mind) - Wikipedia\">Functionalism (philosophy of mind) - Wikipedia<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Functionalism is a theory of the mind in contemporary philosophy, developed largely as an alternative to both the identity theory of mind and behaviorism.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/mind-upload\/functionalism-philosophy-of-mind-wikipedia.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[16],"tags":[],"class_list":["post-214808","post","type-post","status-publish","format-standard","hentry","category-mind-upload"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/214808"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=214808"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/214808\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=214808"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=214808"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=214808"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}