081204
Suppose that what it means to be a particular individual in possession of a particular individual mind is to be a particular individual physical device that implements a particular individual algorithm.
What an Algorithm Is
Searle 2004 (p.67) describes an algorithm as a method for solving a problem by going through a precise series of steps. The steps must be finite in number, and if carried out correctly, they guarantee a solution to the problem.
By talking about solving a problem, the description takes for granted that in order to create or recognize an algorithm we must have1) a problem, 2) a way to describe it, and 3) a way to describe its solution. Within this formulation, an algorithm can only have so-called derived intentionality, viz. before an algorithm can come into existence somebody has to have a purpose that determines what the algorithm is to be about. As Searle points out, a calculator (an instantiation of a calculation algorithm) doesnt do calculations. What it does is change states in a deterministic way in response to physical movements of some of its components (key presses) and alter its appearance (displaying results) as a side effect of its internal state changes. A calculator is usable as a calculator by human beings only because human beings assign calculation-related meanings to key presses and display patterns. The meaning does not reside in the calculator, it resides in the user. Following this line of thought, Searle concludes that neither syntax nor semantics is present in an algorithm. This he says is because syntax and semantics are constructs present only in conscious human beings.
These conclusions are warranted under the given definition of what constitutes an algorithm. However, I will propose an alternative definition that I will argue allows for something we can still reasonably call an algorithm to have syntax and, in its instantiations, semantics without having to provide either from an external source.
I propose to consider algorithms as being about the implementation (realization) of behaviors in time. In a sense, then, an algorithm is an abstraction that specifies a particular deterministic computer program. More formally, an algorithm is a finite series of instructions[1] (steps) that comprise a behavior (a generalization of the idea of performing a task). Algorithms are constructed on the basis of a set of primitive functions (the basis functions) that, taken together specify the operation of an abstract (virtual) machine (computer). It is not possible to specify an algorithm without specifying the set of primitive functions in terms of which the algorithm is expressed, although informally the set of primitives is simply taken to contain whatever functions the specification of a particular algorithm requires. The abstract machine implicit in the set of primitive functions can be described in terms of its computational power (the class of calculations it is capable of). The two most common (and for our purposes most relevant) levels of computational power are 1) computations that are possible for finite state machines and 2) computations that are possible for (unrestricted) Turing machines. The former is formally equivalent to (has identical computational power as) the subset of Turing machines having only a finite tape.
It is, in general, tedious in the extreme to express algorithms in terms of Turing machine functions. And it is also tedious in many cases to make explicit the set of primitive functions that provide the basis for a particular algorithm or set of algorithms. For that reason (and, one assumes, not incidentally because people thought up the list-of-instructions concept long before Turing thought up the machine formalism that bears his name) the specification of most algorithms leaves the specification of the underlying set of primitive functions implicit. That works pretty well and we all learn (maybe now I have to say used to learn) addition, subtraction, multiplication, division, and square root algorithms in elementary school arithmetic, without belaboring or worrying overmuch about the specifics of the underlying primitive functions, e.g., the fact that the set of functions on which the addition algorithm depends includes a function that enables one to write a sort of superscript number above and to the left of a specified decimal position in the topmost number of a column of numbers to be added (the carry) and a function that enables one to read it back to oneself at a later time, and so on.
Special Considerations re: Primitive Functions
Without attempting a rigorous exposition, we may take a mathematical function to be a (deterministic) relation that uniquely associates each element of its domain (the set of possible input values) with an elements of its range (the set of possible output values), in other words, an input-output specification. By definition, mathematical functions do not have side effects. This has all sorts of good consequences for proving theories and doing mathematical logic. However, for the specification of algorithms useful for talking about how the brain works, we need the computer science definition of a function, which generalizes the definition of function to include processes that have side effects.
Side Effects and Referential Opacity
A side effect is an event or process affected by and/or affecting the state of the environment external to the (implementation of the) function and occurring or commencing conceptually within the space or interval between the arrival of its input and the corresponding return of its output. The most common use of computer functions with side effects is to obtain inputs from or deliver outputs to the external environment.[2]
Function side effects are problematical for mathematical theories of computation, because they introduce the unconstrained external world into an otherwise nicely circumscribed theoretical construct. The formal response of computer science has been to expand the boundaries of the theoretical construct to include the possibility of a limited set of side effects explicitly in the domain and range of the function. The drawback this creates is that the more the domain and range include of the external world, the more difficult it is to formally prove program correctness. Nonetheless, functions with side effects are an integral part of the standard description of a Turing machine: specifically, the operations (functions) that read from and write to the machines (infinitely extensible) tape.[3]
At the most fundamental level, the issues raised by side-effects and referential opacity relate to the (theoretically, at least) arbitrary boundary between a system selected for analysis and the external environment in which it is embedded.[4] Because a theory of the mind must I think be about the brain in the context of an (external) environment that is affected by and affects brains, we need to be able to draw a boundary between an entity in possession of a mind and the environment in which it operates.[5] We thus need to allow for side effects in a theory of the mind, simply in recognition of the fact that not everything that happens in a mind originates in processes within the mind. There is an outside world, and it matters.[6]
Side effects show up in the realization (instantiation, physical implementation) of an algorithm in two ways. 1) The set of basis functions for an algorithmic system may include functions that explicitly query and manipulate the physical environment. 2) The physical processes that implement the algorithm have real physical side effects that are above and beyond (outside of) the abstract description of the algorithmoutside, even, the abstractly specified side effects that may be required in the basis functions. For example, a computer needs a power source that meets certain specifications and will operate only under a specified range of environmental conditions.
When analyzing or describing the behavior of a particular physical realization of an algorithm, we generally concentrate on side effects of the first kind and take for granted that the physical side effects of the second kindthose that ground or enable the basis functions themselvesare in place.
Significance of Side Effects
[Didnt I say whats in this paragraph earlier?]
The introduction of functions with open-ended side-effects has the effect of vastly complicating (often to the point of practical impossibility) any complete formal analysis of algorithmic behavior. This, because a formal analysis can only take place within a formal description of a closed system (one in which all events occur deterministically through fully specified rules and circumstances). To the extent that side effects bring aspects of the external world into a system, a formal description of the system must comprehend a formal description of at least those aspects of the external world. In effect, the less constrained the range of allowable side effects, the broader must be the scope of the formal description of the system.
Sequencing and Counterfactuals
Philosophers appeal to the idea of counterfactuals in order to deal with the fact that at the macro physical (as opposed to the quantum physical) level events only happen one way, although our intuitions tell us that if things had been sufficiently different (a counterfactual condition) events would have turned out differently. In planning for the future, where there are no facts yet, we just use conditionals (e.g., If the creek don’t rise and the sky don’t fall, I’ll visit you next Thursday). Computer programming is a quintessential case of formal planning for the future. The programmer’s task is to devise an algorithm that will accomplish whatever it is supposed to accomplish under constrained, but as yet undetermined conditions.
Sequencing and counterfactuals are at the heart of causal descriptions. Sequencing and conditionals are at the heart of algorithmic descriptions. Every entry in the state table of a Turing machine consists in a conditional (e.g., if the symbol under the read/write head on the tape is “1”), an action (e.g., write “0” on the tape and move the read/write head to the left), and a sequencing directive (e.g., go to state 275). In the abstract, sequencing begs the question of causality. If B must follow A, does that mean that A must cause B or is it sufficient that A must be correlated with B? Does it matter? In an algorithmic specification, the answer is no. In fact, it is not even a meaningful question because sequencing is primitive (and thus opaque). So causality is not an issue in an algorithmic specification.
It feels like we have two choices with respect to causality. We can stumble ahead in despite of all that Hume warned us about, and we will fall into the mire he described. Alternatively, we can take the view that informs physics, viz. systems evolve over time according to empirically reliable formulas. On this view, the attribution of causality requires drawing arbitrary physical and temporal boundaries in the midst of an evolving process to interpret as objects of interest at particular points in time. We then examine the state equations of the system in a small neighborhood near each selected point in time and we assign causality to the infinitesimally prior state.
In effect relativistic limits and the flow of time delimit the causes of everything. If there is anything that counts as Humes necessary connexion, it is to be found in the empirically based physical theory that the state of the universe everywhere in a light-cone of radius delta-t centered on the point P at time T is what determines what will happen at P at time T plus delta-t. The instant that one focuses attention on a proper subset of that light cone as a cause, necessary connexion becomes subject to ceteris paribus conditions.
If we want to say that some proper subset of the light cone centered on P at time T caused what happened at P at time T plus delta-t, we must recognize that this is a counterfactual that is falsifiable. Such an assertion requires a ceteris paribus qualification if we are to accept it as a true statement.
====================== Notes ======================