< 587 older entriesHome184 newer entries >

Supposing the truth

Here is a coin. What would have happened if I had just tossed it? It might have landed heads, and it might have landed tails. If the coin is biased towards tails, it is more likely that it would have landed heads. If it's a fair coin, both outcomes are equally likely. That is, they are equally likely on the supposition that the coin had been tossed. Let's write this as P(Heads // Toss) = 1/2, where the double slash indicates that the supposition in question is "subjunctive" rather than "indicative".

Decision-making under determinism

Suppose you have a choice between two options, say, raising your arm and lowering your arm. To evaluate these options, we should compare their outcomes: what would happen if you raise your arm, what if you don't? But we don't want to be misled by merely evidential correlations. Your raising your arm might be evidence that your twin raised their arm in a similar decision problem yesterday, but since you have no causal control over other people's past actions, we should hold them fixed when evaluating your options. Similarly, your choice might be evidentially relevant to hypotheses about the laws of nature, but you have no causal control over the laws, so we should hold them fixed. But now we have a problem. The class of facts outside your causal control is not closed under logical consequence. On the contrary, if the laws are deterministic then facts about the distant past together with the laws logically settle what you will do. We can't hold fixed both the past and the laws and vary your choice.

Review of Tyler Burge: Origins of Objectivity

Earlier this year, I read Tyler Burge's Origins of Objectivity. It's a very long book. Here is an abridged version. A few comments below.

Origins of Objectivity

Representation is a basic explanatory kind in psychology that should be distinguished from mere information-carrying. The most fundamental type of representational state is perception. In perception, an organism attributes properties to objects in its environment. To do this, the organism does not need linguistic capacities, nor does it need to know (or otherwise represent) necessary and sufficient conditions for being the relevant object. Instead, the science of perception reveals that it is sufficient that the organism stands in a suitable causal relation to the object and that its perceptual state involves certain constancies (for shape or colour or distance or whatever) which characterize the object "objectively", abstracting away from contingencies of the present stimulus.

I like the starting point — to think of intentional states as explanatory scientific kinds. Burge doesn't say what exactly he means by this. I would put it as a kind of functionalism: intentional states are characterized (at least in part) by their functional inter-connections and their relationship to environmental causes, behaviour and other psychologically relevant facts.

Some counterexamples to the Benchmark Theory

In "Gandalf's solution to the Newcomb problem" (2013), Ralph Wedgwood proposes a new form of decision theory, Benchmark Theory, that is supposed to combine the good parts of Causal and Evidential Decision Theory.

Like many formulations of Causal Decision Theory, Benchmark Theory (BT) assumes a privileged partition of states that are outside the agent's causal control. Like Evidential Decision Theory, BT only considers the probability of these states conditional on a given act A. However, what is weighted by the conditional probabilities P(S_i/A) is not the absolute utility of S_i & A, but the comparative utility of S_i & A, which is determined by the difference between the absolute utility U(S_i & A) and the average absolute utility U(S_i & A') for all options A'. (This average is the benchmark B(S_i).) So the degree of choiceworthiness of an act A is given by

The lure of free energy

There's an exciting new theory in cognitive science. The theory began as an account of message-passing in the visual cortex, but it quickly expanded into a unified explanation of perception, action, attention, learning, homeostasis, and the very possibility of life. In its most general and ambitious form, the theory was mainly developed by Karl Friston -- see e.g. Friston 2006, Friston and Stephan 2007, Friston 2009, Friston 2010, or the Wikipedia page on the free-energy principle.

Indefinites and implicatures

Suppose I say (*), with respect to a particular gambling occasion.

(*) A gambler lost some of her savings. Another lost all of hers.

There is an implicature here that the first gambler, unlike the second, didn't lose all her savings. How does this implicature arise?

On the standard account of scalar implicatures, we should consider certain alternatives to the uttered sentences. In particular, I could have said 'A gambler lost all of her savings' instead of 'A gambler lost some of her savings'. If true, this alternative would have been more informative. Since I chose the weaker sentence, you can infer that I wasn't in a position to assert the stronger sentence. Assuming I am well-informed, you can further infer that the stronger sentence is false.

Against countable additivity

Imagine the universe has a centre that regularly produces new stars which then drift away at a constant speed. This has been going on forever, so there are infinitely many stars. We can label them by age, or equivalently by their distance from the centre: star 1 is the youngest, then comes star 2, then star 3, and so on, without end. The stars in turn produce planets at regular intervals. So the older a star, the more planets surround it. Today, something happened to one (and only one) of the planets. Let's say it exploded. Given all this, what is your credence that the unfortunate planet belonged to the first 100 stars? What about the second 100? It would be odd to think that the event is more likely to have happened at one of the first 100 stars than at one of the next 100, since the latter have far more planets. Similarly if we compare the first 1000 stars with the next 1000, or the first million with the next million, and so on. But there is no countably additive (real-valued) probability measure that satisfies this constraint.

Conditional chance and rational credence

Two initially plausible claims:

  1. Sometimes, a possible chance function conditionalized on a proposition A yields another possible chance function.
  2. Any rational prior credence function Cr conditional on the hypothesis Ch=f that f is the (actual, present) chance function should coincide with f; i.e., Cr(A / Ch=f) = f(A) for all A (provided that Cr(Ch=f)>0).

Claim 1 is a supported by the popular idea that chances evolve by conditionalizing on history, so that the chance at time t2 equals the chance at t1 conditional on the history of events between t1 and t2. Claim 2 is a weak form of the Principal Principle and often taken to be a defining feature of chance.

Moving to Germany

Inga got a postdoc in Hamburg, so it looks like we'll be moving back to Germany at the end of the year. It's sad to leave the ANU, but we'll probably return here for at least a few months in 2014. (If only because I don't have another job yet.)

Second-order logic and Newman's problem

How much can you say about the world in purely logical terms? In first-order logic with identity, one can construct formulas like '(Ex)(Ey)~(x=y)'. But arguably, this doesn't yet mean anything. As we learned in intro logic, formulas of first-order logic have no fixed interpretation; they mean something only once we provide a domain of quantification and an assignment of values to predicate and function symbols. As it happens, '(Ex)(Ey)~(x=y)' doesn't contain any non-logical predicate and function symbols, so to make it mean anything we just need to specify a domain of quantification. For example, if the domain is the class of Western black rhinos, then the formula says that there are at least two Western black rhinos.

< 587 older entriesHome184 newer entries >