How cold is really cold enough?

[This is another guest post by Doug Clow — thanks Doug! I asked a question on LongeCity: is the 29 kJ/mol figure for the activation energy of the “catalase reaction” given in How Cold Is Cold Enough correct? Doug was kind enough to give a detailed answer, and permission to edit it a little and reproduce here.]

I did do a chemistry degree, with a lot of biochemistry in it. And even as a postgraduate student attended a research seminar by another postgrad who was investigating catalase analogues, which almost certainly touched directly on the question. But that was long ago and I haven’t done this stuff in anger for decades.

Alas, I don’t have good data books to hand and can’t answer the direct question (“What is the activation energy of the decomposition of hydrogen peroxide when catalysed by catalase”) authoritatively.

Partly, it’s because there isn’t a single answer, and anyone who tells you there is is fibbing. There are shedloads of different catalases (more if you include general peroxidases). They are indeed legendarily fast, and they are more or less ubiquitous in oxygen-metabolising species. The story goes that it’s as perfectly evolved an enzyme as you can hope for. It’s not a bad choice for the worst-case scenario for this context, although I wouldn’t go as far as to say it was the very worst without checking up for other very-fast enzymes in metabolic pathways and signal transduction (e.g. acetylcholinesterase is also legendarily fast). Which would be overkill.

I’d say using any value between 1 kJ/mol and 20 kJ/mol is not unreasonable, and if you pressed me for a value, I’d probably settle on 10 kJ/mol as a round value. (See e.g. http://www.ncbi.nlm.nih.gov/pubmed/8320233 which gave 10 kJ/mol for a catalase from a halophile bacterium — no reason for choice except I alighted on it quickly, or http://www.sciencedirect.com/… which found 11 kJ/mol but looks odd for several reasons.)

At the very top end, a value of 50 kJ/mol for a reaction that happens at a reasonable rate for practical experimental purposes at room temperature is fairly typical. There’s a (sorely abused) rule of thumb that says that reaction rate doubles with an increase of 10 C, which only applies under fairly restrictive conditions, one of which is that the Ea is 50 kJ/mol.

This does, of course, yield materially different results. I tried duplicating that big table in ‘How Cold is Cold Enough’ in a toy spreadsheet, and couldn’t quite reproduce his results, but did get within an order of magnitude which is close enough for these purposes. I played around looking at his ‘Rate relative to liquid N2’ column, for different values for the activation energy.

  • 50 kJ/mol -> 2.2 x 1025 times faster at 37C than at LN2
  • 20 kJ/mol -> 1.4 x 1010 times faster
  • 10 kJ/mol -> 1.2 x 105 times faster
  • 8 kJ/mol -> 1.1 x 104
  • 5 kJ/mol -> 340 times faster
  • 2 kJ/mol -> 10 times faster
  • 1 kJ/mol -> 3.2 times faster

For the question at hand, this makes a huge difference — to this analysis.

This analysis is likely to be wrong, anyway.

A quick look at the Arrhenius equation:

k = A e-Ea/RT

Let’s take a very, very simplified reaction, where one molecule of reactant hits one catalyst to produce one product. The pre-exponential factor A represents the number of collisions that occur; the bit in the exponent tells you what proportion of those collisions have energy above the activation energy for the reaction.

Now, mathematical instinct might tell you that the bit in the exponent will give you all the action, but that’s not necessarily true. For practical purposes, the rate of catalase in vivo is limited by the rate at which molecules collide, not by the proportion of the molecules colliding which have greater than the activation energy needed for the reaction. Essentially, if a molecule of hydrogen peroxide bumps in to catalase, it’s breaking down. My biochemical intuition is likely to be sorely astray at LN2 temperatures, but I’d guess the same situation applied.

The pre-exponential factor A is probably more key: it’s the rate at which collisions occur between molecules that might react. If you have a perfectly efficient catalyst, this is the main factor affecting the rate of reaction — which makes sense, since they’d reduce the activation energy to a negligible value. Some enzymes — catalase is an excellent example — have been under geological periods of selection pressure in that direction. The Arrhenius equation is a simplification that works (better than it ought to) across a lot of practically-important situations. (One simplification is that the activation energy is not temperature-dependent. It sometimes is.)

If you get a phase change to solid — vitrification at very low temperatures — then you’ll get a staggering-number-of-orders-of-magnitude change in A. Those molecules are going nowhere fast, and so are flat out not going to bump in to each other. Never mind how much energy they’ve got when they do.

So I think that all is not lost for cryonics on this point.

Doug Clow on the Whole Brain Emulation roadmap

[A guest post from Doug Clow. This was a comment in this article on Bostrom and Sandberg’s Whole Brain Emulation: a Roadmap, but given its length and substance I am with permission putting it here as a new blog post.]

I too am short of time, but have given this paper a quick run through. Here are some unstructured and unedited quick notes I made while I was at it. Apologies for brevity and errors — I almost certainly missed some of their points and have misrepresented parts of their case.

It does seem to be a serious and reasonably well-informed piece of work on speculative science and technology. Emphasis on the speculative, though — which they acknowledge.

The distinction between emulating a brain generically (which I reckon is probably feasible, eventually) and emulating a specific person’s brain (which I reckon is a lot harder), and emulating a specific dead person’s brain (which I reckon is probably not possible), is a crucial one. They do make this point and spell it out in Table 1 on p11, and rightly say it’s very hard.

p8 “An important hypothesis for WBE is that in order to emulate the brain we do not need to understand the whole system, but rather we just need a database containing all necessary low‐level information about the brain and knowledge of the local update rules that change brain states from moment to moment.”

I agree entirely. Without this the ambitious bit of the enterprise fails. (They make the case, correctly, that progress down these lines is useful even if it turns out the big project can’t be done.) I suspect that this hypothesis may be true, but we certainly need to know a lot more about how the whole system works in order to work out what the necessary low-level information and update rules are. And in fact we’ll make interesting scientific progress – as suggested here – by running emulations of bits of the brain we think we might understand and seeing if that produces emergent properties that look like what the brain does. Actually they say this on p15 “WBE appears to be a way of testing many of these assumptions experimentally” – I’d be a bit stronger than that.

Table 2 on levels of emulation makes sense. My gut instinct (note evidence base) is that we will need at least level 8 (states of protein complexes – i.e. what shape conformations the (important) proteins are in) to do WBE, and quite possibly higher ones (though I doubt the quantum level, 11, is needed but Roger Penrose would disagree). Proteins are the actually-existing nanobots that make our cells work. The 3D shape of proteins is critical to their role. Many proteins change shape – and hence what they do or don’t do – in to a smallish fixed number of conformations, and we already know that this can be hugely important to brain function at the gross level. (E.g. transmissible spongiform encaphalopathies – mad cow and all that – are essentially caused by prion proteins in the brain switching from the ordinary shape to the disease-causing one.)

The whole approach is based on scanning an existing brain, in sufficient detail that you can then implement an emulation. I think that’s possibly useful, but I think a more likely successful route to a simulated (!) intelligence will be to grow it, rather than to bring it in to existence fully-formed. By growing, I mean some process akin to the developmental process by which humans come to consciousness: an interaction between an environment and a substrate that can develop in the light of feedback from that environment. But based on their approach, their analysis of technological capabilities needed seems plausible.

The one that leaps out as really, really hard (to the point of impossibility in my mind) is the scanning component. There is the unknown of whether the thing is doable at all (what they call scale separation), which is a biggy, but falsifiable by trying out experiments in this direction.

 continue reading

David Matthewman on the Whole Brain Emulation roadmap

By far the best technical objection I’ve heard so far is David Matthewman’s comments here, discussing Bostrom and Sandberg’s Whole Brain Emulation: a Roadmap:

[…] I still wouldn’t get your hopes up. As far as I can see, it offers no way to discern the state of the neurons, and admits that while it might be possible to get the structure for a small slice of the brain, getting it in 5nm detail for the whole volume is currently impossible, with no known way to overcome the current technological limitations. Many of those limitations are imposed by the wavelength of the medium you’re scanning with, and there’s just no easy way round that. The speed of scanning (which is also currently a showstopper for the ~5nm technologies that might otherwise be attractive) might be able to be improved, but bear in mind that you’re working at levels where the energy of the electrons/photons that you’re using to scan risk damaging the sample, and using more of them in parallel may damage it more. The data transfer/storage problem probably is solvable, by contrast.

I find it a bit worrying that the most promising technologies in the table on page [53] — SOM and SEM, especially combined with array tomography — have relatively little discussion in the text that I can see. This makes me suspect that they’re only even superficially attractive because not enough is known about them to know they don’t work.

Also, given that the conclusion says ‘this sets a resolution requirement on the order of 5 nm at least in two directions,’ there’s far too much discussion of technologies that can only scan down to resolutions two orders of magnitude higher than this. So the text gives the optimistic prediction that ‘[KESM] enables the imaging of an entire macroscopic tissue volume such as a mouse brain in reasonable time’, but what good is that given that KSEM only scans down to 300nm x 500nm? It’s an obvious question, and I’d expect an honestly-written paper to answer it. Because this paper doesn’t, I smell a rat (or, more likely, someone clutching at straws).

The discussion starts ‘As this review shows, WBE on the neuronal/synaptic level requires relatively modest increases in microscopy resolution…’ which may be technically true but vastly understates the difficulty of increasing the resolution of the techniques discussed.

Again, though, I’ll defer to someone who’s done this stuff more recently than I have (and in a medical area — I was mostly looking at metal-matrix composites rather than anything organic).

This stands out from the field in that it is actually in reply to something that someone who believes in cryonics has actually said; it only doesn’t meet the criteria that I asked for in my open letter in that it is blog comment rather than an article, but knowing how busy David is it’s hard to imagine him finding the time to rewrite it in article form any time soon, so with his permission I’m posting it as is.

Updated 2010-02-20: Liam Proven steps up to the plate and meets three of my four criteria. Thanks Liam!