Archive for January, 2008

Deccelerating change?

Sunday, January 27th, 2008

Everyone knows the first words spoken by a man on the moon, but what were the last words? This isn’t just a good pub quiz question, it’s also an affront to the notion that technological progress moves inexorably forward. To critics of the idea that technology is relentlessly accelerating, the fact that space travel now constitutes a technology that the world has essentially relinquished is a prime argument against the idea of inevitable technological progress. The latest of such critics is David Edgerton, whose book The Shock of the Old is now out in paperback.

Edgerton’s book has many good arguments, and serves as a useful corrective to the technological determinism that characterises quite a lot of discussion about technology. His aim is to give a history of innovation which de-emphasises the importance of invention, and to this end he helpfully draws attention to the importance of those innovations which occur during the use and adaptation of technologies, often quite old ones. One very important thing this emphasis on innovation in use does is bring into focus neglected innovations of the developing world, like the auto-rickshaw of India and Bangladesh and the long-tailed boat of Thailand. This said, I couldn’t help finding the book frequently rather annoying. Its standard rhetorical starting point is to present, generally without any reference, a “standard view” of the history of technology that wouldn’t be shared by anyone who knows anything about the subject: a series of straw men, in other words. This isn’t to say that there aren’t a lot of naive views about technology in wide circulation, but to suggest, for example, that it is the “conventional story” that the atomic bomb was the product of academic science, rather than the gigantic military-industrial engineering activity of the Manhatten Project, seems particularly far-fetched.

The style of the book is essentially polemic and anecdotal, the statistics that buttress the argument tending to be of the factoid kind (such as the striking assertion that the UK is home to 3.8 million unused fondue sets). In this and many other respects I found it a much less satisfying book than Vaclav Smil’s excellent 2-volume history of modern technology, Transforming the Twentieth Century: Technical Innovations and Their Consequences and Creating the Twentieth Century: Technical Innovations of 1867-1914 and Their Lasting Impact. These books reach similar conclusions, though Smil’s arguments are supported by substantially more data and carry a greater impact for being less self-consciously contrarian.

Smil’s view – and I suspect that Edgerton would share it, though I don’t think he states it so explicitly – is that the period of history in which there was the greatest leap forward in technology wasn’t present times, but the thirty or forty years of the late 19th and early 20th century that saw the invention of the telephone, the automobile, the aeroplane, electricity, mass production, and most important of all, the Haber-Bosch process. What then of that symbol of what many people think of as the current period of accelerating change – Moore’s law? Moore’s law is an observation about exponential growth of computer power with time, and one should start with an obvious point about exponential growth – it doesn’t come from accelerating change, but constant fractional change. If you are able to improve a process by x% a year, you get exponential growth. Moore’s law simply tells us that the semiconductor industry has been immensely successful at implementing incremental improvements to their technology, albeit at a rapid rate. Stated this way, Moore’s law doesn’t seem so out of place in Edgerton’s narrative of technology as being dominated, not by dramatic new inventions, but by many continuous small improvements in technologies old and new. This story, though, also makes clear how difficult it is to predict, before several generations of this kind of incremental improvement, which technologies are destined to have a major and lasting impact and which ones will peter out and disappoint their proponents. For me, therefore, the lesson to take away is not that new developments in science and technology might not have major and lasting impacts on society, it is simply that some humility is needed when one tries to identify in advance what will have lasting impact and what those impacts will end up being.

On December 17th, 1972, Eugene A. Cernan said the last words by a man on the moon: “OK Jack, let’s get this mutha outta here.”

Invisibility cloaks and perfect lenses – the promise of optical metamaterials

Sunday, January 20th, 2008

The idea of an invisibility cloak – a material which would divert light undetectably around an object – captured the imagination of the media a couple of years ago. For visible light, the possibility of an invisibility cloak remains a prediction, but it graphically illustrates the potential power of a line of research initiated a few years ago by the theoretical physicist Sir John Pendry of Imperial College, London. Pendry realised that constructing structures with peculiar internal structures of conductors and dielectrics would allow one to make what are in effect new materials with very unusual optical properties. The most spectacular of these new metamaterials would have a negative refractive index. In addition to making an invisibility cloak possible one could in principle use negative refractive index metamaterials to make a perfect lens, allowing one to use ordinary light to image structures much smaller than the limit of a few hundred nanometers currently set by the wavelength of light for ordinary optical microscopy. Metamaterials have been made which operate in the microwave range of the electromagnetic spectrum. But to make an optical metamaterial one needs to be able to fabricate rather intricate structures at the nanoscale. A recent paper in Nature Materials (abstract, subscription needed for full article) describes exciting and significant progress towards this goal. The paper, whose lead author is Na Liu, a student in the group of Harald Giessen at the University of Stuttgart, describes the fabrication of an optical metamaterial. This consists of a regular, three dimensional array of horseshoe shaped, sub-micron sized pieces of gold embedded in a transparent polymer – see the electron micrograph below. This metamaterial doesn’t yet have a negative refractive index, but it shows that a similar structure could have this remarkable property.

An optical metamaterial
An optical metamaterial consisting of split rings of gold in a polymer matrix. Electron micrograph from Harald Giessen’s group at 4. Physikalisches Institut, Universität Stuttgart.

To get a feel for how these things work, it’s worth recalling what happens when light goes through an ordinary material. Light, of course, consists of electromagnetic waves, so as a light wave passes a point in space there’s a rapidly alternating electric field. So any charged particle will feel a force from this alternating field. This leads to something of a paradox – when light passes through a transparent material, like glass or a clear crystal, it seems at first that the light isn’t interacting very much with the material. But since the material is full of electrons and positive nuclei, this can’t be right – all the charged particles in the material must be being wiggled around, and as they are wiggled around they in turn must be behaving like little aerials and emitting electromagetic radiation themselves. The solution to the paradox comes when one realises that all these waves emitted by the wiggled electrons interfere with each other, and it turns out that the net effect is of a wave propagating forward in the same direction as the light thats propagating through the material, only with a somewhat different velocity. It’s the ratio of this effective velocity in the material to the velocity the wave would have in free space that defines the refractive index. Now, in a structure like the one in the picture, we have sub-micron shapes of a metal, which is an electrical conductor. When this sees the oscillating electric field due to an incident light wave, the free electrons in the metal slosh around in a collective oscillation called a plasmon mode. These plasmons generate both electric and magnetic fields, whose behaviour depends very sensitively on the size and shape of the object in which the electrons are sloshing around in (to be strictly accurate, the plasmons are restricted to the region near the surface of the object; its the geometry of the surface that matters). If you design the geometry right, you can find a frequency at which both the magnetic and electric fields generated by the motion of the electrons is out of phase with the fields in the light wave that are exciting the plasmons – this is the condition for the negative refractive index which is needed for perfect lenses and other exciting possibilities.

The metamaterial shown in the diagram has a perfectly periodic pattern, and this is what’s needed if you want a uniform plane wave arriving at the material to excite another uniform plane wave. But, in principle, you should be able to design an metamaterial that isn’t periodic to direct and concentrate the light radiation any way you like, on length scales well below the wavelength of light. Some of the possibilities this might lead to were discussed in an article in Science last year, Circuits with Light at Nanoscales: Optical Nanocircuits Inspired by Metamaterials (abstract, subscription required for full article) by Nader Engheta at the University of Pennsylvania. If we can learn how to make precisely specified, non-periodic arrays of metallic, dielectric and semiconducting shaped elements, we should be able to direct light waves where we want them to go on the nanoscale – well below light’s wavelength. This might allow us to store information, to process information in all-optical computers, to interact with electrons in structures like quantum dots, for quantum computing applications, to image structures using light down to the molecular level, and to detect individual molecules with great sensitivity. I’ve said this before, but I’m more and more convinced that this is a potential killer application for advanced nanotechnology – if one really could place atoms in arbitrary, pre-prescribed positions with nanoscale accuracy, this is what one could do with the resulting materials.

The Tata Nano

Monday, January 14th, 2008

The Tata Nano – the newly announced one lakh (100,000 rupees) car from India’s Tata group – hasn’t got a lot to do with nanotechnology (see this somewhat bemused and bemusing piece from the BBC), but since it raises some interesting issues I’ll use the name as an excuse to discuss it here.

The extensive media coverage in the Western media has been characterised by some fairly outrageous hypocrisy – for example, the UK’s Independent newspaper wonders “Can the world afford the Tata Nano?” (The answer, of course, is that what the world can’t afford are the much bigger cars parked outside all those prosperous Independent readers’ houses). With a visit to India fresh in my mind, it’s completely obvious to me why all those families one sees precariously perched on motor-bikes would want a small, cheap, economical car, and not at all obvious that those of us in the West, who are used to enjoying on average 11 times (for the UK) or 23 times (for the USA) more energy per head than the Indians, have any right to complain about the extra carbon dioxide emissions that will result. It’s almost certainly true that the world couldn’t sustain a situation in which all its 6.6 billion population used as much energy as the Americans and Europeans; the way that equation will be squared, though, ultimately must be by the rich countries getting by with less energy rather than by poorer countries being denied the opportunity to use more. It is to be hoped that this transformation takes place in a way that uses better technology to achieve the same or better living standards for everybody from a lot less energy; the probable alternative is the economic disruption and widespread involuntary cuts in living standards that will follow from a prolonged imbalance of energy supply and demand.

A more interesting question to ask about the Tata Nano is to wonder why it was not possible to leapfrog current technology to achieve something even more economical and sustainable – using, one hesitates to suggest, actual nanotechnology? Why is the Nano made from old-fashioned steel, with an internal combustion engine in the back, rather than, say, being made from advanced lightweight composites and propelled by an electric motor and a hydrogen fuel cell? The answers are actually fairly clear – because of cost, the technological capacity of this (or any other) company, and the requirement for maintainability. Aside from these questions, there’s the problem of infrastructure. The problems of creating an infrastructure for hydrogen as a fuel are huge for any country; liquid hydrocarbons are a very convenient store of energy, and, old though it is, the internal combustion engine is a pretty effective and robust device for converting energy. Of course, we can hope that new technologies will lead to new versions of the Tata Nano and similar cars of far greater efficiency, though realism demands that we understand the need for new technology to fit into existing techno-social systems to be viable.

Grand challenges for UK nanotechnology

Saturday, January 12th, 2008

The UK’s Engineering and Physical Sciences Research Council introduced a new strategy for nanotechnology last year, and some of the new measures proposed are beginning to come into effect (including, of course, my own appointment as the Senior Strategic Advisor for Nanotechnology). Just before Christmas the Science Minister announced the funding allocations for research for the next few years. Nanotechnology is one of six priority programmes that cut across all the Research Councils (to be precise, the cross-council programme has the imposing title: Nanoscience through Engineering to Application).

One strand of the strategy involves the funding of large scale integrated research programmes in areas where nanotechnology can contribute to issues of pressing societal or economic need. The first of these Grand Challenges – in the area of using nanotechnology to enable cheap, efficient and scalable ways to harvest solar energy – was launched last summer. An announcement on which proposals will be funded will be made within the next few months.

The second grand challenge will be launched next summer, and it will be in the general area of nanotechnology for healthcare. This is a very broad theme, of course – I discussed some of the potential areas, which include devices for delivering drugs and for rapid diagnosis, in an earlier post. To narrow the area down, there’s going to be an extensive process of consultation with researchers and people in the relevant industries – for details, see the EPSRC website. There’ll also be a role for public engagement; EPSRC is commissioning a citizens’ jury to consider the options and have an input into the decision of what area to focus on.

UK Government outlines nanorisk research needs

Friday, January 4th, 2008

The UK government has released a second report reviewing progress and identifying knowledge gaps about the potential environmental and health risks arising from engineered nanoparticles. This is a comprehensive document, breaking down the problem into five areas. The first of these is the question of how you detect and measure nanoparticles and the second considers the ways in which people and the environment might be exposed to nanoparticles. The third area concerns the assessment of the degree to which some nanoparticles might be toxic to humans, while the fourth area considers potential environmental impacts. Finally, a fifth section considers wider social and economic dimensions of nanotechnology.

The document represents, in part, a response to the very critical verdict on the UK government’s response on nanotoxicology given by the Council for Science and Technology last March. It isn’t, of course, able to address the fundamental criticism: that the Government didn’t act on the recommendation of the Royal Society and set up a coordinated programme of research into the toxicology and health and environmental effects of nanomaterials, with dedicated funding, but instead relied on an ad-hoc process of waiting for proposals to come in through peer review with opportunistic funding from a number of sources. The response from the Royal Society reflects the continuing frustration at opportunities lost: “The Government has recognised the huge potential of nanotechnology and recognised what needs to be done to ensure that advances are realised safely, but by their own admission progress has been slow in some areas. Given the wealth of expertise in UK universities and industries we should be further ahead.

That’s old ground now, of course, so perhaps it’s worth focusing on some of the positive outcomes reported in the report. Quite a lot of work has been carried out or at least started. In the area of nanoparticles in the environment, for example, the Natural Environment Research Council has funded more than £2.3 million worth of projects, in areas ranging from studies of the toxicity of nanoparticles to fish and other aquatic organisms, to studies of the fate of silicon dioxide nanoparticles from pharmaceutical and cosmetic formulations in wastewaters and of the effect of silver nanoparticles on natural bacterial populations.

For another view of the positives and negatives of this report, it’s interesting to see the response of nanoparticle expert Andrew Maynard. More shocking is the way this report is mendaciously misquoted in an article in the Daily Mail: Alert over the march of the ‘grey goo’ in nanotechnology Frankenfoods (via TNTlog).