PART ONE The Garden-of-Eden Configuration

1 (Remit not paucity) NOVEMBER 2050

Maria Deluca had ridden past the stinking hole in Pyrmont Bridge Road for six days running, certain each time, as she’d approached, that she’d be greeted by the reassuring sight of a work team putting things right. She knew that there was no money for road works or drainage repairs this year, but a burst sewage main was a serious health risk; she couldn’t believe it would be neglected for long.

On the seventh day, the stench was so bad from half a kilometer away that she turned into a side street, determined to find a detour.

This end of Pyrmont was a depressing sight; not every warehouse was empty, not every factory abandoned, but they all displayed the same neglected look, the same peeling paint and crumbling brickwork. Half a dozen blocks west, she turned again—to be confronted by a vista of lavish gardens, marble statues, fountains and olive groves, stretching into the distance beneath a cloudless azure sky.

Maria accelerated without thinking—for a few seconds, almost believing that she’d chanced upon a park of some kind, an impossibly well-kept secret in this decaying corner of the city. Then, as the illusion collapsed—punctured by sheer implausibility as much as any visible flaw—she pedaled on wilfully, as if hoping to blur the imperfections and contradictions out of existence. She braked just in time, mounting the narrow footpath at the end of the cul-de-sac, the front wheel of her cycle coming to a halt centimeters from the warehouse wall.

Close up, the mural was unimpressive, the brushstrokes clearly visible, the perspective obviously false. Maria backed away—and she didn’t have to retreat far to see why she’d been fooled. At a distance of twenty meters or so, the painted sky suddenly seemed to merge with the real thing; with a conscious effort, she could make the border reappear, but it was hard work keeping the slight difference in hue from being smoothed out of existence before her eyes—as if some subsystem deep in her visual cortex had shrugged off the unlikely notion of a sky-blue wall and was actively collaborating in the deception. Further back, the grass and statues began to lose their two-dimensional, painted look—and at the corner where she’d turned into the cul-de-sac, every element of the composition fell into place, the mural’s central avenue now apparently converging toward the very same vanishing point as the interrupted road.

Having found the perfect viewing position, she stood there awhile, propping up her cycle. Sweat on the back of her neck cooled in the faint breeze, then the morning sun began to bite. The vision was entrancing—and it was heartening to think that the local artists had gone to so much trouble to relieve the monotony of the neighborhood. At the same time, Maria couldn’t help feeling cheated. She didn’t mind having been taken in, briefly; what she resented was not being able to be fooled again. She could stand there admiring the artistry of the illusion for as long as she liked, but nothing could bring back the surge of elation she’d felt when she’d been deceived.

She turned away.

+ + +

Home, Maria unpacked the day’s food, then lifted her cycle and hooked it into its frame on the livingroom ceiling. The terrace house, one hundred and forty years old, was shaped like a cereal box; two stories high, but scarcely wide enough for a staircase. It had originally been part of a row of eight; four on one side had been gutted and remodeled into offices for a firm of architects; the other three had been demolished at the turn of the century to make way for a road that had never been built. The lone survivor was now untouchable under some bizarre piece of heritage legislation, and Maria had bought it for a quarter of the price of the cheapest modern flats. She liked the odd proportions—and with more space, she was certain, she would have felt less in control. She had as clear a mental image of the layout and contents of the house as she had of her own body, and she couldn’t recall ever misplacing even the smallest object. She couldn’t have shared the place with anyone, but having it to herself seemed to strike the right balance between her territorial and organizational needs. Besides, she believed that houses were meant to be thought of as vehicles—physically fixed, but logically mobile—and compared to a one-person space capsule or submarine, the size was more than generous.

Upstairs, in the bedroom that doubled as an office, Maria switched on her terminal and glanced at a summary of the twenty-one items of mail which had arrived since she’d last checked. All were classified as “Junk”; there was nothing from anyone she knew—and nothing remotely like an offer of paid work. Camel’s Eye, her screening software, had identified six pleas for donations from charities (all worthy causes, but Maria hardened her heart); five invitations to enter lotteries and competitions; seven retail catalogues (all of which boasted that they’d been tailored to her personality and “current lifestyle requirements”—but Camel’s Eye had assessed their contents and found nothing of interest); and three interactives.

The “dumb” audio-visual mail was all in standard transparent data formats, but interactives were executable programs, machine code with heavily encrypted data, intentionally designed to be easier for a human to talk to than for screening software to examine and summarize. Camel’s Eye had run all three interactives (on a doubly quarantined virtual machine—a simulation of a computer running a simulation of a computer) and tried to fool them into thinking that they were making their pitch to the real Maria Deluca. Two sales programs—superannuation and health insurance—had fallen for it, but the third had somehow deduced its true environment and clammed up before disclosing anything. In theory, it was possible for Camel’s Eye to analyze the program and figure out exactly what it would have said if it had been fooled; in practice, that could take weeks. The choice came down to trashing it blind, or talking to it in person.

Maria ran the interactive. A man’s face appeared on the terminal; “he” met her gaze and smiled warmly, and she suddenly realized that “he” bore a slight resemblance to Aden. Close enough to elicit a flicker of recognition which the mask of herself she’d set up for Camel’s Eye would not have exhibited? Maria felt a mixture of annoyance and grudging admiration. She’d never shared an address with Aden—but no doubt the data analysis agencies correlated credit card use in restaurants, or whatever, to pick up relationships which didn’t involve cohabitation. Mapping useful connections between consumers had been going on for decades—but employing the data in this way, as a reality test, was a new twist.

The junk mail, now rightly convinced that it was talking to a human being, began the spiel it had refused to waste on her digital proxy. “Maria, I know your time is valuable, but I hope you can spare a few seconds to hear me out.” It paused for a moment, to make her feel that her silence was some kind of assent. “I also know that you’re a highly intelligent, discerning woman, with no interest whatsoever in the muddled, irrational superstitions of the past, the fairy tales that comforted humanity in its infancy.” Maria guessed what was coming next; the interactive saw it on her face—she hadn’t bothered to hide behind any kind of filter—and it rushed to get a hook in. “No truly intelligent person, though, ever dismisses an idea without taking the trouble to evaluate it—skeptically, but fairly—and here at the Church of the God Who Makes No Difference—”

Maria pointed two fingers at the interactive, and it died. She wondered if it was her mother who’d set the Church onto her, but that was unlikely. They must have targeted their new member’s family automatically; if consulted, Francesca would have told them that they’d be wasting their time.

Maria invoked Camel’s Eye and told it, “Update my mask so it reacts as I did in that exchange.”

A brief silence followed. Maria imagined the synaptic weighting parameters being juggled in the mask’s neural net, as the training algorithm hunted for values which would guarantee the required response. She thought: If I keep on doing this, the mask is going to end up as much like me as a fully fledged Copy. And what’s the point of saving yourself from the tedium of talking to junk mail if… you’re not? It was a deeply unpleasant notion… but masks were orders of magnitude less sophisticated than Copies; they had about as many neurons as the average goldfish—organized in a far less human fashion. Worrying about their “experience” would be as ludicrous as feeling guilty about terminating junk mail.

Camel’s Eye said, “Done.”

It was only 8:15. The whole day loomed ahead, promising nothing but bills. With no contract work coming in for the past two months, Maria had written half a dozen pieces of consumer software—mostly home-security upgrades, supposedly in high demand. So far, she’d sold none of them; a few thousand people had read the catalogue entries, but nobody had been persuaded to download. The prospect of embarking on another such project wasn’t exactly electrifying—but she had no real alternative. And once the recession was over and people started buying again, it would have been time well spent.

First, though, she needed to cheer herself up. If she worked in the Autoverse, just for half an hour or so—until nine o’clock at the latest—then she’d be able to face the rest of the day…

Then again, she could always try to face the rest of the day without bribing herself, just once. The Autoverse was a waste of money, and a waste of time—a hobby she could justify when things were going well, but an indulgence she could ill afford right now.

Maria put an end to her indecision in the usual way. She logged on to her Joint Supercomputer Network account—paying a fifty-dollar fee for the privilege, which she now had to make worthwhile. She slipped on her force gloves and prodded an icon, a wireframe of a cube, on the terminal’s flatscreen—and the three-dimensional workspace in front of the screen came to life, borders outlined by a faint holographic grid. For a second, it felt like she’d plunged her hand into some kind of invisible vortex: magnetic fields gripped and twisted her glove, as start-up surges tugged at the coils in each joint at random—until the electronics settled into equilibrium, and a message flashed up in the middle of the workspace: you may now put on your gloves.

She jabbed another icon, a starburst labeled FIAT. The only visible effect was the appearance of a small menu strip hovering low in the foreground—but to the cluster of programs she’d invoked, the cube of thin air in front of her terminal now corresponded to a small, empty universe.

Maria summoned up a single molecule of nutrose, represented as a ball-and-stick model, and, with a flick of a gloved forefinger, imparted a slow spin. The vertices of the crimped hexagonal ring zig-zagged above and below the molecule’s average plane; one vertex was a divalent blue atom, linked only to its neighbors in the ring; the other five were all tetravalent greens, with two bonds left over for other attachments. Each green was joined to a small, monovalent red—on the top side if the vertex was raised, on the bottom if it was lowered—and four of them also sprouted short horizontal spikes, built from a blue and a red, pointing away from the ring. The fifth green held out a small cluster of atoms instead: a green with two reds, and its own blue-red spike.

The viewing software rendered the molecule plausibly solid, taking into account the effects of ambient light; Maria watched it spin above the desktop, admiring the not-quite-symmetrical form. A real-world chemist, she mused, would take one look at this and say: Glucose. Green is carbon, blue is oxygen, red is hydrogen… no? No. They’d stare awhile; put on the gloves and give the impostor a thorough grope; whip a protractor out of the toolbox and measure a few angles; invoke tables of bond formation energies and vibrational modes; maybe even demand to see nuclear magnetic resonance spectra (not available—or, to put it less coyly, not applicable). Finally, with the realization of blasphemy dawning, they’d tear their hands from the infernal machinery, and bolt from the room screaming, “There is no Periodic Table but Mendeleev’s! There is no Periodic Table but Mendeleev’s!”

The Autoverse was a “toy” universe, a computer model which obeyed its own simplified “laws of physics”—laws far easier to deal with mathematically than the equations of real-world quantum mechanics. Atoms could exist in this stylized universe, but they were subtly different from their real-world counterparts; the Autoverse was no more a faithful simulation of the real world than the game of chess was a faithful simulation of medieval warfare. It was far more insidious than chess, though, in the eyes of many real-world chemists. The false chemistry it supported was too rich, too complex, too seductive by far.

Maria reached into the workspace again, halted the molecule’s spin, deftly plucked both the lone red and the blue-red spike from one of the greens, then reattached them, swapped, so that the spike now pointed upward. The gloves’ force and tactile feedback, the molecule’s laser-painted image, and the faint clicks that might have been plastic on plastic as she pushed the atoms into place, combined to create a convincing impression of manipulating a tangible object built out of solid spheres and rods.

This virtual ball-and-stick model was easy to work with—but its placid behavior in her hands had nothing to do with the physics of the Autoverse, temporarily held in abeyance. Only when she released her grip was the molecule allowed to express its true dynamics, oscillating wildly as the stresses induced by the alteration were redistributed from atom to atom, until a new equilibrium geometry was found.

Maria watched the delayed response with a familiar sense of frustration; she could never quite resign herself to accepting the handling rules, however convenient they were. She’d thought about trying to devise a more authentic mode of interaction, offering the chance to feel what it was “really like” to grasp an Autoverse molecule, to break and re-form its bonds—instead of everything turning to simulated plastic at the touch of a glove. The catch was, if a molecule obeyed only Autoverse physics—the internal logic of the self-contained computer model—then how could she, outside the model, interact with it at all? By constructing little surrogate hands in the Autoverse, to act as remote manipulators? Construct them out of what? There were no molecules small enough to build anything finely structured, at that scale; the smallest rigid polymers which could act as “fingers” would be half as thick as the entire nutrose ring. In any case, although the target molecule would be free to interact with these surrogate hands according to pure Autoverse physics, there’d be nothing authentic about the way the hands themselves magically followed the movements of her gloves. Maria could see no joy in simply shifting the point where the rules were broken—and the rules had to be broken, somewhere. Manipulating the contents of the Autoverse meant violating its laws. That was obvious… but it was still frustrating.

She saved the modified sugar, optimistically dubbing it mutose. Then, changing the length scale by a factor of a million, she started up twenty-one tiny cultures of Autobacterium lamberti, in solutions ranging from pure nutrose, to a fifty-fifty mixture, to one hundred percent mutose.

She gazed at the array of Petri dishes floating in the workspace, their contents portrayed in colors which coded for the health of the bacteria. “False colors”… but that phrase was tautological. Any view of the Autoverse was necessarily stylized: a color-coded map, displaying selected attributes of the region in question. Some views were more abstract, more heavily processed than others—in the sense that a map of the Earth, color-coded to show the health of its people, would be arguably more abstract than one displaying altitude or rain-fall—but the real-world ideal of an unadulterated, naked-eye view was simply untranslatable.

A few of the cultures were already looking decidedly sick, fading from electric blue to dull brown. Maria summoned up a three-dimensional graph, showing population versus time for the full range of nutrient mixtures. The cultures with only a trace of the new stuff were, predictably, growing at almost the pace of the control; with increasing mutose substitution the ascent gradually slowed, until, around the eighty-five percent line, the population was static. Beyond that were ever steeper trajectories into extinction. In small doses, mutose was simply irrelevant, but at high enough concentrations it was insidious: similar enough to nutroseA. lamberti’s usual food—to be taken part-way through the metabolic process, competing for the same enzymes, tying up valuable biochemical resources… but eventually reaching a step where that one stray blue-red spike formed an insurmountable barrier to the reaction geometry, leaving the bacterium with nothing but a useless byproduct and a net energy loss. A culture with ninety percent mutose was a world where ninety per cent of the food supply had no nutritional value whatsoever—but had to be ingested indiscriminately along with the worthwhile ten percent. Consuming ten times as much for the same return wasn’t a viable solution; to survive in the long term, A. lamberti would have to chance upon some means of rejecting mutose before wasting energy on it—or, better still, find a way to turn it back into nutrose, transforming it from a virtual poison into a source of food.

Maria displayed a histogram of mutations occurring in the bacteria’s three nutrose epimerase genes; the enzymes these genes coded for were the closest things A. lamberti had to a tool to render mutose digestible—although none, in their original form, would do the job. No mutants had yet persisted for more than a couple of generations; all the changes so far had evidently done more harm than good. Partial sequences of the mutant genes scrolled by in a small window; Maria gazed at the blur of codons, and mentally urged the process on—if not straight toward the target (since she had no idea what that was), then at least… outward, blindly, into the space of all possible mistakes.

It was a nice thought. The only trouble was, certain portions of the genes were especially prone to particular copying errors, so most of the mutants were “exploring” the same dead ends again and again.

Arranging for A. lamberti to mutate was easy; like a real-world bacterium, it made frequent errors every time it duplicated its analogue of DNA. Persuading it to mutate “usefully” was something else. Max Lambert himself—inventor of the Autoverse, creator of A. lamberti, hero to a generation of cellular-automaton and artificial-life freaks—had spent much of the last fifteen years of his life trying to discover why the subtle differences between real-world and Autoverse biochemistry made natural selection so common in one system, and so elusive in the other. Exposed to the kind of stressful opportunities which E. coli would have exploited within a few dozen generations, strain after strain of A. lamberti had simply died out.

Only a few die-hard enthusiasts still continued Lambert’s work. Maria knew of just seventy-two people who’d have the slightest idea what it meant if she ever succeeded. The artificial life scene, now, was dominated by the study of Copies—patchwork creatures, mosaics of ten thousand different ad hoc rules… the antithesis of everything the Autoverse stood for.

Real-world biochemistry was far too complex to simulate in every last detail for a creature the size of a gnat, let alone a human being. Computers could model all the processes of life—but not on every scale, from atom to organism, all at the same time. So the field had split three ways. In one camp, traditional molecular biochemists continued to extend their painstaking calculations, solving Schrödinger’s equation more or less exactly for ever larger systems, working their way up to entire replicating strands of DNA, whole mitochondrial sub-assemblies, significant patches of the giant carbohydrate chain-link fence of a cell wall… but spending ever more on computing power for ever diminishing returns.

At the other end of the scale were Copies: elaborate refinements of whole-body medical simulations, originally designed to help train surgeons with virtual operations, and to take the place of animals in drug tests. A Copy was like a high-resolution CAT scan come to life, linked to a medical encyclopedia to spell out how its every tissue and organ should behave… walking around inside a state-of-the-art architectural simulation. A Copy possessed no individual atoms or molecules; every organ in its virtual body came in the guise of specialized sub-programs which knew (in encyclopedic, but not atomic, detail) how a real liver or brain or thyroid gland functioned… but which couldn’t have solved Schrödinger’s equation for so much as a single protein molecule. All physiology, no physics.

Lambert and his followers had staked out the middle ground. They’d invented a new physics, simple enough to allow several thousand bacteria to fit into a modest computer simulation, with a consistent, unbroken hierarchy of details existing right down to the subatomic scale. Everything was driven from the bottom up, by the lowest level of physical laws, just as it was in the real world.

The price of this simplicity was that an Autoverse bacterium didn’t necessarily behave like its real-world counterparts. A. lamberti had a habit of confounding traditional expectations in bizarre and unpredictable ways—and for most serious microbiologists, that was enough to render it worthless.

For Autoverse junkies, though, that was the whole point.

Maria brushed aside the diagrams concealing her view of the Petri dishes, then zoomed in on one thriving culture, until a single bacterium filled the workspace. Color-coded by “health,” it was a featureless blue blob; but even when she switched to a standard chemical map there was no real structure visible, apart from the cell wall—no nucleus, no organelles, no flagella; A. lamberti wasn’t much more than a sac of protoplasm. She played with the representation, making the fine strands of the unraveled chromosomes appear; highlighting regions where protein synthesis was taking place; rendering visible the concentration gradients of nutrose and its immediate metabolites. Computationally expensive views; she cursed herself (as always) for wasting money, but failed (as always) to shut down everything but the essential analysis software (and the Autoverse itself), failed to sit gazing into thin air, waiting patiently for a result.

Instead, she zoomed in closer, switched to atomic colors (but left the pervasive aqua molecules invisible), temporarily halted time to freeze the blur of thermal motion, then zoomed in still further until the vague specks scattered throughout the workspace sharpened into the intricate tangles of long-chain lipids, polysaccharides, peptidoglycans. Names stolen unmodified from their real-world analogues—but screw it, who wanted to spend their life devising a whole new biochemical nomenclature? Maria was sufficiently impressed that Lambert had come up with distinguishable colors for all thirty-two Autoverse atoms, and unambiguous names to match.

She tracked through the sea of elaborate molecules—all of them synthesized by A. lamberti from nothing but nutrose, aqua, pneuma, and a few trace elements. Unable to spot any mutose molecules, she invoked Maxwell’s Demon and asked it to find one. The perceptible delay before the program responded always drove home to her the sheer quantity of information she was playing with—and the way in which it was organized. A traditional biochemical simulation would have been keeping track of every molecule, and could have told her the exact location of the nearest altered sugar almost instantaneously. For a traditional simulation, this catalogue of molecules would have been the “ultimate truth”—nothing would have “existed,” except by virtue of an entry in the Big List. In contrast, the “ultimate truth” of the Autoverse was a vast array of cubic cells of subatomic dimensions—and the primary software dealt only with these cells, oblivious to any larger structures. Atoms in the Autoverse were like hurricanes in an atmospheric model (only far more stable); they arose from the simple rules governing the smallest elements of the system. There was no need to explicitly calculate their behavior; the laws governing individual cells drove everything that happened at higher levels. Of course, a swarm of demons could have been used to compile and maintain a kind of census of atoms and molecules—at great computational expense, rather defeating the point. And the Autoverse itself would have churned on, regardless.

Maria locked her viewpoint to the mutose molecule, then restarted time, and everything but that one hexagonal ring smeared into translucence. The molecule itself was only slightly blurred; the current representational conventions made the average positions of the atoms clearly visible, with the deviations due to bond vibration merely suggested by faint ghostly streaks.

She zoomed in until the molecule filled the workspace. She didn’t know what she was hoping to see: a successful mutant epimerase enzyme suddenly latch onto the ring and shift the aberrant blue-red spike back into the horizontal position? Questions of probability aside, it would have been over before she even knew it had begun. That part was easily fixed: she instructed Maxwell’s Demon to keep a rolling buffer of a few million clock ticks of the molecule’s history, and to replay it at a suitable rate if any structural change occurred.

Embedded in a “living” organism, the mutose ring looked exactly the same as the prototype she’d handled minutes before: red, green and blue billiard balls, linked by thin white rods. It seemed like an insult for even a bacterium to be composed of such comic-book molecules. The viewing software was constantly inspecting this tiny region of the Autoverse, identifying the patterns that constituted atoms, checking for overlaps between them to decide which was bonded to which, and then displaying a nice, neat, stylized picture of its conclusions. Like the handling rules which took this representation at face value, it was a useful fiction, but…

Maria slowed down the Autoverse clock by a factor of ten billion, then popped up the viewing menu and hit the button marked RAW. The tidy assembly of spheres and rods melted into a jagged crown of writhing polychromatic liquid metal, waves of color boiling away from the vertices to collide, merge, flow back again, wisps licking out into space.

She slowed down time a further hundredfold, almost freezing the turmoil, and then zoomed in to the same degree. The individual cubic cells which made up the Autoverse were visible now, changing state about once a second. Each cell’s “state”—a whole number between zero and two hundred and fifty-five—was recomputed every clock cycle, according to a simple set of rules applied to its own previous state, and the states of its closest neighbors in the three-dimensional grid. The cellular automaton which was the Autoverse did nothing whatsoever but apply these rules uniformly to every cell; these were its fundamental “laws of physics.” Here, there were no daunting quantum-mechanical equations to struggle with—just a handful of trivial arithmetic operations, performed on integers. And yet the impossibly crude laws of the Autoverse still managed to give rise to “atoms” and “molecules” with a “chemistry” rich enough to sustain “life.”

Maria followed the fate of a cluster of golden cells spreading through the lattice—the cells themselves didn’t move, by definition, but the pattern advanced—infiltrating and conquering a region of metallic blue, only to be invaded and consumed in turn by a wave of magenta.

If the Autoverse had a “true” appearance, this was it. The palette which assigned a color to each state was still “false”—still completely arbitrary—but at least this view revealed the elaborate three-dimensional chess game which underpinned everything else.

Everything except the hardware, the computer itself.

Maria reverted to the standard clock rate, and a macroscopic view of her twenty-one Petri dishes—just as a message popped up in the foreground:

JSN regrets to advise you that your resources have been diverted to a higher bidder. A snapshot of your task has been preserved in mass storage, and will be available to you when you next log on. Thank you for using our services.

Maria sat and swore angrily for half a minute—then stopped abruptly, and buried her face in her hands. She shouldn’t have been logged on in the first place. It was insane, squandering her savings playing around with mutant A. lamberti—but she kept on doing it. The Autoverse was so seductive, so hypnotic… so addictive.

Whoever had elbowed her off the network had done her a favor—and she’d even have her fifty-dollar log-on fee refunded, since she’d been thrown right out, not merely slowed down to a snail’s pace.

Curious to discover the identity of her unintentional benefactor, she logged on directly to the QIPS Exchange—the marketplace where processing power was bought and sold. The connection to JSN had passed through the Exchange, transparently; her terminal was programmed to bid at the market rate automatically, up to a certain ceiling. Right now, though, some outfit calling itself Operation Butterfly was buying QIPS—quadrillions of instructions per second—at six hundred times that ceiling, and had managed to acquire one hundred percent of the planet’s traded computing power.

Maria was stunned; she’d never seen anything like it. The pie chart of successful bidders—normally a flickering kaleidoscope of thousands of needle-thin slices—was a solid, static disk of blue. Aircraft would not be dropping out of the sky, world commerce would not have ground to a halt… but tens of thousands of academic and industrial researchers relied on the Exchange every day for tasks it wasn’t worth owning the power to perform in-house. Not to mention a few thousand Copies. For one user to muscle in and outbid everyone else was unprecedented. Who needed that much computing power? Big business, big science, the military? All had their own private hardware—usually in excess of their requirements. If they traded at all, it was to sell their surplus capacity.

Operation Butterfly? The name sounded vaguely familiar. Maria logged on to a news system and searched for reports which mentioned the phrase. The most recent was three months ago:

Kuala Lumpar—Monday, August 8th, 2050: A meeting of environmental ministers from the Association of South-East Asian Nations (ASEAN) today agreed to proceed with the latest stage of Operation Butterfly, a controversial plan to attempt to limit the damage and loss of life caused by Greenhouse Typhoons in the region.

The long-term aim of the project is to utilize the so-called Butterfly Effect to divert typhoons away from vulnerable populated areas—or perhaps prevent them from forming in the first place.

Maria said, “Define ‘Butterfly Effect.’” A second window opened up in front of the news report:

Butterfly Effect: This term was coined by meteorologist Edward Lorenz in the late 1970s, to dramatize the futility of trying to make long-term weather forecasts. Lorenz pointed out that meteorological systems were so sensitive to their initial conditions that a butterfly flapping its wings in Brazil could be enough to determine whether or not there was a tornado in Texas a month later. No computer model could ever include such minute details—so any attempt to forecast the weather more than a few days in advance was doomed to failure.

However, in the 1990s the term began to lose its original, pessimistic connotations. A number of researchers discovered that, although the effects of small, random influences made a chaotic system unpredictable, under certain conditions the same sensitivity could be deliberately exploited to steer the system in a chosen direction. The same kind of processes which magnified the flapping of butterflies’ wings into tornadoes could also magnify the effects of systematic intervention, allowing a degree of control out of all proportion to the energy expended.

The Butterfly Effect now commonly refers to the principle of controlling a chaotic system with minimum force, through a detailed knowledge of its dynamics. This technique has been applied in a number of fields, including chemical engineering, stock-market manipulation, fly-by-wire aeronautics, and the proposed ASEAN weather-control system, Operation Butterfly.

There was more, but Maria took the cue and switched back to the article.

Meteorologists envisage dotting the waters of the tropical western Pacific and the South China Sea with a grid of hundreds of thousands of “weather-control” rigs—solar-powered devices designed to alter the local temperature on demand by pumping water between different depths. Theoretical models suggest that a sufficient number of rigs, under elaborate computer control, could be used to influence large-scale weather patterns, “nudging” them toward the least harmful of a number of finely balanced possible outcomes.

Eight different rig prototypes have been tested in the open ocean, but before engineers select one design for mass production, an extensive feasibility study will be conducted. Over a three-year period, any potentially threatening typhoon will be analyzed by a computer model of the highest possible resolution, and the effects of various numbers and types of the as yet nonexistent rigs will be included in the model. If these simulations demonstrate that intervention could have yielded significant savings in life and property, ASEAN’s ministerial council will have to decide whether or not to spend the estimated sixty billion dollars required to make the system a reality. Other nations are observing the experiment with interest.

Maria leaned back from the screen, impressed. A computer model of the highest possible resolution. And they’d meant it, literally. They’d bought up all the number-crunching power on offer—paying a small fortune, but only a fraction of what it would have cost to buy the same hardware outright.

Nudging typhoons! Not yet, not in reality… but who could begrudge Operation Butterfly their brief monopoly, for such a grand experiment? Maria felt a vicarious thrill at the sheer scale of the endeavor—and then a mixture of guilt and resentment at being a mere bystander. She had no qualifications in atmospheric or oceanic physics, no PhD in chaos theory—but in a project of that size, there must have been a few hundred jobs offered to mere programmers. When the tenders had gone out over the network, she’d probably been busy on some shitty contract to improve the tactile qualities of beach sand for visitors to the Virtual Gold Coast—either that, or tinkering with the genome of A. lamberti, trying to become the first person in the world to bludgeon a simulated bacterium into exhibiting natural selection.

It wasn’t clear how long Operation Butterfly would spend monitoring each typhoon—but she could forget about returning to the Autoverse for the day.

Reluctantly, she logged off the news system—fighting the temptation to sit and wait for the first reports of the typhoon in question, or the response of other supercomputer users to the great processing buy-out—and began reviewing her plans for a new intruder surveillance package.

2 (Remit not paucity) NOVEMBER 2050

“What I’m asking for is two million ecus. What I’m offering you is immortality.”

Thomas Riemann’s office was compact but uncluttered, smartly furnished without being ostentatious. The single large window offered a sweeping view of Frankfurt—looking north across the river, as if from Sachsenhausen, toward the three jet-black towers of the Siemens/Deutsche Bank Center—which Thomas believed was as honest as any conceivable alternative. Half the offices in Frankfurt itself looked out over recorded tropical rainforests, stunning desert gorges, Antarctic ice shelves—or wholly synthetic landscapes: rural-idyllic, futuristic, interplanetary, or simply surreal. With the freedom to choose whatever he liked, he’d selected this familiar sight from his corporeal days; sentimental, perhaps, but at least it wasn’t ludicrously inappropriate.

Thomas turned away from the window, and regarded his visitor with good-natured skepticism. He replied in English; the office software could have translated for him—and would have chosen the very same words and syntax, having been cloned from his own language centers—but Thomas still preferred to use the version “residing inside” his own “skull.”

“Two million? What’s the scheme? Let me guess. Under your skillful management, my capital will grow at the highest possible rate consistent with the need for total security. The price of computation is sure to fall again, sooner or later; the fact that it’s risen for the last fifteen years only makes that more likely than ever. So: it may take a decade or two—or three, or four—but eventually, the income from my modest investment will be enough to keep me running on the latest hardware, indefinitely… while also providing you with a small commission, of course.” Thomas laughed, without malice. “You don’t seem to have researched your prospective client very thoroughly. You people usually have immaculate intelligence—but I’m afraid you’ve really missed the target with me. I’m in no danger of being shut down. The hardware we’re using, right now, isn’t leased from anyone; it’s wholly owned by a foundation I set up before my death. My estate is being managed to my complete satisfaction. I have no problems—financial, legal, peace of mind—for you to solve. And the last thing in the world I need is a cheap and nasty perpetuity fund. Your offer is useless to me.”

Paul Durham chose to display no sign of disappointment. He said, “I’m not talking about a perpetuity fund. I’m not selling any kind of financial service. Will you give me a chance to explain?”

Thomas nodded affably. “Go ahead. I’m listening.” Durham had flatly refused to state his business in advance, but Thomas had decided to see him anyway—anticipating a perverse satisfaction in confirming that the man’s mysterious coyness hid nothing out of the ordinary. Thomas almost always agreed to meet visitors from outside—even though experience had shown that most were simply begging for money, one way or another. He believed that anyone willing to slow down their brain by a factor of seventeen, solely for the privilege of talking to him face to face, deserved a hearing—and he wasn’t immune to the intrinsic flattery of the process, the unequal sacrifice of time.

There was more to it, though, than flattery.

When other Copies called on him in his office, or sat beside him at a boardroom table, everyone was “present” in exactly the same sense. However bizarre the algorithmic underpinnings of the encounter, it was a meeting of equals. No boundaries were crossed.

A visitor, though, who could lift and empty a coffee cup, who could sign a document and shake your hand—but who was, indisputably, lying motionless on a couch in another (higher?) metaphysical plane—came charged with too many implicit reminders of the nature of things to be faced with the same equanimity. Thomas valued that. He didn’t want to grow complacent—or worse. Visitors helped him to retain a clear sense of what he’d become.

Durham said, “Of course I’m aware of your situation—you have one of the most secure arrangements I’ve seen. I’ve read the incorporation documents of the Soliton Foundation, and they’re close to watertight. Under present legislation.”

Thomas laughed heartily. “But you think you can do better? Soliton pays its most senior lawyers almost a million a year; you should have got yourself some forged qualifications and asked me to employ you. Under present legislation! When the laws change, believe me, they’ll change for the better. I expect you know that Soliton spends a small fortune lobbying for improvement—and it’s not alone. The trend is in one direction: there are more Copies every year, and most of them have de facto control over virtually all of the wealth they owned when they were alive. I’m afraid your timing’s atrocious if you’re planning on using scare tactics; I received a report last week predicting full human rights—in Europe, at least—by the early sixties. Ten years isn’t long for me to wait. I’ve grown used to the current slowdown factor; even if processor speeds improve, I could easily choose to keep living at the rate I’m living now, for another six or seven subjective months, rather than pushing all the things I’m looking forward to—like European citizenship—further into the future.”

Durham’s puppet inclined its head in a gesture of polite assent; Thomas had a sudden vision of a second puppet—one Durham truly felt himself to be inhabiting—hunched over a control panel, hitting a button on an etiquette sub-menu. Was that paranoid? But any sensible mendicant visitor would do just that, conducting the meeting at a distance rather than exposing their true body language to scrutiny.

The visible puppet said, “Why spend a fortune upgrading, for the sake of effectively slowing down progress? And I agree with you about the outlook for reform—in the short term. Of course people begrudge Copies their longevity, but the PR has been handled remarkably well. A few carefully chosen terminally ill children are scanned and resurrected every year: better than a trip to Disney World. There’s discreet sponsorship of a sitcom about working-class Copies, which makes the whole idea less threatening. The legal status of Copies is being framed as a human rights issue, especially in Europe: Copies are disabled people, no more, no less—really just a kind of radical amputee—and anyone who talks about decadent rich immortals getting their hands on all the wealth is shouted down as a neo-Nazi.

“So you might well achieve citizenship in a decade. And if you’re lucky, the situation could be stable for another twenty or thirty years after that. But… what’s twenty or thirty years to you? Do you honestly think that the status quo will be tolerated for ever?’”

Thomas said, “Of course not—but I’ll tell you what would be “tolerated”: scanning facilities, and computing power, so cheap that everyone on the planet could be resurrected. Everyone who wanted it. And when I say cheap, I mean at a cost comparable to a dose of vaccine at the turn of the century. Imagine that. Death could be eradicated—like smallpox or malaria. And I’m not talking about some solipsistic nightmare; by then, telepresence robots will let Copies interact with the physical world as fully as if they were human. Civilization wouldn’t have deserted reality—just transcended biology.”

“That’s a long, long way in the future.”

“Certainly. But don’t accuse me of thinking in the short term.”

“And in the meantime? The privileged class of Copies will grow larger, more powerful—and more threatening to the vast majority of people, who still won’t be able to join them. The costs will come down, but not drastically—just enough to meet some of the explosion in demand from the executive class, once they throw off their qualms, en masse. Even in secular Europe, there’s a deeply ingrained prejudice that says dying is the responsible, the moral thing to do. There’s a Death Ethic—and the first substantial segment of the population abandoning it will trigger a huge backlash. A small enough elite of giga-rich Copies is accepted as a freak show; tycoons can get away with anything, they’re not expected to act like ordinary people. But just wait until the numbers go up by a factor of ten.”

Thomas had heard it all before. “We may be unpopular for a while. I can live with that. But you know, even now we’re vilified far less than people who strive for organic hyper-longevity—transplants, cellular rejuvenation, whatever—because at least we’re no longer pushing up the cost of health care, competing for the use of overburdened medical facilities. Nor are we consuming natural resources at anything like the rate we did when we were alive. If the technology improves sufficiently, the environmental impact of the wealthiest Copy could end up being less than that of the most ascetic living human. Who’ll have the high moral ground then? We’ll be the most ecologically sound people on the planet.”

Durham smiled. The puppet. “Sure—and it could lead to some nice ironies if it ever came true. But even low environmental impact might not seem so saintly, when the same computing power could be used to save tens of thousands of lives through weather control.”

“Operation Butterfly has inconvenienced some of my fellow Copies very slightly. And myself not at all.”

“Operation Butterfly is only the beginning. Crisis management, for a tiny part of the planet. Imagine how much computing power it would take to render sub-Saharan Africa free from drought.”

“Why should I imagine that, when the most modest schemes are still unproven? And even if weather control turns out to be viable, more supercomputers can always be built. It doesn’t have to be a matter of Copies versus flood victims.”

“There’s a limited supply of computing power right now, isn’t there? Of course it will grow—but the demand, from Copies, and for weather control, is almost certain to grow faster. Long before we get to your deathless Utopia, we’ll hit a bottle-neck—and I believe that will bring on a time when Copies are declared illegal. Worldwide. If they’ve been granted human rights, those rights will be taken away. Trusts and foundations will have their assets confiscated. Supercomputers will be heavily policed. Scanners—and scan files—will be destroyed. It may be forty years before any of this happens—or it may be sooner. Either way, you need to be prepared.”

Thomas said mildly, “If you’re fishing for a job as a futurology consultant, I’m afraid I already employ several—highly qualified—people who do nothing but investigate these trends. Right now, everything they tell me gives me reason to be optimistic—and even if they’re wrong, Soliton is ready for a very wide range of contingencies.”

“If your whole foundation is eviscerated, do you honestly believe it will be able to ensure that a snapshot of you is hidden away safely—and then resurrected after a hundred years or more of social upheaval? A vault full of ROM chips at the bottom of a mine shaft could end up taking a one-way trip into geological time.”

Thomas laughed. “And a meteor could hit the planet tomorrow, wiping out this computer, all of my backups, your organic body… anything and everything. Yes, there could be a revolution which pulls the plug on my world. It’s unlikely, but it’s not impossible. Or there could be a plague, or an ecological disaster, which kills billions of organic humans but leaves all the Copies untouched. There are no certainties for anyone.”

“But Copies have so much more to lose.”

Thomas was emphatic; this was part of his personal litany. “I’ve never mistaken what I have—a very good chance of a prolonged existence—for a guarantee of immortality.”

Durham said flatly, “Quite right. You have no such thing. Which is why I’m here offering it to you.”

Thomas regarded him uneasily. Although he’d had all the ravages of surgery edited out of his final scan file, he’d kept a scar on his right forearm, a small memento of a youthful misadventure. He stroked it, not quite absentmindedly; conscious of the habit, conscious of the memories that the scar encoded—but practiced at refusing to allow those memories to hold his gaze.

Finally, he said, “Offering it how? What can you possibly do—for two million ecus—that Soliton can’t do a thousand times better?”

“I can run a second version of you, entirely out of harm’s way. I can give you a kind of insurance—against an anti-Copy backlash… or a meteor strike… or whatever else might go wrong.”

Thomas was momentarily speechless. The subject wasn’t entirely taboo, but he couldn’t recall anyone raising it quite so bluntly before. He recovered swiftly. “I have no wish to run a second version, thank you. And… what do you mean, “out of harm’s way”? Where’s your invulnerable computer going to be? In orbit? Up where it would only take a pebble-sized meteor to destroy it, instead of a boulder?”

“No, not in orbit. And if you don’t want a second version, that’s fine. You could simply move.”

“Move where? Underground? To the bottom of the ocean? You don’t even know where this office is being implemented, do you? What makes you think you can offer a superior site—for such a ridiculous price—when you don’t have the faintest idea how secure I am already?” Thomas was growing disappointed, and uncharacteristically irritable. “Stop making these inflated claims, and get to the point. What are you selling?”

Durham shook his head apologetically. “I can’t tell you that. Not yet. If I tried to explain it, out of the blue, it would make no sense. You have to do something first. Something very simple.”

“Yes? And what’s that?”

“You have to conduct a small experiment.”

Thomas scowled. “What kind of experiment? Why?”

And Durham—the software puppet, the lifeless shell animated by a being from another plane—looked him in the eye and said, “You have to let me show you exactly what you are.”

3 (Rip, tie, cut toy man) JUNE 2045

Paul—or the flesh-and-blood man whose memories he’d inherited—had traced the history of Copies back to the turn of the century, when researchers had begun to fine-tune the generic computer models used for surgical training and pharmacology, transforming them into customized versions able to predict the needs and problems of individual patients. Drug therapies were tried out in advance on models which incorporated specific genetic and biochemical traits, allowing doses to be optimized and any idiosyncratic side-effects anticipated and avoided. Elaborate operations were rehearsed and perfected in Virtual Reality, on software bodies with anatomical details—down to the finest capillaries—based on the flesh-and-blood patient’s tomographic scans.

These early models included a crude approximation of the brain, perfectly adequate for heart surgery or immunotherapy—and even useful to a degree when dealing with gross cerebral injuries and tumours—but worthless for exploring more subtle neurological problems.

Imaging technology steadily improved, though—and by 2020, it had reached the point where individual neurons could be mapped, and the properties of individual synapses measured, non-invasively. With a combination of scanners, every psychologically relevant detail of the brain could be read from the living organ—and duplicated on a sufficiently powerful computer.

At first, only isolated neural pathways were modeled: portions of the visual cortex of interest to designers of machine vision, or sections of the limbic system whose role had been in dispute. These fragmentary neural models yielded valuable results, but a functionally complete representation of the whole organ—embedded in a whole body—would have allowed the most delicate feats of neurosurgery and psychopharmacology to be tested in advance. For several years, though, no such model was built—in part, because of a scarcely articulated unease at the prospect of what it would mean. There were no formal barriers standing in the way—government regulatory bodies and institutional ethics committees were concerned only with human and animal welfare, and no laboratory had yet been fire-bombed by activists for its inhumane treatment of physiological software—but still, someone had to be the first to break all the unspoken taboos.

Someone had to make a high-resolution, whole-brain Copy—and let it wake, and talk.

In 2024, John Vines, a Boston neurosurgeon, ran a fully conscious Copy of himself in a crude Virtual Reality. Taking slightly less than three hours of real time (pulse racing, hyper-ventilating, stress hormones elevated), the first Copy’s first words were: “This is like being buried alive. I’ve changed my mind. Get me out of here.”

His original obligingly shut him down—but then later repeated the demonstration several times, without variation, reasoning that it was impossible to cause additional distress by running exactly the same simulation more than once.

When Vines went public, the prospects for advancing neurological research didn’t rate a mention; within twenty-four hours—despite the Copy’s discouraging testimony—the headlines were all immortality, mass migration into Virtual Reality, and the imminent desertion of the physical world.

Paul was twenty-four years old at the time, with no idea what to make of his life. His father had died the year before—leaving him a modest business empire, centered on a thriving retail chain, which he had no interest in managing. He’d spent seven years traveling and studying—science, history and philosophy—doing well enough at everything he tried, but unable to discover anything that kindled real intellectual passion. With no struggle for financial security ahead, he’d been sinking quietly into a state of bemused complacency.

The news of John Vines’s Copy blasted away his indifference. It was as if every dubious promise technology had ever made to transform human life was about to be fulfilled, with a vengeance. Longevity would only be the start of it; Copies could evolve in ways almost impossible for organic beings: modifying their minds, redefining their goals, endlessly transmuting themselves. The possibilities were intoxicating—even as the costs and drawbacks of the earliest versions sank in, even as the inevitable backlash began, Paul was a child of the millennium; he was ready to embrace it all.

But the more time he spent contemplating what Vines had done, the more bizarre the implications seemed to be.

The public debate the experiment had triggered was heated, but depressingly superficial. Decades-old arguments raged again over just how much computer programs could ever have in common with human beings (psychologically, morally, metaphysically, information-theoretically… ) and even whether or not Copies could be “truly” intelligent, “truly” conscious. As more workers repeated Vines’s result, their Copies soon passed the Turing test: no panel of experts quizzing a group of Copies and humans—by delayed video, to mask the time-rate difference—could tell which were which. But some philosophers and psychologists continued to insist that this demonstrated nothing more than “simulated consciousness,” and that Copies were merely programs capable of faking a detailed inner life which didn’t actually exist at all.

Supporters of the Strong AI Hypothesis insisted that consciousness was a property of certain algorithms—a result of information being processed in certain ways, regardless of what machine, or organ, was used to perform the task. A computer model which manipulated data about itself and its “surroundings” in essentially the same way as an organic brain would have to possess essentially the same mental states. “Simulated consciousness” was as oxymoronic as “simulated addition.”

Opponents replied that when you modeled a hurricane, nobody got wet. When you modeled a fusion power plant, no energy was produced. When you modeled digestion and metabolism, no nutrients were consumed—no real digestion took place. So, when you modeled the human brain, why should you expect real thought to occur? A computer running a Copy might be able to generate plausible descriptions of human behavior in hypothetical scenarios—and even appear to carry on a conversation, by correctly predicting what a human would have done in the same situation—but that hardly made the machine itself conscious.

Paul had rapidly decided that this whole debate was a distraction. For any human, absolute proof of a Copy’s sentience was impossible. For any Copy, the truth was self-evident: cogito ergo sum. End of discussion.

But for any human willing to grant Copies the same reasonable presumption of consciousness that they granted their fellow humans—and any Copy willing to reciprocate—the real point was this:

There were questions about the nature of this shared condition which the existence of Copies illuminated more starkly than anything which had come before them. Questions which needed to be explored, before the human race could confidently begin to bequeath its culture, its memories, its purpose and identity, to its successors.

Questions which only a Copy could answer.

+ + +

Paul sat in his study, in his favorite armchair (unconvinced that the texture of the surface had been accurately reproduced), taking what comfort he could from the undeniable absurdity of being afraid to experiment on himself further. He’d already “survived” the “transition” from flesh-and-blood human to computerized physiological model—the most radical stage of the project, by far. In comparison, tinkering with a few of the model’s parameters should have seemed trivial.

Durham appeared on the terminal—which was otherwise still dysfunctional. Paul was already beginning to think of him as a bossy little djinn trapped inside the screen—rather than a vast, omnipotent deity striding the halls of Reality, pulling all the strings. The pitch of his voice was enough to deflate any aura of power and grandeur.

Squeak. “Experiment one, trial zero. Baseline data. Time resolution one millisecond—system standard. Just count to ten, at one-second intervals, as near as you can judge it. Okay?”

“I think I can manage that.” He’d planned all this himself, he didn’t need step-by-step instructions. Durham’s image vanished; during the experiments, there could be no cues from real time.

Paul counted to ten. The djinn returned. Staring at the face on the screen, Paul realized that he had no inclination to think of it as “his own.” Perhaps that was a legacy of distancing himself from the earlier Copies. Or perhaps his mental image of himself had never been much like his true appearance—and now, in defense of sanity, was moving even further away.

Squeak. “Okay. Experiment one, trial number one. Time resolution five milliseconds. Are you ready?”

“Yes.”

The djinn vanished. Paul counted: “One. Two. Three. Four. Five. Six. Seven. Eight. Nine. Ten.”

Squeak. “Anything to report?”

“No. I mean, I can’t help feeling slightly apprehensive, just knowing that you’re screwing around with my… infrastructure. But apart from that, nothing.”

Durham’s eyes no longer glazed over while he was waiting for the speeded-up reply; either he’d gained a degree of self-discipline, or—more likely—he’d interposed some smart editing software to conceal his boredom.

Squeak. “Don’t worry about apprehension. We’re running a control, remember?”

Paul would have preferred not to have been reminded. He’d known that Durham must have cloned him, and would be feeding exactly the same sensorium to both Copies—while only making changes in the model’s time resolution for one of them. It was an essential part of the experiment—but he didn’t want to dwell on it. A third self, shadowing his thoughts, was too much to acknowledge on top of everything else.

Squeak. “Trial number two. Time resolution ten milliseconds.”

Paul counted. The easiest thing in the world, he thought, when you’re made of flesh, when you’re made of matter, when the quarks and the electrons just do what comes naturally. Human beings were embodied, ultimately, in fields of fundamental particles—incapable, surely, of being anything other than themselves. Copies were embodied in computer memories as vast sets of numbers. Numbers which certainly could be interpreted as describing a human body sitting in a room… but it was hard to see that meaning as intrinsic, as necessary, when tens of thousands of arbitrary choices had been made about the way in which the model had been coded. Is this my blood sugar here… or my testosterone level? Is this the firing rate of a motor neuron as I raise my right hand… or a signal coming in from my retina as I watch myself doing it? Anybody given access to the raw data, but unaware of the conventions, could spend a lifetime sifting through the numbers without deciphering what any of it meant.

And yet no Copy buried in the data itself—ignorant of the details or not—could have the slightest trouble making sense of it all in an instant.

Squeak. “Trial number three. Time resolution twenty milliseconds.”

“One. Two. Three.”

For time to pass for a Copy, the numbers which defined it had to change from moment to moment. Recomputed over and over again, a Copy was a sequence of snapshots, frames of a movie—or frames of computer animation.

But… when, exactly, did these snapshots give rise to conscious thought? While they were being computed? Or in the brief interludes when they sat in the computer’s memory, unchanging, doing nothing but representing one static instant of the Copy’s life? When both stages were taking place a thousand times per subjective second, it hardly seemed to matter, but very soon—

Squeak. “Trial number four. Time resolution fifty milliseconds.”

What am I? The data? The process that generates it? The relationships between the numbers?

All of the above?

“One hundred milliseconds.”

“One. Two. Three.”

Paul listened to his voice as he counted—as if half expecting to begin to notice the encroachment of silence, to start perceiving the gaps in himself.

“Two hundred milliseconds.”

A fifth of a second. “One. Two.” Was he strobing in and out of existence now, at five subjective hertz? The crudest of celluloid movies had never flickered at this rate. “Three. Four.” He waved his hand in front of his face; the motion looked perfectly smooth, perfectly normal. And of course it did; he wasn’t watching from the outside. “Five. Six. Seven.” A sudden, intense wave of nausea passed through him but he fought it down, and continued. “Eight. Nine. Ten.”

The djinn reappeared and emitted a brief, solicitous squeak. “What’s wrong? Do you want to stop for a while?”

“No, I’m fine.” Paul glanced around the innocent, sun-dappled room, and laughed. How would Durham handle it if the control and the subject had just given two different replies? He tried to recall his plans for such a contingency, but couldn’t remember them—and didn’t much care. It wasn’t his problem any more.

Squeak. “Trial number seven. Time resolution five hundred milliseconds.”

Paul counted—and the truth was, he felt no different. A little uneasy, yes—but factoring out any squeamishness, everything about his experience seemed to remain the same. And that made sense, at least in the long run—because nothing was being omitted, in the long run. His model-of-a-brain was only being fully described at half-second (model time) intervals—but each description still included the results of everything that “would have happened” in between. Every half-second, his brain was ending up in exactly the state it would have been in if nothing had been left out.

“One thousand milliseconds.”

But… what was going on, in between? The equations controlling the model were far too complex to solve in a single step. In the process of calculating the solutions, vast arrays of partial results were being generated and discarded along the way. In a sense, these partial results implied—even if they didn’t directly represent—events taking place within the gaps between successive complete descriptions. And when the whole model was arbitrary, who was to say that these implied events, buried a little more deeply in the torrent of data, were any “less real” than those which were directly described?

“Two thousand milliseconds.”

“One. Two. Three. Four.”

If he seemed to speak (and hear himself speak) every number, it was because the effects of having said “three” (and having heard himself say it) were implicit in the details of calculating how his brain evolved from the time when he’d just said “two” to the time when he’d just said “four.”

“Five thousand milliseconds.”

“One. Two. Three. Four. Five.”

Besides, hearing words that he’d never “really” spoken wasn’t much stranger than a Copy hearing anything at all. Even the standard millisecond clock rate of this world was far too coarse to resolve the full range of audible tones. Sound wasn’t represented in the model by fluctuations in air pressure values—which couldn’t change fast enough—but in terms of audio power spectra: profiles of intensity versus frequency. Twenty kilohertz was just a number here, a label; nothing could actually oscillate at that rate. Real ears analyzed pressure waves into components of various pitch; Paul knew that his brain was being fed the preexisting power spectrum values directly, plucked out of the nonexistent air by a crude patch in the model.

“Ten thousand milliseconds.”

“One. Two. Three.”

Ten seconds free-falling from frame to frame.

Fighting down vertigo, still counting steadily, Paul prodded the shallow cut he’d made in his forearm with the kitchen knife. It stung, convincingly. So where was this experience coming from? Once the ten seconds were up, his fully described brain would remember all of this… but that didn’t account for what was happening now. Pain was more than the memory of pain. He struggled to imagine the tangle of billions of intermediate calculations, somehow “making sense” of themselves, bridging the gap.

And he wondered: What would happen if someone shut down the computer, just pulled the plug—right now?

He didn’t know what that meant, though. In any terms but his own, he didn’t know when “right now” was.

“Eight. Nine. Ten.”

Squeak. “Paul—I’m seeing a slight blood pressure drop. Are you okay? How are you feeling?”

Giddy—but he said, “The same as always.” And if that wasn’t quite true, no doubt the control had told the same lie. Assuming…

“Tell me—which was I? Control, or subject?”

Squeak. Durham replied, “I can’t answer that—I’m still speaking to both of you. I’ll tell you one thing, though: the two of you are still identical. There were some very small, transitory discrepancies, but they’ve died away completely now—and whenever the two of you were in comparable representations, all firing patterns of more than a couple of neurons were the same.”

Paul grunted dismissively; he had no intention of letting Durham know how unsettling the experiment had been. “What did you expect? Solve the same set of equations two different ways, and of course you get the same results—give or take some minor differences in round-off errors along the way. You must. It’s a mathematical certainty.”

Squeak. “Oh, I agree.” The djinn wrote with one finger on the screen:

(1 + 2) + 3 = 1 + (2 + 3)

Paul said, “So why bother with this stage at all? I know—I wanted to be rigorous, I wanted to establish solid foundations. But the truth is, it’s a waste of our resources. Why not skip the bleeding obvious, and get on with the kind of experiment where the answer isn’t a foregone conclusion?”

Squeak. Durham frowned reprovingly. “I didn’t realize you’d grown so cynical so quickly. AI isn’t a branch of pure mathematics; it’s an empirical science. Assumptions have to be tested. Confirming the so-called “obvious” isn’t such a dishonourable thing, is it? And if it’s all so straightforward, why should you be afraid?”

“I’m not afraid: I just want to get it over with. But… go ahead. Prove whatever you think you have to prove, and then we can move on.”

Squeak. “That’s the plan. But I think we could both use a break now. I’ll enable your communications—for incoming data only.” He turned away, reached off-screen, and hit a few keys on a second terminal.

Then he turned back to the camera, smiling—and Paul knew exactly what he was going to say.

Squeak. “By the way, I just deleted one of you. I couldn’t afford to keep you both running, when all you’re going to do is laze around.”

Paul smiled back at him, although something inside him was screaming. “Which one did you terminate?”

Squeak. “What difference does it make? I told you, they were identical. And you’re still here, aren’t you? Whoever you are. Whichever you were.”

+ + +

Three weeks had passed outside since the day of the scan, but it didn’t take Paul long to catch up with the state of the world; most of the fine details had been rendered irrelevant by subsequent events, and much of the ebb and flow had simply canceled itself out. Israel and Palestine had come close to war again, over alleged water treaty violations on both sides—but a joint peace rally had brought more than a million people onto the glassy plain that used to be Jerusalem, and the two governments had been forced to back down. Former US President Martin Sandover was still fighting extradition to Palau, to face charges arising from his role in the bloody coup d’etat of thirty-five; the Supreme Court had finally reversed a long-standing ruling which had granted him immunity from all foreign laws, and for a day or two things had looked promising—but then his legal team had discovered a whole new set of delaying tactics. In Canberra, another leadership challenge had come and gone, with the Prime Minister remaining undeposed. In a week-old report, one journalist described this, straight-faced, as “high drama.” Paul thought: I guess you had to be there. Inflation had fallen by half a per-centage point; unemployment had risen by the same amount.

Paul scanned the old news reports rapidly, skimming over articles and fast-forwarding scenes which he felt sure he would have studied scrupulously, had they been fresh. He felt a curious sense of resentment, at having “missed” so much—it was all there in front of him, now, but that wasn’t the same at all.

And yet, he wondered, shouldn’t he be relieved that he hadn’t wasted his time on so much ephemeral detail? The very fact that he was now less than enthralled only proved how little of it had really mattered, in the long run.

Then again, what did? People didn’t inhabit geological time. People inhabited hours and days; they had to care about things on that time scale.

People.

Paul plugged into real-time TV, and watched an episode of The Unclear Family flash by in less than two minutes, the soundtrack an incomprehensible squeal. A game show. A war movie. The evening news. It was as if he was in deep space, rushing back toward the Earth through a sea of Doppler-shifted broadcasts. The image was strangely comforting; his situation wasn’t so bizarre, after all, if flesh-and-blood humans could find themselves in much the same relationship with the world as he did. Nobody would claim that the Doppler shift could rob someone of their humanity.

Dusk fell over the recorded city. He ate a microwaved soya protein stew—wondering if there was any good reason, moral or otherwise, to continue to be a vegetarian.

He listened to music until long after midnight. Tsang Chao, Michael Nyman, Philip Glass. It made no difference that each note “really” lasted seventeen times as long as it should have, or that the audio ROM sitting in the player “really” possessed no microstructure, or that the “sound” itself was being fed into his model-of-a-brain by a computerized sleight-of-hand that bore no resemblance to the ordinary process of hearing. The climax of Glass’s Mishima still seized him like a grappling hook through the heart.

And if the computations behind all this had been performed over millennia, by people flicking abacus beads, would he have felt exactly the same?

It was outrageous to admit it—but the answer had to be yes.

He lay in bed, wondering: Do I still want to wake from this dream?

The question remained academic, though; he still had no choice.

4 (Remit not paucity) NOVEMBER 2050

Maria had arranged to meet Aden at the Nadir, an Oxford Street nightclub where he sometimes played and often went to write. He could usually get them both in for free, and the door—an intimidating, airlock-like contraption of ribbed black anodized steel—let her pass unchallenged after a brief security scan. Maria had once had a nightmare in which she’d been trapped in that chamber, a knife inexplicably strapped to her right boot—and, worse, her credit rating canceled. The thing had digested her like an insect in a Venus flytrap, while Aden stood on stage, singing one of his cut-up love songs.

Inside, the place was crowded for a Thursday night, and poorly lit as always; she finally spotted Aden sitting at a table near a side wall, listening to one of the bands and jotting down music, his face catching the glow of his notepad. So far as Maria could tell, he never seemed to be unduly influenced by anything he listened to while composing, but he claimed to be unable to work in silence, and preferred live performances for inspiration—or catalysis, or whatever it was.

She touched him on the shoulder. He looked up, took off his headset, and stood to kiss her. He tasted of orange juice.

He gestured with the headset. “You should listen. Crooked Buddhist Lawyers on Crack. They’re quite good.”

Maria glanced at the stage, although there was no way of telling who he meant. A dozen performers—four bands in all—stood enclosed in individual soundproof plastic cylinders. Most of the patrons were tuned in, wearing headsets to pick up one band’s sound, and liquid crystal shades, flickering in synch with one group of cylinders, to render the other bands invisible. A few people were chatting quietly—and of the room’s five possible soundtracks, Maria decided that this tranquil near-silence best suited her mood. Besides, she never much liked using nerve current inducers; although physically unable to damage the eardrums (sparing the management any risk of litigation), they always seemed to leave her ears—or her auditory pathways—ringing, regardless of the volume setting she chose.

“Maybe later.”

She sat beside Aden, and felt him tense slightly when their shoulders brushed, then force himself to relax. Or maybe not. Often when she thought she was reading his body language, she was making signals out of noise. She said, “I got some junk mail today that looked just like you.”

“How flattering. I think. What was it selling?”

“The Church of the God Who Makes No Difference.”

He laughed. “Every time I hear that, I think: they’ve got to change the name. A God which makes no difference doesn’t rate the definite article or the pronoun ‘who.’”

“I’ll rerun the program, and the two of you can fight it out.”

“No thanks.” He took a sip of his drink. “Any non-junk mail? Any contracts?”

“No.”

“So… another day of terminal boredom?”

“Mostly.” Maria hesitated. Aden usually only pressed her for news when he had something to announce himself—and she was curious to find out what it was. But he volunteered nothing, so she went on to describe her encounter with Operation Butterfly.

Aden said, “I remember hearing something about that. But I thought it was decades away.”

“The real thing probably is, but the simulations have definitely started. In a big way.”

He looked pained. “Weather control? Who do they think they’re kidding?”

Maria suppressed her irritation. “The theory must look promising, or they wouldn’t have taken it this far. Nobody spends a few million dollars an hour on supercomputer time without a good chance of a payoff.”

Aden snickered. “Oh yes they do. And it’s usually called Operation something-or-other. Remember Operation Radiant Way?”

“Yes, I remember.”

“They were going to seed the upper atmosphere with nanomachines which could monitor the temperature—and supposedly do something about it.”

“Manufacture particles which reflected certain wavelengths of solar radiation—and then disassemble them, as required.”

“In other words, cover the planet with a giant thermostatic blanket.”

“What’s so terrible about that?”

“You mean, apart from the sheer technocratic hubris? And apart from the fact that releasing any kind of replicator into the environment is—still, thankfully—illegal? It wouldn’t have worked. There were complications nobody had predicted—unstable mixing of air layers, wasn’t it?—which would have counteracted most of the effect.”

Maria said, “Exactly. But how would anyone have known that, if they hadn’t run a proper simulation?”

“Common sense. This whole idea of throwing technology at problems created by technology… “

Maria felt her patience desert her. “What would you rather do? Be humble in the presence of nature, and hope you’ll be rewarded for it? You think Mother Gaia is going to forgive us, and put everything right—just as soon as we throw away our wicked computers and promise to stop trying to fix things ourselves?” Should have made that “Nanny Gaia.”

Aden scowled. “No—but the only way to “fix things” is to have less impact on the planet, not more. Instead of thinking up these grandiose schemes to bludgeon everything into shape, we have to back off, leave it alone, give it a chance to heal.”

Maria was bemused. “It’s too late for that. If that had started a hundred years ago… fine. Everything might have turned out differently. But it’s not enough any more; too much damage has already been done. Tip-toeing through the debris, hoping all the systems we’ve fucked up will magically restore themselves—and tip-toeing twice as carefully every time the population doubles—just won’t work. The whole planetary ecosystem is as much of an artifact, now, as… a city’s microclimate. Believe me, I wish that wasn’t the case, but it is—and now that we’ve created an artificial world, intentionally or not, we’d better learn to control it. Because if we stand back and leave it all to chance, it’s just going to collapse around us in some random fashion that isn’t likely to be any better than our worst well-intentioned mistakes.”

Aden was horrified. “An artificial world? You honestly believe that?”

“Yes.”

“Only because you spend so much time in Virtual Reality you don’t know the difference anymore.”

Maria was indignant. “I hardly ever—“ Then she stopped herself, realizing that he meant the Autoverse. She’d long ago given up trying to drum the distinction into his head.

Aden said, “I’m sorry. That was a cheap shot.” He made a gesture of retraction, a wave of the hand more impatient than apologetic. “Look, forget all this depressing ecoshit. I’ve got some good news, for a change. We’re going to Seoul.”

Maria laughed. “Are we? Why?”

“I’ve been offered a job. University Music Department.”

She looked at him sharply. “Thanks for telling me you’d applied.”

He shrugged it off lightly. “I didn’t want to get your hopes up. Or mine. I only heard this afternoon; I can still hardly believe it. Composer-in-residence, for a year; a couple of hours a week teaching, the rest of the time I can do what I like: writing, performing, producing, whatever. And they throw in free accommodation. For two.”

“Just… hold it. A few hours’ teaching? Then why do you have to go there in person?”

“They want me, physically. It’s a prestige thing. Every Mickey Mouse university can plug into the networks and bring in a dozen lecturers from around the world—”

“That’s not Mickey Mouse, it’s efficient.”

“Cheap and efficient. This place doesn’t want to be cheap. They want a piece of exotic cultural decoration. Stop laughing. Australia is flavor of the month in Seoul; it only happens once every twenty years, so we’d better take advantage of it. And they want a composer-in-residence. In residence.”

Maria sat back and digested it.

Aden said, “I don’t know about you, but I have a lot of trouble imagining us ever being able to afford to spend a year in Korea, any other way.”

“And you’ve said yes?”

“I said maybe. I said probably.”

“Accommodation for two. What am I supposed to do while you’re being exotic and decorative?”

“Whatever you like. Anything you do here, you could do just as easily there. You’re the one who keeps telling me how you’re plugged into the world, you’re a node in a logical data space, your physical location is entirely irrelevant… “

“Yes, and the whole point of that is not having to move. I like it where I am.”

“That shoebox.”

“A campus apartment in Seoul won’t be much bigger.”

“We’ll go out! It’s an exciting city—there’s a whole cultural renaissance going on there, it’s not just the music scene. And who knows? You might find some exciting project to work on. Not everything gets broadcast over the nets.”

That was true enough. Korea had full membership of ASEAN, as opposed to Australia’s probationary status; if she’d been living in Seoul at the right time, if she’d had the right contacts, she might have ended up part of Operation Butterfly. And even if that was wishful thinking—the right contacts probably took a decade to make—she could hardly do worse than she’d been doing in Sydney.

Maria fell silent. It was good news, a rare opportunity for both of them, but she still couldn’t understand why he was unloading it on her out of the blue. He should have told her everything when he’d applied, however poorly he’d rated his chances.

She glanced at the stage, at the twelve sweating musicians playing their hearts out, then looked away. There was something disconcertingly voyeuristic about watching them without tuning in: not just the sight of them emoting in silence, but also the realization that none of the bands could see each other, despite the fact that she could see them all.

Aden said, “There’s no rush to make up your mind. The academic year starts on January ninth. Two months away.”

“Won’t they need to know, long before then?”

“They’ll need to know by Monday if I’ve accepted the job—but I don’t think the accommodation will be a big deal. I mean, if I end up alone in an apartment for two, it’ll hardly be the end of the world.” He looked at her innocently, as if daring her to give the time and place he’d ever promised to turn down a chance like this, just because she didn’t want to come along for the ride.

Maria said, “No, of course not. How stupid of me.”

Home, Maria couldn’t resist logging on to the QIPS Exchange, just to find out what was going on. Operation Butterfly had vanished from the market. Omniaveritas, her knowledge miner, had picked up no news reports of a typhoon in the region; perhaps the predicted one had failed to eventuate—or perhaps it was yet to appear, but the simulations had already given their verdict. It was strange to think that it could all be over before the storm was a reality… but then, by the time anything newsworthy happened, the actual meteorological data would—hopefully—bear no relationship at all to what would have happened if the weather control rigs had been in use. The only real-world data needed for the simulations was the common starting point, a snapshot of the planet’s weather the moment before intervention would have begun.

The QIPS rate was still about fifty percent higher than normal, as ordinary users jostled to get their delayed work done. Maria hesitated; she felt like she needed cheering up, but running the Autoverse now would be stupid; it would make far more sense to wait until morning.

She logged on to the JSN, slipped on her gloves, activated the workspace. An icon of a man tripping on a banana skin, frozen in mid-fall, represented the snapshot of her interrupted task. She prodded it, and the Petri dishes reappeared in front of her instantly, the A. lamberti feeding, dividing and dying, as if the past fifteen hours had never happened.

She could have asked Aden to his face: Do you want to go to Seoul alone? Do you want a year away from me? If that’s it, why don’t you just say so? But he would have denied it, whether or not it was the truth. And she wouldn’t have believed him, whether or not he was lying. Why ask the question, if the answer told you nothing?

And it hardly seemed to matter, now: Seoul or Sydney, welcome or not. She could reach this place from anywhere—geographically or emotionally. She stared into the workspace, ran a gloved finger around the rim of one of the Petri dishes, and declaimed mockingly, “My name is Maria, and I am an Autoverse addict.”

As she watched, the culture in the dish she’d touched faded from muddy blue to pure brown, and then began to turn transparent, as the viewing software ceased classifying dead A. lamberti as anything more than chance arrangements of organic molecules.

As the brown mass dissolved, though, Maria noticed something she’d missed.

A tiny speck of electric blue.

She zoomed in on it, refusing to leap to conclusions. The speck was a small cluster of surviving bacteria, growing slowly—but that didn’t prove anything. Some strains always lasted longer than others; in the most pedantic sense, there was always a degree of “natural selection” taking place—but the honor of being the last of the dinosaurs wasn’t the kind of evolutionary triumph she was looking for.

She summoned up a histogram showing the prevalence of different forms of the epimerase enzymes, the tools she’d been pinning her hopes on to turn mutose back into nutrose… but there was nothing out of the ordinary, just the usual scatter of short-lived, unsuccessful mutations. No hint of how this strain was different from all of its extinct cousins.

So why was it doing so well?

Maria “tagged” a portion of the mutose molecules in the culture medium, assigning multiple clones of Maxwell’s Demon to track their movements and render them visible… the Autoverse equivalent of the real-world biochemist’s technique of radioactive labeling—along with something like nuclear magnetic resonance, since the demons would signal any chemical changes, as well as indicating position. She zoomed in on one surviving A. lamberti, rendered neutral gray now, and watched a swarm of phosphorescent green pin-pricks pass through the cell wall and jostle around the protoplasm in the sway of Brownian motion.

One by one, a fraction of the tags changed from green to red, marking passage through the first stage of the metabolic pathway: the attachment of an energy-rich cluster of atoms—more or less the Autoverse equivalent of a phosphate group. But there was nothing new in that; for the first three stages of the process, the enzymes which worked with nutrose would squander energy on the impostor as if it were the real thing.

Strictly speaking, these red specks weren’t mutose any more, but Maria had instructed the demons to turn an unmistakable violet, not only in the presence of nutrose itself, but also if the molecules under scrutiny were rehabilitated at a later stage—salvaged in mid-digestion. With the epimerase enzymes unchanged, she doubted that this was happening… but the bacteria were thriving, somehow.

The red-tagged molecules wandered the cell at random, part-digested mixed with raw indiscriminately. Neat process diagrams of metabolism—the real-world Embden-Meyerhof pathway, or the Autoverse’s Lambert pathway—always gave the impression of some orderly molecular conveyor belt, but the truth was, life in either system was powered by nothing at the deepest level but a sequence of chance collisions.

A few red tags turned orange. Stage two: an enzyme tightening the molecule’s hexagonal ring into a pentagon, transforming the spare vertex into a protruding cluster, more exposed and reactive than before.

Still nothing new. And still no hint of violet.

Nothing further seemed to happen for so long that Maria glanced at her watch and said “Globe,” to see if some major population center had just come on-line for the day—but the authentic Earth-from-space view showed dawn well into the Pacific. California would have been busy since before she’d arrived home.

A few orange tags turned yellow. Stage three of the Lambert pathway, like stage one, consisted of bonding an energy-rich group of atoms to the sugar. With nutrose, there was a payoff for this, eventually, with twice as many of the molecules which supplied the energy ending up “recharged” as had been “drained.” Stage four, though—the cleaving of the ring into two smaller fragments—was the point where mutose gummed up the works irretrievably…

Except that one yellow speck had just split into two, before her eyes… and both new tags were colored violet.

Maria, startled, lost track of the evidence. Then she caught sight of the same thing happening again. And then a third time.

It took her a minute to think it through, and understand what this meant. The bacterium wasn’t reversing the change she’d made to the sugar, converting mutose back into nutrose—or doing the same to some part-digested metabolite. Instead, it must have modified the enzyme which broke the ring, coming up with a version which worked directly on the metabolite of mutose.

Maria froze the action, zoomed in, and watched a molecular-scale replay. The enzyme in question was constructed of thousands of atoms; it was impossible to spot the difference at a glance—but there was no doubt about what it was doing. The two-atom blue-red spike she’d repositioned on the sugar was never shifted back into its “proper” place; instead, the enzyme now accommodated the altered geometry perfectly.

She summoned up old and new versions of the enzyme, highlighted the regions where the tertiary structure was different, and probed them with her fingertips—confirming, palpably, that the cavity in the giant molecule where the reaction took place had changed shape.

And once the ring was cleaved? The fragments were the same, whether the original sugar had been nutrose or mutose. The rest of the Lambert pathway went on as if nothing had changed.

Maria was elated, and a little dazed. People had been trying to achieve a spontaneous adaptation like this for sixteen years. She didn’t even know why she’d finally succeeded; for five years she’d been tinkering with the bacterium’s error correction mechanisms, trying to force A. lamberti to mutate, not more rapidly, but more randomly. Every time, she’d ended up with a strain which—like Lambert’s original, like those of other workers—suffered the same handful of predictable, useless mutations again and again… almost as if something deep in the clockwork of the Autoverse itself ruled out the exuberant diversity which came so effortlessly to real-world biology. Calvin and others had suggested that, because Autoverse physics omitted the deep indeterminacy of real-world quantum mechanics—because it lacked this vital inflow of “true unpredictability”—the same richness of phenomena could never be expected, at any level.

But that had always been absurd—and now she’d proved it was absurd.

For a moment she thought of phoning Aden, or Francesca—but Aden wouldn’t understand enough to do more than nod politely, and her mother didn’t deserve to be woken at this hour.

She got up and paced the tiny bedroom for a while, too excited to remain still. She’d upload a letter to Autoverse Review (total subscription, seventy-three), with the genome of the strain she’d started out with appended as a footnote, so everyone else could try the experiment…

She sat down and began composing the letter—popping up a word processor in the foreground of the workspace—then decided that was premature; there was still a lot more to be done to form the basis of even a brief report.

She cloned a small colony of the mutose-eating strain, and watched it grow steadily in a culture of pure mutose. No surprise, but it was still worth doing.

Then she did the same, with pure nutrose, and the colony, of course, died out at once. The original ring-cleaving enzyme had been lost; the original roles of nutrose and mutose as food and poison had been swapped.

Maria pondered this. A. lamberti had adapted—but not in the way she’d expected. Why hadn’t it found a means of consuming both sugars, instead of exchanging one kind of exclusive reliance for another? It would have been a far better strategy. It was what a real-world bacterium would have done.

She brooded over the question for a while—then started laughing. Sixteen years, people had been hunting for a single, convincing example of natural selection in the Autoverse—and here she was worrying that it wasn’t the best of all possible adaptations. Evolution was a random walk across a minefield, not a preordained trajectory, onward and upward toward “perfection.” A. lamberti had stumbled on a successful way to turn poison into food. It was tough luck if the corollary was: vice versa.

Maria ran a dozen more experiments. She lost all track of time; when dawn came, the software brightened the images in front of her, keeping the daylight from washing them out. It was only when her concentration faltered, and she looked around the room, that she realized how late it was.

She started again on the letter. After three drafts of the first paragraph—all eliciting the same response from Camel’s Eye: You’ll hate this when you reread it later. Trust me.—she finally admitted to herself that she was wasted. She shut down everything and crawled into bed.

She lay there awhile in a stupor, burying her face in the pillow, waiting for the ghost images of Petri dishes and enzymes to fade. Five years ago, she could have worked all night, and suffered nothing worse than a fit of yawning in the middle of the afternoon. Now, she felt like she’d been hit by a train—and she knew she’d be a wreck for days. Thirty-one is old, old, old.

Her head throbbed, her whole body ached. She didn’t care. All the time and money she’d squandered on the Autoverse was worth it, now. Every moment she’d spent there had been vindicated.

Yeah? She rolled onto her back and opened her eyes. What, exactly, had changed? It was still nothing but a self-indulgent hobby, an elaborate computer game. She’d be famous with seventy-two other anal-retentive Autoverse freaks. How many bills would that pay? How many typhoons would it neutralize?

She wrapped her head in the pillow, feeling crippled, stupid, hopeless—and defiantly happy—until her limbs went numb, her mouth went dry, and the room seemed to rock her to sleep.

5 (Remit not paucity) NOVEMBER 2050

Peer anchored the soles of both feet and the palm of one hand firmly against the glass, and rested for a while. He tipped his head back to take in, one more time, the silver wall of the skyscraper stretching to infinity above him. Cotton-wool clouds drifted by, higher than any part of the building—even though the building went on forever.

He freed his right foot, reanchored it higher up the wall, then turned and looked down at the neat grid of the city below, surrounded by suburbs as orderly as ploughed fields. The foreshortened countryside beyond formed a green-brown rim to the hemispherical bowl of the Earth; a blue-hazed horizon bisected the view precisely. The features of the landscape, like the clouds, were “infinitely large,” and “infinitely distant”; a finite city, however grand, would have shrunk to invisibility, like the base of the skyscraper. The distance was more than a trick of perspective, though; Peer knew he could keep on approaching the ground for as long as he liked, without ever reaching it. Hours, days, centuries.

He couldn’t remember beginning the descent, although he understood clearly—cloud-knowledge, cloud-memories—the sense in which there was a beginning, and the sense in which there was none. His memories of the skyscraper, like his view of it, seemed to converge toward a vanishing point; looking back from the present moment, all he could recall was the act of descending, punctuated by rest. And although his mind had wandered, he’d never lost consciousness; his past seemed to stretch back seamlessly, forever—yet he could hold it all in his finite gaze, thanks to some law of mental perspective, some calculus of memory limiting the sum of ever diminishing contributions to his state of mind from ever more distant moments in the past. But he had his cloud-memories, too; memories from before the descent. He couldn’t join them to the present, but they existed nonetheless, a backdrop informing everything else. He knew exactly who he’d been, and what he’d done, in that time before the time he now inhabited.

Peer had been exhausted when he’d stopped, but after a minute’s rest he felt, literally, as energetic and enthusiastic as ever. Back in cloud-time, preparing himself, he’d edited out any need or desire for food, drink, sleep, sex, companionship, or even a change of scenery, and he’d preprogrammed his exoself—the sophisticated, but nonconscious, supervisory software which could reach into the model of his brain and body and fine-tune any part of it as required—to ensure that these conditions remained true. He resumed the descent gladly, a happy Sisyphus. Making his way down the smooth mirrored face of the skyscraper was, still, the purest joy he could imagine: the warmth of the sun reflecting back on him, the sharp cool gusts of wind, the faint creak of steel and concrete. Adrenaline and tranquility. The cycle of exertion and perfect recovery. Perpetual motion. Touching infinity.

The building, the Earth, the sky, and his body vanished. Stripped down to vision and hearing, Peer found himself observing his Bunker: a cluster of display screens floating in a black void. Kate was on one screen; two-dimensional, black-and-white, nothing but her lips moving.

She said, “You set your threshold pretty damn high. You’d be hearing about this a decade later if I hadn’t called you in.”

Peer grunted—disconcerted for a moment by the lack of tactile feedback from the conventional organs of speech—and glanced, by way of eye-movement-intention, at the screen beside her, a graph of the recent history of Bunker time versus realtime.

Observing the Bunker—“being in it” would have been an overstatement—was the most computing-efficient state a Copy could adopt, short of losing consciousness. Peer’s body was no longer being simulated at all; the essential parts of his model-of-a-brain had been mapped into an abstract neural network, a collection of idealized digital gates with no pretensions to physiological verisimilitude. He didn’t enter this state very often, but Bunker time was still a useful standard as a basis for comparisons. At best—on the rare occasions when demand slackened, and he shared a processor cluster with only two or three other users—his Bunker-time slowdown factor dropped to about thirty. At worst? Up until a few minutes ago, the worst had been happening: a section of the graph was perfectly flat. For more than ten hours of real time, he hadn’t been computed at all.

Kate said, “Operation Butterfly. Weather control simulations. The fuckers bought up everything.”

She sounded shaken and angry. Peer said calmly, “No great loss. Solipsist Nation means making your own world, on your own terms. Whatever the risks. Real time doesn’t matter. Let them give us one computation per year. What would it change? Nothing.” He glanced at another display, and realized that he’d only been in the skyscraper model for seven subjective minutes. The false memories had meshed perfectly; he would never have believed it had been so short a time. Pre-computing the memories had taken time, of course—but far less than it would have taken to accumulate the same effect by conventional experience.

Kate said, “You’re wrong. You don’t—”

“Let them run one moment of model time for one Copy on every processor cluster, the day it’s commissioned—and then dedicate it entirely to other users. Each Copy would thread its way from machine to machine, with a slowdown of a few billion… and it wouldn’t matter. The manufacturers could run us all for free—turn it into a kind of ritual, a blessing of the hardware by the spirits of the dead. Then we could abolish all the trust funds, and stop worrying about money altogether. The cheaper we are, the less vulnerable we are.”

“That’s only half the truth. The more we’re marginalized, the more we’re at risk.”

Peer tried to sigh; the sound that emerged was plausible enough, but the lack of sensation was annoying.

“Is there any reason to stay in emergency mode? Is there some snap decision I’m going to have to make? Are there missiles heading for—“ He checked a display. “-- Dallas?” Dallas? The US dollar must have fallen sharply against the yen.

Kate said nothing, so Peer glanced at icons for a body and a room, and willed them to be active. His disembodied consciousness, and the floating screens of the Bunker, fleshed out into a young man, barefoot in blue jeans and a T-shirt, sitting in a windowless control room—what might have been the operations center for a medium-sized office building.

The body’s physiological state continued directly from its last moments on the wall of the skyscraper—and it felt good: loose-limbed, invigorated. Peer recorded a snapshot, so he could get the feeling back again at will. He looked at Kate imploringly; she relented and joined him, vanishing from the screen and appearing on a chair beside him.

She said, “I am Solipsist Nation. What happens outside doesn’t matter to me… but we still need certain guarantees, certain minimum standards.”

Peer laughed. “So what are you going to do? Become a lobbyist now? Spend all your time petitioning Brussels and Geneva? “Human rights” are for people who want to play at being human. I know who I am. I am not human.” He plunged his fist into his chest, effortlessly penetrating shirt, skin and ribs, and tore his heart out. He felt the parting of his flesh, and the aftermath—but although aspects of the pain were “realistic,” preprogrammed barriers kept it isolated within his brain, a perception without any emotional, or even metabolic, consequences. And his heart kept beating in his hand as if nothing had happened; the blood passed straight between the ragged ends of each broken artery, ignoring the “intervening distance.”

Kate said, “Blink and ten hours are gone. That’s no disaster—but where is it heading? State-of-emergency decrees, nationalizing all the computing power in Tokyo for weather control?”

“Tokyo?”

“Some models show Greenhouse Typhoons reaching the Japanese islands in the next thirty years.”

“Fuck Tokyo. We’re in Dallas.”

“Not any more.” She pointed to the status display; exchange-rate fluctuations, and the hunt for the cheapest QIPS, had flung them back across the Pacific. “Not that it matters. There are plans for the Gulf of Mexico, too.”

Peer put his heart on the floor and shrugged, then groped around in his chest cavity in search of other organs. He finally settled on a handful of lung. Torn free, the pink tissue continued to expand and contract in time with his breathing; functionally, it was still inside his rib cage. “Start looking for security, and you end up controlled by the demands of the old world. Are you Solipsist Nation, or not?”

Kate eyed his bloodless wound, and said quietly, “Solipsist Nation doesn’t mean dying of stupidity. You take your body apart, and you think it proves you’re invulnerable? You plant a few forced-perspective memories, and you think you’ve already lived forever? I don’t want some cheap illusion of immortality. I want the real thing.”

Peer frowned, and started paying attention to her latest choice of body. It was still recognizably “Kate”—albeit the most severe variation on the theme he’d seen. Short-haired, sharp-boned, with piercing gray eyes; leaner than ever, plainly dressed in loose-fitting white. She looked ascetic, functional, determined.

She said—mock-casually, as if changing the subject—“Interesting news: there’s a man—a visitor—approaching the richest Copies, selling prime real estate for second versions at a ludicrous rate.”

“How much?”

“Two million ecus.”

“What—per month?”

“No. Forever.”

Peer snorted. “It’s a con.”

“And outside, he’s been contracting programmers, designers, architects. Commissioning—and paying for—work that will need at least a few dozen processor clusters to run on.”

“Good move. That might actually persuade a few of the doddering old farts that he can deliver what he’s promising. Not many, though. Who’s going to pay without getting the hardware on-line and running performance tests? How’s he going to fake that? He can show them simulations of glossy machines, but if the things aren’t real, they won’t crunch. End of scam.”

“Sanderson has paid. Repetto has paid. The last word I had was he’d talked to Riemann.”

“I don’t believe any of this. They all have their own hardware—why would they bother?”

“They all have a high profile. People know that they have their own hardware. If things get ugly, it can be confiscated. Whereas this man, Paul Durham, is nobody. He’s a broker for someone else, obviously—but whoever it is, they’re acting like they have access to more computing power than Fujitsu, at about a thousandth of the cost. And none of it is on the open market. Nobody officially knows it exists.”

“Or unofficially. Because it doesn’t. Two million ecus!”

“Sanderson has paid. Repetto has paid.”

“According to your sources.”

“Durham’s getting money from somewhere. I spoke to Malcolm Carter myself. Durham’s commissioned a city from him, thousands of square kilometers—and none of it passive. Architectural detail everywhere down to visual acuity, or better. Pseudo-autonomous crowds—hundreds of thousands of people. Zoos and wildlife parks with the latest behavioral algorithms. A waterfall the size of nothing on Earth.”

Peer pulled out a coil of intestine and playfully wrapped it around his neck. “You could have a city like that, all to yourself, if you really wanted it—if you were willing to live with the slowdown. Why are you so interested in this con man Durham? Even if he’s genuine, you can’t afford his price. Face it: you’re stuck here in the slums with me—and it doesn’t matter.” Peer indulged in a brief flashback to the last time they’d made love. He merged it with the current scene, so he saw both Kates, and the new lean gray-eyed one seemed to look on as he lay on the floor gasping beneath his tangible memory of her earlier body—although in truth she saw him still sitting in the chair, smiling faintly.

All memory is theft, Daniel Lebesgue had written. Peer felt a sudden pang of post-coital guilt. But what was he guilty of? Perfect recollection, nothing more.

Kate said, “I can’t afford Durham’s price—but I can afford Carter’s.”

Peer was caught off guard for a second, but then he grinned at her admiringly. “You’re serious, aren’t you?”

She nodded soberly. “Yes. I’ve been thinking about it for some time, but after being flatlined for ten hours—”

“Are you sure Carter is serious? How do you know he really has something to sell?”

She hesitated. “I hired him myself, when I was outside. I used to spend a lot of time in VR, as a visitor, and he made some of my favorite places: the winter beach; that cottage I took you to. And others. He was one of the people I talked it over with, before I made up my mind to come in for good.” Peer regarded her uneasily—she rarely talked about the past, which suited him fine—and mercifully, she returned to the point. “With slowdown, filters, masks, it’s hard to judge anyone… but I don’t think he’s changed that much. I still trust him.”

Peer nodded slowly, absentmindedly sliding his intestine back and forth across his shoulders. “But how much does Durham trust him? How thoroughly will he check the city for stowaways?”

“Carter’s sure he can hide me. He has software that can break up my model and bury it deep in the city’s algorithms—as a few billion trivial redundancies and inefficiencies.”

Inefficiencies can get optimized out. If Durham—”

Kate cut him off impatiently. “Carter’s not stupid. He knows how optimizers work—and he knows how to keep them from touching his stuff.”

“Okay. But… once you’re in there, what sort of communications will you have?”

“Not much. Only limited powers to eavesdrop on what the legitimate inhabitants choose to access—and if the whole point of this place is secrecy, that may not be much. I get the impression from Carter that they’re planning to drag in everything they need, then pull up the drawbridge.”

Peer let that sink in, but chose not to ask the obvious question, or to show that he’d even thought of it. “So what do you get to take with you?”

“All the software and all the environments I’ve been using here—which doesn’t amount to all that much data, compared to me. And once I’m in, I’ll have read-only access to all of the city’s public facilities: all the information, all the entertainment, all the shared environments. I’ll be able to walk down the main street—invisible and intangible—staring at the trillionaires. But my presence won’t affect anything—except to slow it all down by a negligible amount—so even the most rigorous verification should pass the total package as contamination-free.”

“What rate will you run at?”

Kate snorted. “I should refuse to answer that. You’re the champion of one computation per year.”

“I’m just curious.”

“It depends how many QIPS are allocated to the city.” She hesitated. “Carter has no real evidence for this—but he thinks there’s a good chance that Durham’s employers have got their hands on some kind of new high-powered hardware—”

Peer groaned. “Please, this whole deal is already suspect enough—don’t start invoking the mythical breakthrough. What makes people think that anyone could keep that a secret? Or that anyone would even want to?”

“They might not want to, in the long run. But the best way to exploit the technology might be to sell the first of the new generation of processors to the richest Copies—before they hit the open market and the QIPS rate crashes.”

Peer laughed. “Then why stow away at all? If that happens, there’ll be nothing to fear from weather control.”

“Because there might not have been any breakthrough. The only thing that’s certain is that some of the wealthiest—and best-informed—Copies have decided that it’s worth going into this… sanctuary. And I’ve got the chance to go with them.”

Peer was silent for a while. Finally, he asked, “So are you moving—or cloning yourself?”

“Cloning.”

He could have concealed his relief, easily—but he didn’t. He said, “I’m glad. I would have missed you.”

“And I’d have missed you. I want you to come with me.”

“You want—?”

Kate leaned toward him. “Carter has said he’ll include you—and your baggage—for another fifty percent. Clone yourself and come with me. I don’t want to lose you—either of me.”

Peer felt a rush of excitement—and fear. He took a snap-shot of the emotion, then said, “I don’t know. I’ve never—”

“A second version, running on the most secure hardware on the planet. That’s not surrendering to outside—it’s just finally gaining some true independence.”

“Independence? What if these Copies get bored with Carter’s city and decide to trash it—trade it in for something new?”

Kate was unfazed. “That’s not impossible. But there are no guarantees on the public networks, either. This way, at least you have a greater chance that one version will survive.”

Peer tried to imagine it. “Stowaways. No communications. Just us, and whatever software we bring.”

“You’re Solipsist Nation, aren’t you?”

“You know I am. But… I’ve never run a second version before. I don’t know how I’ll feel about that, after the split.”

How who will feel about it?

Kate bent over and picked up his heart. “Having a second version won’t bother you.” She fixed her new gray eyes on him. “We’re running at a slowdown of sixty-seven. Carter will be delivering his city to Durham, six real-time months from now. But who knows when Operation Butterfly will flat-line us again? So you don’t have long to decide.”

Peer continued to show Kate his body sitting in the chair, thinking it over, while in truth he rose to his feet and walked across the room, escaping her formidable gaze.

Who am I? Is this what I want?

He couldn’t concentrate. He manually invoked a menu on one of the control screens, an array of a dozen identical images: a nineteenth-century anatomical drawing of the brain, with the surface divided into regions labeled with various emotions and skills. Each icon represented a package of mental parameters: snapshots of previous states of mind, or purely synthetic combinations.

Peer hit the icon named clarity.

In twelve short real-time years as a Copy, he’d tried to explore every possibility, map out every consequence of what he’d become. He’d transformed his surroundings, his body, his personality, his perceptionsbut he’d always owned the experience himself. The tricks he’d played on his memory had added, never erased—and whatever changes he’d been through, there was always only one person, in the end, taking responsibility, picking up the pieces. One witness, unifying it all.

The truth was, the thought of finally surrendering that unity made him dizzy with fear. It was the last vestige of his delusion of humanity. The last big lie.

And as Daniel Lebesgue, founder of Solipsist Nation, had written: “My goal is to take everything which might be revered as quintessentially human… and grind it into dust.”

He returned to his seated body, and said, “I’ll do it.”

Kate smiled, raised his beating heart to her lips, and gave it a long, lingering kiss.

6 (Rip, tie, cut toy man) JUNE 2045

Paul woke without any confusion. He dressed and ate, trying to feel optimistic. He’d demonstrated his willingness to cooperate; now it was time to ask for something in return. He walked into the study, switched on the terminal, and called his own number. The djinn answered at once.

Paul said, “I’d like to talk to Elizabeth.”

Squeak. “That’s not possible.”

“Not possible? Why don’t you just ask her?”

Squeak. “I can’t do that. She doesn’t even know you exist.”

Paul stared at him coldly. “Don’t lie to me, it’s a waste of time. As soon as I had a Copy who survived, I was going to explain everything—”

Squeak. The djinn said drily, “Or so we thought.”

Paul’s certainty wavered. “You’re telling me that your great ambition is finally being fulfilled—and you haven’t even mentioned it to the one woman… ?”

Squeak. Durham’s face turned to stone. “I really don’t wish to discuss it. Can we get on with the experiment, please?”

Paul opened his mouth to protest—and then found he had nothing to say. All his anger and jealousy suddenly dissipated into… embarrassment. It was as if he’d just come to his senses from a daydream, an elaborate fantasy of a relationship with someone else’s lover. Paul and Elizabeth. Elizabeth and Paul. What happened between them was none of his business. Whatever his memories suggested, that life wasn’t his to live anymore.

He said, “Sure, let’s get on with the experiment. Time is just rushing by. You must have turned forty-five… what, a day ago? Many happy returns.”

Squeak. “Thanks—but you’re wrong. I took some shortcuts while you were asleep: I shut down part of the model—and cheated on most of the rest. It’s only the fourth of June; you got six hours’ sleep in ten hours’ real time. Not a bad job, I thought.”

Paul was outraged. “You had no right to do that!”

Squeak. Durham sighed. “Be practical. Ask yourself what you’d have done in my place.”

“It’s not a joke!

Squeak. “So you slept without a whole body. I cleaned a few toxins out of your blood at a non-physiological rate.” The djinn seemed genuinely puzzled. “Compared to the experiments, that’s nothing. Why should it bother you? You’ve woken up in exactly the same condition as you’d be in if you’d slept in the normal way.”

Paul caught himself. He didn’t want to explain how vulnerable it made him feel to have someone reach through the cracks in the universe and relieve him of unnecessary organs while he slept. And the less the bastard knew about his Copy’s insecurities, the better—he’d only exploit them.

He said, “It bothers me because the experiments are worthless if you’re going to intervene at random. Precise, controlled changes—that’s the whole point. You have to promise me you won’t do it again.”

Squeak. “You’re the one who was complaining about waste. Someone has to think about conserving our dwindling resources.”

“Do you want me to keep on cooperating? Or do you want to start everything again from scratch?”

Squeak. The djinn said mildly, “All right, you don’t have to threaten me. You have my word: no more ad hoc intervention.”

“Thank you.”

Conserving our dwindling resources? Paul had been trying hard not to think about money. What would the djinn do when he could no longer afford to keep him running—if Paul chose not to bale out once the experiments were over? Store a snapshot of the model, of course, until he could raise the cash flow to start it up again. In the long term, set up a trust fund; it would only have to earn enough to run him part-time, at first: keep him in touch with the world, stave off excessive culture shock… until the technology became cheap enough to let him live continuously.

Of course, all these reassuring plans had been made by a man with two futures. Would he really want to keep an old Copy running, when he could save his money for a deathbed scan, and “his own” immortality?

Squeak. “Can we get to work, now?”

“That’s what I’m here for.”

This time, the model would be described at the standard time resolution of one millisecond, throughout—but the order in which the states were computed would be varied.

Squeak. “Experiment two, trial number one. Reverse order.”

Paul counted. “One. Two. Three.” Reverse order. After an initial leap into the future, he was now traveling backward through real time. It would have been a nice touch if he’d been able to view an external event on the terminal—some entropic cliche like a vase being smashed—knowing that it was himself, and not the scene, that was being “rewound”… but he knew that it couldn’t be done (quite apart from the fact that it would have ruined the experiment, betraying the difference between subject and control). In real time, the first thing to be computed would be his model-time-final brain state, complete with memories of everything that “had happened” in the “preceding” ten seconds. Those memories couldn’t include having seen a real broken vase assemble itself from fragments, if the vase hadn’t even been smashed yet. The trick could have been done with a simulation, or a video recording of the real thing—but that wouldn’t have been the same.

“Eight. Nine. Ten.” Another imperceptible leap into the future, and the djinn reappeared.

Squeak. “Trial number two. Odd numbered states, then even.”

In external terms: he would count to ten, skipping every second model-time moment… then forget having done so, and count again, going back and filling in the gaps.

And from his own point of view? As he counted, once only, the external world—even if he couldn’t see it—was flickering back and forth between two separate regions of time, which had been chopped up into seventeen-millisecond portions, and interleaved.

So… who was right? Paul thought it over, half seriously. Maybe both, descriptions were equally valid; after all, relativity had abolished absolute time. Everybody was entitled to their own frame of reference; crossing deep space at close to lightspeed, or skimming the event horizon of a black hole. Why shouldn’t a Copy’s experience of time be as sacrosanct as that of any astronaut?

The analogy was flawed, though. Relativistic transformations were smooth—possibly extreme, but always continuous. One observer’s space-time could be stretched and deformed in the eyes of another—but it couldn’t be sliced like a loaf of bread and then shuffled like a deck of cards.

“Every tenth state, in ten sets.”

Paul counted—and for argument’s sake, tried to defend his own perspective, tried to imagine the outside world actually cycling through fragments of time drawn from ten distinct periods. The trouble was… this allegedly shuddering universe contained the computer which ran the whole model, the infrastructure upon which everything else depended. If its orderly chronology had been torn to shreds, what was keeping him together, enabling him to ponder the question?

“Every twentieth state, in twenty sets.”

Nineteen episodes of amnesia, nineteen new beginnings.

(Unless, of course, he was the control.)

“Every hundredth state, in one hundred sets.”

He’d lost any real feeling for what was happening. He just counted.

“Pseudo-random ordering of states.”

“One. Two. Three.”

Now he was… dust. To an outside observer, these ten seconds had been ground up into ten thousand uncorrelated moments and scattered throughout real time—and in model time, the outside world had suffered an equivalent fate. Yet the pattern of his awareness remained perfectly intact: somehow he found himself, “assembled himself from these scrambled fragments. He’d been taken apart like a jigsaw puzzle—but his dissection and shuffling were transparent to him. Somehow—on their own terms—the pieces remained connected.

“Eight. Nine. Ten.”

Squeak. “You’re sweating.”

“Both of me?”

Squeak. The djinn laughed. “What do you think?”

Paul said, “Do me one small favor. The experiment is over. Shut down one of me—control or subject, I don’t care.”

Squeak. “Done.”

“Now there’s no need to conceal anything, is there? So run the pseudo-random effect on me again—and stay on-line. This time, you count to ten.”

Squeak. Durham shook his head. “Can’t do it, Paul. Think about it: you can’t be computed non-sequentially when past perceptions aren’t known.”

Of course; the broken vase problem all over again.

Paul said, “Record yourself, then, and use that.”

The djinn seemed to find the request amusing, but he agreed; he even slowed down the recording so it lasted ten model-time seconds. Paul watched the blurred lips and jaws intently, listened carefully to the drone of white noise.

Squeak. “Happy now?”

“You did scramble me, and not the recording?”

Squeak. “Of course. Your wish is my command.”

“Yeah? Then do it again.”

Durham grimaced, but obliged.

Paul said, “Now, scramble the recording.”

It looked just the same. Of course.

“Again.”

Squeak. “What’s the point of all this?”

“Just do it.”

Paul watched, the hairs on the back of his neck rising, convinced that he was on the verge of… what? Finally confronting the “obvious” fact that the wildest permutations in the relationship between model time and real time would be undetectable to an isolated Copy? He’d accepted the near certainty of that, tacitly, for almost twenty years… but the firsthand experience of having his mind literally scrambled—to absolutely no effect—was still provocative in a way that the abstract understanding had never been.

He said, “When do we move on to the next stage?”

Squeak. “Why so keen all of a sudden?”

“Nothing’s changed. I just want to get it over and done with.”

Squeak. “Lining up all the other machines is taking some delicate negotiations. The network allocation software isn’t designed to accommodate whims about geography. It’s a bit like going to a bank and asking to deposit some money… at a certain location in a particular computer’s memory. Basically, people think I’m crazy.”

Paul felt a momentary pang of empathy, recalling his own anticipation of these difficulties. Empathy verging on identification. He smothered it. The two of them were irreversibly different people now, with different problems and different goals—and the stupidest thing he could do would be to forget that.

Squeak. “I could suspend you while I finalize the arrangements, save you the boredom—if that’s what you want.”

“You’re too kind. But I’d rather stay conscious. I’ve got a lot to think about.”

7 (Remit not paucity) NOVEMBER 2050

“Twelve to eighteen months? Are they sure?”

Francesca Deluca said drily, “What can I say? They modeled it.”

Maria did her best to sound calm. “That’s plenty of time. We’ll get you scanned. We’ll get the money together. I can sell the house, and borrow some from Aden—”

Francesca smiled but shook her head. “No, darling.” Her hair had grayed a little since Maria had last really looked at her, last consciously gauged her appearance, but she showed no obvious signs of ill health. “What’s the point? Even if I wanted that—and I don’t—what’s the use of a scan that will never be run?”

“It will be run. Computing power will get cheaper. Everybody’s counting on that. Thousands of people have scan files waiting—”

“How many frozen corpses have ever been revived?”

“That’s not the same thing at all.”

“How many?”

“Physically, none. But some have been scanned—”

“And proved non-viable. All the interesting ones—the celebrities, the dictators—are brain-damaged, and nobody cares about the rest.”

“A scan file is nothing like a frozen corpse. You’d never become non-viable.”

“No, but I’d never become worth bringing back to life, either.”

Maria stared at her angrily. “I’ll bring you back to life. Or don’t you believe I’ll ever have the money?”

Francesca said, “Maybe you will. But I’m not going to be scanned, so forget about it.”

Maria hunched forward on the couch, not knowing how to sit, not knowing where to put her hands. Sunlight streamed into the room, obscenely bright, revealing every speck of lint on the carpet; she had to make an effort not to get up and close the blinds. Why hadn’t Francesca told her on the phone? All of this would have been a thousand times easier by phone.

She said, “All right, you’re not going to be scanned. Someone in the world must be making nanomachines for liver cancer. Even just experimental ones.”

“Not for this cell type. It’s not one of the common onco-genes, and nobody’s sure of the cell surface markers.”

“So? They can find them, can’t they? They can look at the cells, identify the markers, and modify an existing nanomachine. All the information they need is there in your body.” Maria pictured the mutant proteins which enabled metastasis poking through the cell walls, highlighted in ominous yellow.

Francesca said, “With enough time and money and expertise, I’m sure that would be possible… but as it happens, nobody plans to do it in the next eighteen months.”

Maria started shuddering. It came in waves. She didn’t make a sound; she just sat and waited for it to pass.

Finally, she said, “There must be drugs.”

Francesca nodded. “I’m on medication to slow the growth of the primary tumor, and limit further metastasis. There’s no point in a transplant; I already have too many secondary tumors—actual liver failure is the least of my worries. There are general cytotoxic drugs I could take, and there’s always radiation therapy—but I don’t think the benefits are worth the side effects.”

“Would you like me to stay with you?”

“No.”

“It’d be no trouble. You know I can work from anywhere.”

“There’s no need for it. I’m not going to be an invalid.”

Maria closed her eyes. She couldn’t imagine feeling this way for another hour, let alone another year. When her father had died of a heart attack, three years before, she’d promised herself that she’d raise the money to have Francesca scanned by her sixtieth birthday. She was nowhere near on target. I screwed up. I wasted time. And now it’s almost too late.

Thinking aloud, she said, “Maybe I’ll get some work in Seoul.”

“I thought you’d decided not to go.”

Maria looked up at her, uncomprehending. “Why don’t you want to be scanned? What are you afraid of? I’d protect you, I’d do whatever you asked. If you didn’t want to be run until slowdown is abolished, I’d wait. If you wanted to wake up in a physical body—an organic body—I’d wait.”

Francesca smiled. “I know you would, darling. That’s not the point.”

“Then what is the point?”

“I don’t want to argue about it.”

Maria was desperate. “I won’t argue. But can’t you tell me? Please?”

Francesca relented. “Listen, I was thirty-three when the first Copy was made. You were five years old, you grew up with the idea—but to me, it’s still… too strange. It’s something rich eccentrics do—the way they used to freeze their corpses. To me, spending hundreds of thousands of dollars for the chance to be imitated by a computer after my death is just… farcical. I’m not an eccentric millionaire, I don’t want to spend my money—or yours—building some kind of… talking monument to my ego. I still have a sense of proportion.” She looked at Maria imploringly. “Doesn’t that count for anything any more?”

“You wouldn’t be imitated. You’d be you.”

“Yes and no.”

“What’s that supposed to mean? You always told me you believed—”

“I do believe that Copies are intelligent. I just wouldn’t say that they are—or they aren’t—“the same person as” the person they were based on. There’s no right or wrong answer to that; it’s a question of semantics, not a question of truth.

“The thing is, I have my own sense—right now—of who I am… what my boundaries are… and it doesn’t include a Copy of me, run at some time in the indefinite future. Can you understand that? Being scanned wouldn’t make me feel any better about dying. Whatever a Copy of me might think, if one was ever run.”

Maria said, angrily, “That’s just being perverse. That’s as stupid as… saying when you’re twenty years old, “I can’t picture myself at fifty, a woman that old wouldn’t really be me.” And then killing yourself because there’s nothing to lose but that older woman, and she’s not inside your ‘boundaries.’”

“I thought you said you weren’t going to argue.”

Maria looked away. “You never used to talk like this. You’re the one who always told me that Copies had to be treated exactly like human beings. If you hadn’t been brain-washed by that ‘religion’—”

“The Church of the God Who Makes No Difference has no position on Copies, one way or the other.”

“It has no position on anything.”

“That’s right. So it can hardly be their fault that I don’t want to be scanned, can it?”

Maria felt physically sick. She’d held off saying anything on the subject for almost a year; she’d been astonished and appalled, but she’d struggled to respect her mother’s choice—and now she could see that that had been insane, irresponsible beyond belief. You don’t stand by and let someone you lovesomeone who gave you your own understanding of the worldhave their brain turned to pulp.

She said, “It’s their fault, because they’ve undermined your judgment. They’ve fed you so much bullshit that you can’t think straight about anything, anymore.”

Francesca just looked at her reprovingly. Maria felt a pang of guilt—How can you make things harder for her, now? How can you start attacking her, when she’s just told you that she’s dying?—but she wasn’t going to fold now, take the easy way out, be “supportive.”

She said, “‘God makes no difference… because God is the reason why everything is exactly what it is?’ That’s supposed to make us all feel at peace with the cosmos, is it?”

Francesca shook her head. “At peace? No. It’s just a matter of clearing away, once and for all, old ideas like divine intervention—and the need for some kind of proof, or even faith, in order to believe.”

Maria said, “What do you need, then? I don’t believe, so what am I missing?”

“Belief?”

“And a love of tautology.”

“Don’t knock tautology. Better to base a religion on tautology than fantasy.”

“But it’s worse than tautology. It’s… redefining words arbitrarily, it’s like something out of Lewis Carroll. Or George Orwell. “God is the reason for everything… whatever that reason is.” So what any sane person would simply call the laws of physics, you’ve decided to rename G-O-D… solely because the word carries all kinds of historical resonances—all kinds of misleading connotations. You claim to have nothing to do with the old religions—so why keep using their terminology?”

Francesca said, “We don’t deny the history of the word. We make a break from the past in a lot of ways—but we also acknowledge our origins. God is a concept people have been using for millennia. The fact that we’ve refined the idea beyond primitive superstitions and wish-fulfilment doesn’t mean we’re not part of the same tradition.”

“But you haven’t refined the idea, you’ve made it meaningless! And rightly so—but you don’t seem to realize it. You’ve stripped away all the obvious stupidities—all the anthropomorphism, the miracles, the answered prayers—but you don’t seem to have noticed that once you’ve done that, there’s absolutely nothing left that needs to be called religion. Physics is not theology. Ethics is not theology. Why pretend that they are?”

Francesca said, “But don’t you see? We talk about God for the simple reason that we still want to. There’s a deeply ingrained human compulsion to keep using that word, that concept—to keep honing it, rather than discarding it—despite the fact that it no longer means what it did five thousand years ago.”

“And you know perfectly well where that compulsion comes from! It has nothing to do with any real divine being; it’s just a product of culture and neurobiology—a few accidents of evolution and history.”

“Of course it is. What human trait isn’t?”

“So why give in to it?”

Francesca laughed. “Why give in to anything? The religious impulse isn’t some kind of… alien mind virus. It’s not—in its purest form, stripped of all content—the product of brain-washing. It’s a part of who I am.”

Maria put her face in her hands. “Is it? When you talk like this, it doesn’t sound like you.”

Francesca said, “Don’t you ever want to give thanks to God when things are going well for you? Don’t you ever want to ask God for strength when you need it?”

“No.”

“Well, I do. Even though I know God makes no difference. And if God is the reason for everything, then God includes the urge to use the word God. So whenever I gain some strength, or comfort, or meaning, from that urge, then God is the source of that strength, that comfort, that meaning.

“And if God—while making no difference—helps me to accept what’s going to happen to me, why should that make you sad?”

+ + +

On the train home, Maria sat next to a boy of about seven, who twitched all the way to the silent rhythms of a nerve-induced PMV—participatory music video. Nerve induction had been developed to treat epilepsy, but now its most common use seemed to bring about the symptoms it was meant to alleviate. Glancing at him sideways, she could see his eyeballs fluttering behind his mirror shades.

As the shock of the news diminished, slightly, Maria began to see things more clearly. It was really all about money, not religion. She wants to be a martyr, to save me from spending a cent. All the rest is rationalization. She must have picked up a load of archaic bullshit from her own parents about the virtues of not being a “burden”—not imposing too much on the next generation, not “ruining the best years of their lives.”

She’d left her cycle in a locker at Central Station. She rode home slowly through the leisurely Sunday evening traffic, still feeling drained and shaky, but a little more confident, now that she’d had a chance to think it through. Twelve to eighteen months? She’d raise the money in less than a year. Somehow. She’d show Francesca that she could shoulder the burden—and once that was done, her mother could stop inventing excuses.

Home, she started some vegetables boiling, then went upstairs and checked for mail. There were six items under “Junk,” four under “Autoverse”—and nothing under “Boring But Lucrative.” Since her letter in Autoverse Review, almost every subscriber had been in touch, with congratulations, requests for more data, offers of collaboration, and a few borderline crank calls full of misunderstandings and complaints. Her success with A. lamberti had even made the big time—a slightly less specialized journal, Cellular Automaton World. It was all strangely anticlimactic—and in a way, she was glad of that; it put things in perspective.

She trashed all the junk mail with a sweep of her hand across the touch screen, then sat for a moment gazing at the icons for the Autoverse messages, contemplating doing the same to them. I have to get my act together. Concentrate on earning money, and stop wasting time on this shit.

She ran the first message. A teenage girl in Kansas City complained that she couldn’t duplicate Maria’s results, and proceeded to describe her own tortuous version of the experiment. Maria stopped and deleted the file after viewing twenty seconds; she’d already replied at length to half a dozen like it, and any sense of obligation she’d felt to the “Autoverse community” had vanished in the process.

As she started the second message running, she smelled something burning downstairs, and suddenly remembered that the stove had been brain-dead since Friday—everything had to be watched, and she couldn’t even switch off the hotplates remotely. She turned up the volume on the terminal, and headed for the kitchen.

The spinach was a blackened mess. She threw the saucepan across the narrow room; it rebounded, almost to her feet. She picked it up again and started smashing it against the wall beside the stove, until the tiles began to crack and fall to the floor. Damaging the house was more satisfying than she’d ever imagined; it felt like rending her clothes, like tearing out her hair, like self-mutilation. She pounded the wall relentlessly, until she was breathless, giddy, running with sweat, her face flushed with a strange heat she hadn’t felt since childhood tantrums. Her mother touched her cheek with the back of her hand, brushing away tears of anger. The cool skin, the wedding ring. “Sssh. Look at the state you’re in. You’re burning up!”

After a while, she calmed down, and noticed that the message was still playing upstairs; the sender must have programmed it to repeat indefinitely until she acknowledged it. She sat on the floor and listened.

“My name is Paul Durham. I read your article in Autoverse Review. I was very impressed by what you’ve done with A. lamberti—and if you think you might be interested in being funded to take it further, call me back on this number and we can talk about it.”

Maria had to listen three more times before she was certain she’d understood the message. Being funded to take it further. The phrasing seemed deliberately coy and ambiguous, but in the end it could only really mean one thing.

Some idiot was offering her a job.

+ + +

When Durham asked to meet her in person, Maria was too surprised to do anything but agree. Durham said he lived in north Sydney, and suggested that they meet the next morning in the city, at the Market Street Cafe. Maria, unable to think of a plausible excuse on the spot, just nodded—thankful that she’d made the call through a software filter which would erase any trace of anxiety from her face and tone of voice. Most programming contracts did not involve interviews, even by phone—the tendering process was usually fully automated, based entirely on the quotes submitted and the tenderer’s audited performance record. Maria hadn’t faced an interview in the flesh since she’d applied for part-time cleaning jobs as a student.

It was only after she’d broken the connection that she realized she still had no idea what Durham wanted from her. A real Autoverse fanatic might, just conceivably, part with money for the privilege of collaborating with her—perhaps footing the bills for computer time, for the sake of sharing the kudos of any further results. It was hard to think of any other explanation.

Maria lay awake half the night, looking back on the brief conversation, wondering if she was missing something blindingly obvious—wondering if it could be some kind of hoax. Just before two, she got up and did a hasty literature search of Autoverse Review and a handful of other cellular automaton journals. There were no articles by anyone named Durham.

Around three o’clock, she gave up pondering the question and managed to force herself to sleep. She dreamed that she was still awake, distraught at the news of her mother’s illness—and then, realizing that she was only dreaming, cursed herself angrily because this proof of her love was nothing but an illusion.

8 (Remit not paucity) NOVEMBER 2050

Thomas took the elevator from his office to his home. In life, the journey had been a ten-minute ride on the S-Bahn, but after almost four subjective months he was gradually becoming accustomed to the shortcut. Today, he began the ascent without giving it a second thought—admiring the oak panelling, lulled by the faint hum of the motor—but halfway up, for no good reason, he suffered a moment of vertigo, as if the elegant coffin had gone into free fall.

When first resurrected, he’d worried constantly over which aspects of his past he should imitate for the sake of sanity, and which he should discard as a matter of honesty. A window with a view of the city seemed harmless enough—but to walk, and ride, through an artificial crowd scene struck him as grotesque, and the few times he’d tried it, he’d found it acutely distressing. It was too much like life—and too much like his dream of one day being among people again. He had no doubt that he would have become desensitized to the illusion with time, but he didn’t want that. When he finally inhabited a telepresence robot as lifelike as his lost body—when he finally rode a real train again, and walked down a real street—he didn’t want the joy of the experience dulled by years of perfect imitation.

He had no wish to delude himself—but apart from declining to mimic his corporeal life to the point of parody, it was hard to define exactly what that meant. He baulked at the prospect of the nearest door always opening magically onto his chosen destination, and he had no desire to snap his fingers and teleport. Acknowledging—and exploiting—the unlimited plasticity of Virtual Reality might have been the most “honest” thing to do… but Thomas needed a world with a permanent structure, not a dream city which reconfigured itself to his every whim.

Eventually, he’d found a compromise. He’d constructed an auxiliary geography—or architecture—for his private version of Frankfurt; an alternative topology for the city, in which all the buildings he moved between were treated as being stacked one on top of the other, allowing a single elevator shaft to link them all. His house “in the suburbs” began sixteen stories “above” his city office; in between were board rooms, restaurants, galleries and museums. Having decided upon the arrangement, he now regarded it as immutable—and if the view from each place, once he arrived, blatantly contradicted the relationship, he could live with that degree of paradox.

Thomas stepped out of the elevator into the ground floor entrance hall of his home. The two-story building, set in a modest ten hectares of garden, was his alone—as the real-world original had been from the time of his divorce until his terminal illness, when a medical team had moved in. At first, he’d had cleaning robots gliding redundantly through the corridors, and gardening robots at work in the flower beds—viewing them as part of the architecture, as much as the drain pipes, the air-conditioning grilles, and countless other “unnecessary” fixtures. He’d banished the robots after the first week. The drain pipes remained.

His dizziness had passed, but he strode into the library and poured himself a drink from two cut-glass decanters, a bracing mixture of Confidence and Optimism. With a word, he could have summoned up a full mood-control panel—an apparition which always reminded him of a recording studio’s mixing desk—and adjusted the parameters of his state of mind until he reached a point where he no longer wished to change the settings… but he’d become disenchanted with that nakedly technological metaphor. Mood-altering “drugs,” here, could function with a precision, and a lack of side effects, which no real chemical could ever have achieved—pharmacological accuracy was possible, but hardly mandatory—and it felt more natural to gulp down a mouthful of “spirits” for fortification than it did to make adjustments via a hovering bank of sliding potentiometers.

Even if the end result was exactly the same.

Thomas sank into a chair as the drink started to take effect—as a matter of choice, it worked gradually, a pleasant warmth diffusing out from his stomach before his brain itself was gently manipulated—and began trying to make sense of his encounter with Paul Durham.

You have to let me show you exactly what you are.

There was a terminal beside the chair. He hit a button, and one of his personal assistants, Hans Löhr, appeared on the screen.

Thomas said casually, “Find out what you can about my visitor, will you?”

Löhr replied at once, “Yes, sir.”

Thomas had six assistants, on duty in shifts around the clock. All flesh-and-blood humans—but so thoroughly wired that they were able to switch their mental processes back and forth between normal speed and slowdown at will. Thomas kept them at a distance, communicating with them only by terminal; the distinction between a visitor “in the flesh” and a “mere image” on a screen didn’t bear much scrutiny, but in practice it could still be rigorously enforced. He sometimes thought of his staff as working in Munich or Berlin… “far enough away” to “explain” the fact that he never met them in person, and yet “near enough” to make a kind of metaphorical sense of their ability to act as go-betweens with the outside world. He’d never bothered to find out where they really were, in case the facts contradicted this convenient mental image.

He sighed, and took another swig of C & O. It was a balancing act, a tightrope walk. A Copy could go insane, either way. Caring too much about the truth could lead to a pathological obsession with the infrastructure—the algorithms and optical processors, the machinery of “deception” which lay beneath every surface. Caring too little, you could find yourself gradually surrendering to a complacent fantasy in which life had gone on as normal, and everything which contradicted the illusion of ordinary physical existence was avoided, or explained away.

Was that Durham’s real intention? To drive him mad?

Thomas had ordered the usual cursory screening before letting Durham in, revealing only that the man worked as a salesman for Gryphon Financial Products—a moderately successful Anglo-Australian company—and that he possessed no criminal record. Elaborate precautions were hardly warranted; visitors could do no harm. Thomas’s VR consultants had assured him that nothing short of tampering with the hardware in situ could ever damage or corrupt the system; no mere signal coming down the fiber from the outside world could penetrate the protected layers of the software. Visitors who wreaked havoc, introducing viruses by the fiendishly clever binary-modulated snapping of their fingers, were the stuff of fiction. (Literally; Thomas had seen it happen once on The Unclear Family.)

Durham had said: “I’m not going to lie to you. I’ve spent time in a mental institution. Ten years. I suffered delusions. Bizarre, elaborate delusions. And I realize, now, that I was seriously ill. I can look back and understand that.

“But at the very same time, I can look back and remember what it was that I believed was happening when I was insane. And without for one moment ceasing to acknowledge my condition, I still find those memories so convincing… “

Thomas’s skin crawled. He raised his glass… and then put it down. He knew that if he kept on drinking, nothing the man had said would unsettle him in the least—but he hadn’t drunk enough, yet, to be absolutely sure that that was what he wanted.

“If you’re not prepared to perform the experiment yourself, at least think about the implications. Imagine that you’ve modified the way in which you’re computed—and imagine what the consequences would be. A gedanken experiment—is that too much to ask for? In a sense, that’s all I ever performed myself.”

The terminal chimed. Thomas took the call. Löhr said, “I have a preliminary report on Paul Durham. Would you like me to read it?”

Thomas shook his head. “I’ll view the file.”

He skimmed it, at level one detail. Paul Kingsley Durham. Born in Sydney on June 6, 2000. Parents: Elizabeth Anne Maddox and John Arthur Durham… joint owners of a delicatessen in the Sydney suburb of Concord, from 1996 to 2032… retired to Mackay, Queensland… now both deceased by natural causes.

Educated at a government high school. 2017: Higher School Certificate aggregate score in third percentile; best subjects physics and mathematics. 2018: completed one year of a science degree at Sydney University, passed all examinations but discontinued studies. 2019 to 2023: traveled in Thailand, Burma, India, Nepal. 2024: on return to Australia, diagnosed with an organic delusional syndrome, probably congenital… condition partly controlled by medication. Numerous casual laboring jobs until May, 2029. Condition deteriorating… disability pension granted January, 2031. Committed to Psychiatric Ward of Blacktown Hospital on September 4, 2035.

Corrective nanosurgery to the hippocampus and prefrontal cerebral cortex performed on November 11, 2045… declared a complete success.

Thomas switched to level two, to fill in the ten-year gap, but found little more than a long list of the drugs, neural grafts, and gene-therapy vectors which had been injected into Durham’s skull during that period, to no apparent benefit. There were frequent notes that the treatments had been tested first on a set of partial brain models, but hadn’t worked in practice. Thomas wondered if Durham had been told about this—and wondered what the man imagined happened when a drug was evaluated on fifteen separate models of different regions of the brain, which, taken together, encompassed the entire organ…

2046 to 2048: studying finance and administration at Macquarie University. 2049: graduated with first class honors, and immediately hired by Gryphon as a trainee salesman. As of January 17, 2050, working in the Artificial Intelligence Division.

Which meant selling protection, in various guises, to Copies who were afraid that their assets were going to be pulled out from under them. Durham’s job description would certainly cover spending long hours as a visitor—if not quite stretching to matters like disclosing details of his personal psychiatric history, or suggesting metaphysical gedanken experiments to his clients. Or indeed, wasting time on Copies obviously far too secure to need Gryphon’s services.

Thomas leaned back from the terminal. It was almost too simple: Durham had fooled his doctors into believing that they’d cured him—and then, with typical paranoid ingenuity and tenacity, he’d set about getting himself into a position where he could meet Copies, share the Great Truth that had been revealed to him… and try to extract a little money in the process.

If Thomas contacted Gryphon and told them what their mad salesman was up to, Durham would certainly lose his job, probably end up in an institution again—and hopefully benefit from a second attempt at nanosurgery. Durham probably wasn’t harming anybody… but ensuring that he received treatment was, surely, the kindest thing to do.

A confident, optimistic person would make the call at once. Thomas eyed his drink, but decided to hold off a little longer before drowning the alternatives.

Durham had said: “I understand that everything I believe I’ve experienced was “due to” my illness—and I know there’s no easy way to persuade you that I’m not still insane. But even if that were true… why should it make the question I’ve raised any less important to you?

“Most flesh-and-blood humans live and die without knowing or caring what they are—scoffing at the very idea that it should matter. But you’re not flesh and blood, and you can’t afford the luxury of ignorance.”

Thomas rose and walked over to the mirror above the fireplace. Superficially, his appearance was still based largely on his final scan; he had the same unruly thick white hair, the same loose, mottled, translucent eighty-five-year-old skin. He had the bearing of a young man, though; the model constructed from the scan file had been thoroughly rejuvenated, internally, sweeping away sixty years’ worth of deterioration in every joint, every muscle, every vein and artery. He wondered if it was only a matter of time before vanity got the better of him and he did the same with his appearance. Many of his business associates were un-aging gradually—but a few had leaped back twenty, thirty, fifty years, or changed their appearances completely. Which was most honest? Looking like an eighty-five-year-old flesh-and-blood human (which he was not), or looking the way he’d prefer to look… prefer to be… given the choice. And he did have the choice.

He closed his eyes, put his fingertips to his cheek, explored the damaged skin. If he believed these ruins defined him, they defined him… and if he learned to accept a new young body, the same would be true of it. And yet, he couldn’t shake the notion that external rejuvenation would entail nothing more than constructing a youthful “mask”… while his “true face” continued to exist—and age—somewhere. Pure Dorian Gray—a stupid moralistic fable stuffed with “eternal” verities long obsolete.

And it was good just to feel healthy and vigorous, to be free of the arthritis, the aches and cramps and chills, the shortness of breath he could still remember vividly. Anything more seemed too easy, too arbitrary. Any Copy could become a Hollywood Adonis in an instant. And any Copy could outrace a bullet, lift a building, move a planet from its course.

Thomas opened his eyes, reached out and touched the surface of the mirror, aware that he was avoiding making a decision. But one thing still bothered him.

Why had Durham chosen him? The man might be deluded—but he was also intelligent and rational on some level. Of all the Copies whose insecurities he might have tried to exploit, why choose one with a watertight setup, secure hardware, a well-managed trust fund? Why choose a target who appeared to have absolutely nothing to fear?

Thomas felt the vertigo returning. It had been sixty-five years. Not one newspaper story or police report had mentioned his name; no database search, however elaborate, could link him to Anna. Nobody alive could know what he’d done—least of all a fifty-year-old ex-psychiatric patient from the other side of the world.

Even the man who’d committed the crime was dead. Thomas had seen him cremated.

Did he seriously think that Durham’s offer of sanctuary was some elaborately coded euphemism for not dredging up the past? Blackmail?

No. That was ludicrous.

So why not make a few calls, and have the poor man seen to? Why not pay for him to be treated by the best Swiss neurosurgeon (who’d verify the procedure in advance, on the most sophisticated set of partial brain models…)

Or did he believe there was a chance that Durham was telling the truth? That he could run a second Copy, in a place nobody could reach in a billion years?

The terminal chimed. Thomas said, “Yes?”

Heidrich had taken over from Löhr; sometimes the shifts seemed to change so fast that it made Thomas giddy. “You have a meeting of the Geistbank board in five minutes, sir.”

“Thank you, I’ll be right down.”

Thomas checked his appearance in the mirror. He said, “Comb me.” His hair was made passably tidy, his complexion less pale, his eyes clear; certain facial muscles were relaxed, and others tightened. His suit required no attention; as in life, it could not be wrinkled.

He almost laughed, but his newly combed expression discouraged it. Expediency, honesty, complacency, insanity. It was a tightrope walk. He was ninety years old by one measure, eighty-five-and-a-half by another—and he still didn’t know how to live.

On his way out, he picked up his Confidence & Optimism and poured it on the carpet.

9 (Rip, tie, cut toy man) JUNE 2045

Paul took the stairs down, and circled the block a few times, hoping for nothing more than to forget himself for a while. He was tired of having to think about what he was, every waking moment. The streets around the building were familiar enough, not to let him delude himself, but at least to allow him to take himself for granted.

It was hard to separate fact from rumor, but he’d heard that even the giga-rich tended to live in relatively mundane surroundings, favoring realism over power fantasies. A few models-of-psychotics had reportedly set themselves up as dictators in opulent palaces, waited on hand and foot, but most Copies aimed for an illusion of continuity. If you desperately wanted to convince your-self that you were the same person as your memories suggested, the worst thing to do would be to swan around a virtual antiquity (with mod cons), pretending to be Cleopatra or Ramses II.

Paul didn’t believe that he “was” his original. He knew he was nothing but a cloud of ambiguous data. The miracle was that he was capable of believing that he existed at all.

What gave him that sense of identity?

Continuity. Consistency. Thought following thought in a coherent pattern.

But where did that coherence come from?

In a human, or a Copy being run in the usual way, the physics of brain or computer meant that the state of mind at any one moment directly influenced the state of mind that followed. Continuity was a simple matter of cause and effect; what you thought at time A affected what you thought at time B affected what you thought at time C…

But when his subjective time was scrambled, the flow of cause and effect within the computer bore no relationship whatsoever to the flow of his experience—so how could it be an essential part of it? When the program spelled out his life DBCEA, but it still felt exactly like ABCDE… then surely the pattern was all, and cause and effect were irrelevant. The whole experience might just as well have arisen by chance.

Suppose an intentionally haywire computer sat for a thousand years or more, twitching from state to state in the sway of nothing but electrical noise. Might it embody consciousness?

In real time, the answer was: probably not—the probability of any kind of coherence arising at random being so small. Real time, though, was only one possible reference frame; what about all the others? If the states the machine passed through could be rearranged in time arbitrarily, then who could say what kind of elaborate order might emerge from the chaos?

Paul caught himself. Was that fatuous? As absurd as insisting that every room full of monkeys really did type the complete works of Shakespeare—they just happened to put the letters in a slightly different order? As ludicrous as claiming that every large-enough quantity of rock contained Michelangelo’s David, and every warehouse full of paint and canvas contained the complete works of Rembrandt and Picasso—not in any mere latent form, awaiting some skillful forger to physically rearrange them, but solely by virtue of the potential redefinition of the coordinates of space-time?

For a statue or a painting, yes, it was a joke. Where was the observer who perceived the paint to be in contact with the canvas, who saw the stone figure suitably delineated by air?

If the pattern in question was not an isolated object, though, but a self-contained world, complete with at least one observer to join up the dots from within…

There was no doubt that it was possible. He’d done it. In the final trial of the second experiment, he’d assembled himself and his surroundings—effortlessly—from the dust of randomly scattered moments, from apparent white noise in real time. True, what the computer had done had been contrived, guaranteed to contain his thoughts and perceptions coded into its seemingly aimless calculations. But given a large enough collection of truly random numbers, there was no reason to believe that it wouldn’t include, purely by accident, hidden patterns as complex and coherent as the ones which underlay him.

And wouldn’t those patterns, however scrambled they might be in real time, be conscious of themselves, just as he’d been conscious, and piece their own subjective world together, just as he had done?

Paul returned to the apartment, fighting off a sense of giddiness and unreality. So much for forgetting himself; he felt more charged than ever with the truth of his strange nature.

Did he still want to bale out? No. No! How could he declare that he’d happily wake and forget himself—wake and “reclaim” his life—when he was beginning to glimpse the answers to questions which his original had never even dared to ask?

10 (Remit not paucity) NOVEMBER 2050

Maria arrived at the cafe fifteen minutes early—to find Durham already there, seated at a table close to the entrance. She was surprised, but relieved; with the long wait she’d been expecting suddenly canceled, she had no time to grow nervous. Durham spotted her as she walked in; they shook hands, exchanged pleasantries, ordered coffee from the table’s touch-screen menus. Seeing Durham in the flesh did nothing to contradict the impression he’d made by phone: middle-aged, quiet, conservatively dressed; not exactly the archetypical Autoverse junkie.

Maria said, “I always thought I was the only Autoverse Review subscriber living in Sydney. I’ve been in touch with Ian Summers in Hobart a couple of times, but I never realized there was anyone so close.”

Durham was apologetic. “There’s no reason why you would have heard of me. I’m afraid I’ve always confined myself to reading the articles; I’ve never contributed anything or participated in the conferences. I don’t actually work in the Autoverse, myself. I don’t have the time. Or the skills, to be honest.”

Maria absorbed that, trying not to appear too startled. It was like hearing someone admit that they studied chess but never played the game.

“But I’ve followed progress in the field very closely, and I can certainly appreciate what you’ve done with A. lamberti. Perhaps even more so than some of your fellow practitioners. I think I see it in a rather broader context.”

“You mean… cellular automata in general?”

“Cellular automata, artificial life.”

“They’re your main interests?”

“Yes.”

But not as a participant? Maria tried to imagine this man as a patron of the artificial life scene, magnanimously sponsoring promising young practitioners; Lorenzo the Magnificent to the Botticellis and Michelangelos of cellular automaton theory.

It wouldn’t wash. Even if the idea wasn’t intrinsically ludicrous, he just didn’t look that rich.

The coffee arrived. Durham started paying for both of them, but when Maria protested, he let her pay for herself without an argument—which made her feel far more at ease. As the robot trolley slid away, she got straight to the point. “You say you’re interested in funding research that builds on my results with A. lamberti. Is there any particular direction—?”

“Yes. I have something very specific in mind.” Durham hesitated. “I still don’t know the best way to put this. But I want you to help me… prove a point. I want you to construct a seed for a biosphere.”

Maria said nothing. She wasn’t even sure that she’d heard him correctly. A seed for a biosphere was terraforming jargon—for all the plant and animal species required to render a sterile, but theoretically habitable planet ecologically stable. She’d never come across the phrase in any other context.

Durham continued. “I want you to design a pre-biotic environment—a planetary surface, if you’d like to think of it that way—and one simple organism which you believe would be capable, in time, of evolving into a multitude of species and filling all the potential ecological niches.”

“An environment? So… you want a Virtual Reality landscape?” Maria tried not to look disappointed. Had she seriously expected to be paid to work in the Autoverse? “With microscopic primordial life? Some kind of… Precambrian theme park, where the users can shrink to the size of algae and inspect their earliest ancestors?” For all her distaste for patchwork VR, Maria found herself almost warming to the idea. If Durham was offering her the chance to supervise the whole project—and the funds to do the job properly—it would be a thousand times more interesting than any of the tedious VR contracts she’d had in the past. And a lot more lucrative.

But Durham said, “No, please—forget about Virtual Reality. I want you to design an organism, and an environment—in the Autoverse—which would have the properties I’ve described. And forget about Precambrian algae. I don’t expect you to recreate ancestral life on Earth, translated into Autoverse chemistry—if such a thing would even be possible. I just want you to construct a system with… the same potential.”

Maria was now thoroughly confused. “When you mentioned a planetary surface, I thought you meant a full-scale virtual landscape—a few dozen square kilometers. But if you’re talking about the Autoverse… you mean a fissure in a rock on a seabed, something like that? Something vaguely analogous to a microenvironment on the early Earth? Something a bit more ‘natural’ than a culture dish full of two different sugars?”

Durham said, “I’m sorry, I’m not making myself very clear. Of course you’ll want to try out the seed organism in a number of microenvironments; that’s the only way you’ll be able to predict with any confidence that it would actually survive, mutate, adapt… flourish. But once that’s established, I’ll want you to describe the complete picture. Specify an entire planetary environment which the Autoverse could support—and in which the seed would be likely to evolve into higher lifeforms.”

Maria hesitated. She was beginning to wonder if Durham had any idea of the scale on which things were done in the Autoverse. “What exactly do you mean by a ‘planetary environment’?”

“Whatever you think is reasonable. Say—thirty million square kilometers?” He laughed. “Don’t have a heart attack; I don’t expect you to model the whole thing, atom by atom. I do realize that all the computers on Earth couldn’t handle much more than a tide pool. I just want you to describe the essential features. You could do that in a couple of terabytes—probably less. It wouldn’t take much to sum up the topography; it doesn’t matter what the specific shape of every mountain and valley and beach is—all you need is a statistical description, a few relevant fractal dimensions. The meteorology and the geochemistry—for want of a better word—will be a little more complex. But I think you know what I’m getting at. You could summarize everything that matters about a pre-biotic planet with a relatively small amount of data. I don’t expect you to hand over a giant Autoverse grid which contains every atom in every grain of sand.”

Maria said, “No, of course not.” This was getting stranger by the minute. “But… why specify a whole ‘planet’—in any form?”

“The size of the environment, and the variation in climate and terrain, are important factors. Details like that will affect the number of different species which arise in isolation and later migrate and interact. They certainly made a difference to the Earth’s evolutionary history. So they may or may not be crucial, but they’re hardly irrelevant.”

Maria said carefully, “That’s true—but nobody will ever be able to run a system that big in the Autoverse, so what’s the point of describing it? On Earth, the system is that big, we’re stuck with it. The only way to explain the entire fossil record, and the current distribution of species, is to look at things on a planetary scale. Migration has happened, it has to be taken into account. But… in the Autoverse, it hasn’t happened, and it never will. Effects like that will always be completely hypothetical.”

Durham said, “Hypothetical? Absolutely. But that doesn’t mean the results can’t be considered, can’t be imagined, can’t be argued about. Think of this whole project as… an aid to a thought experiment. A sketch of a proof.”

“A proof of what?”

“That Autoverse life could—in theory—be as rich and complex as life on Earth.”

Maria shook her head. “I can’t prove that. Modeling a few thousand generations of bacterial evolution in a few microenvironment.…”

Durham waved a hand reassuringly. “Don’t worry; I don’t have unrealistic expectations. I said ‘a sketch of a proof,’ but maybe even that’s putting it too strongly. I just want… suggestive evidence. I want the best blueprint, the best recipe you can come up with for a world, embedded in the Autoverse, which might eventually develop complex life. A set of results on the short-term evolutionary genetics of the seed organism, plus an outline of an environment in which that organism could, plausibly, evolve into higher forms. All right, it’s impossible to run a planet-sized world. But that’s no reason not to contemplate what such a world would be like—to answer as many questions as can be answered, and to make the whole scenario as concrete as possible. I want you to create a package so thorough, so detailed, that if someone handed it to you out of the blue, it would be enough—not to prove anything—but to persuade you that true biological diversity could arise in the Autoverse.”

Maria laughed. “I’m already persuaded of that, myself. I just doubt that there could ever be a watertight proof.”

“Then imagine persuading someone a little more skeptical.”

“Who exactly did you have in mind? Calvin and his mob?”

“If you like.”

Maria suddenly wondered if Durham was someone she should have known, after all—someone who’d published in other areas of the artificial life scene. Why else would he be concerned with that debate? She should have done a much wider literature search.

She said, “So what it comes down to is… you want to present the strongest possible case that deterministic systems like the Autoverse can generate a biology as complex as real-world biology—that all the subtleties of real-world physics and quantum indeterminacy aren’t essential. And to deal with the objection that a complex biology might only arise in a complex environment, you want a description of a suitable ‘planet’ that could exist in the Autoverse—if not for the minor inconvenience that the hardware that could run it will almost certainly never be built.”

“That’s right.”

Maria hesitated; she didn’t want to argue this bizarre project out of existence, but she could hardly take it on if she wasn’t clear about its goals. “But when it’s all said and done, how much will this really add to the results with A. lamberti?”

“In one sense, not a lot,” Durham conceded. “As you said, there can never be a proof. Natural selection is natural selection, and you’ve shown that it can happen in the Autoverse; maybe that should be enough. But don’t you think a—carefully designed—thought experiment with an entire planet is a bit more… evocative… than any number of real experiments with Petri dishes? Don’t underestimate the need to appeal to people’s imaginations. Maybe you can see all the consequences of your work, already. Other people might need to have them spelled out explicitly.”

Maria couldn’t argue with any of that—but who handed out research grants on the basis of what was evocative! “Sowhich university—?”

Durham cut her off. “I’m not an academic. This is just an interest of mine. A hobby, like it is with you. I’m an insurance salesman, in real life.”

“But how could you get funding without—?”

“I’m paying for this myself.” He laughed. “Don’t worry, I can afford it; if you take me up on this, you’re not going to be shortchanged, I can promise you that. And I know it’s unusual for an amateur to… subcontract. But like I said, I don’t work in the Autoverse. It would take me five years to learn to do, myself, what I’m asking of you. You’ll be free to publish all of this under your own name, of course—all I ask is a footnote acknowledging financial support.”

Maria didn’t know what to say. Lorenzo the insurance salesman? A private citizen—not even an Autoverse junkie—was offering to pay her to carry out the most abstract piece of programming imaginable: not simulating a nonexistent world, but “preparing” a simulation that would never be performed. She could hardly be disdainful of anyone for throwing their hard-earned money away on “pointless” Autoverse research—but everything that had driven her to do that, herself, revolved around firsthand experience. However much intellectual pleasure it had given her, the real obsession, the real addiction, was a matter of putting on the gloves and reaching into that artificial space.

Durham handed her a ROM chip. “There are some detailed notes here—including a few ideas of mine, but don’t feel obliged to follow any of them. What I want is whatever you think is most likely to work, not what’s closest to my preconceptions. And there’s a contract, of course. Have your legal expert system look it over; if you’re not happy with anything, I’m pretty flexible.”

“Thank you.”

Durham stood. “I’m sorry to cut this short, but I’m afraid I have another appointment. Please—read the notes, think it all through. Call me when you’ve made a decision.”

After he’d left, Maria sat at the table, staring at the black epoxy rectangle in her palm, trying to make sense of what had happened.

Babbage had designed the Analytical Engine with no real prospect of seeing it constructed in his lifetime. Space travel enthusiasts had been designing interstellar craft, down to every last nut and bolt, since the 1960s. Terraforming advocates were constantly churning out comprehensive feasibility studies for schemes unlikely to be attempted for a hundred years or more. Why? As aids to thought experiments. As sketches of proofs.

And if Durham, who’d never even worked in the Autoverse, had an infinitely grander vision of its long-term possibilities than she had, then maybe she’d always been too close to it, too wrapped up in the tedious contingencies, to see what he’d seen…

Except that this wasn’t about long-term possibilities. The computer that could run an Autoverse world would be far bigger than the planet it was modeling. If such a device was ever to be constructed, however far into the future, there’d have to be far better reasons for building it than this. It wasn’t a question of a visionary born a generation or two before his time; Autoverse ecology was an entirely theoretical notion, and it always would be. The project was a thought experiment in the purest sense.

It was also too good to be true. The Autoverse addict’s dream contract. But short of some senseless, capricious hoax, why should Durham lie to her?

Maria pocketed the chip and left the cafe, not knowing whether to feel skeptical and pessimistic, or elated—and guilty. Guilty, because Durham—if he was genuine, if he honestly planned to pay her real money for this glorious, senseless exercise—had to be a little insane. If she took this job, she’d be taking advantage of him, exploiting his strange madness.

+ + +

Maria let Aden into the house reluctantly; they usually met at his place, or on neutral ground, but he’d been visiting a friend nearby, and she could think of no excuse to turn him away. She caught a glimpse of the red cloudless sunset behind him, and the open doorway let in the hot concrete smell of dusk, the whirr of evening traffic. After seven hours cloistered in her room, reading Durham’s notes for his Autoverse Garden of Eden, the street outside seemed strange, almost shocking—charged with the two-billion-year gulf between Earth’s equivalent moment of primordial fecundity and all the bizarre consequences.

She walked ahead of Aden down the entrance hall and switched on the light in the living room, while he propped his cycle against the stairs. Alone, the house suited her perfectly, but it took only one more person to make it seem cramped.

He caught up with her and said, “I heard about your mother.”

“How? Who told you?”

“Joe knows one of your cousins in Newcastle. Angela? Is that her name?”

He was leaning sideways against the doorframe, arms folded. Maria said, “Why don’t you come right in if you’re coming in?”

He said, “I’m sorry. Is there anything I can do?”

She shook her head. She’d been planning to ask him how much he could lend her to help with the scan, but she couldn’t raise the subject, not yet. He’d ask, innocently, if Francesca was certain that she wanted to be scanned—and the whole thing would degenerate into an argument about her right to choose a natural death. As if there was any real choice, without the money for a scan.

Maria said, “I saw her yesterday. She’s handling it pretty well. But I don’t want to talk about it right now.”

Aden nodded, then detached himself from the doorway and walked up to her. They kissed for a while, which was comforting in a way, but Aden soon had an erection, and Maria was in no mood for sex. Even at the best of times, it took a willing suspension of disbelief, a conscious decision to bury her awareness of the biological clockwork driving her emotions—and right now, her head was still buzzing with Durham’s suggestion for building a kind of latent diploidism into A. lamberti, a propensity to “mistakenly” make extra copies of chromosomes, which might eventually pave the way to sexual reproduction and all of its evolutionary advantages.

Aden pulled free and went and sat in one of the armchairs.

Maria said, “I think I’ve finally got some work. If I didn’t dream the whole thing.”

“That’s great! Who for?”

She described her meeting with Durham. The commission, the seed.

Aden said, “So you don’t even know what he gets out of this—except not-quite-proving some obscure intellectual point about evolution?” He laughed, incredulous. “How will you know if you’ve not-quite-proved it well enough? And what if Durham disagrees?”

“The contract is all in my favor. He pays the money into a trust fund before I even start. All I have to do is make a genuine effort to complete the project within six months—and if there’s any dispute, he’s legally bound to accept an independent adjudicator’s decision on what constitutes a ‘genuine effort.’ The expert system I hired gave the contract a triple-A rating.”

Aden still looked skeptical. “You should get a second opinion; half the time those things don’t even agree with each other—let alone predict what would happen in court. Anyway, if it all goes smoothly, what do you end up with?”

“Thirty thousand dollars. Not bad, for six months’ work. Plus computing time up to another thirty thousand—billed directly to him.”

“Yeah? How can he afford all this?”

“He’s an insurance salesman. If he’s good, he could be making, I don’t know… two hundred grand a year?”

“Which is one hundred and twenty, after tax. And he’s paying out sixty on this shit?”

“Yes. You have a problem with that? It doesn’t exactly leave him poverty-stricken. And he could be earning twice as much, for all I know. Not to mention savings, investments… tax dodges. His personal finances are none of my business; once the money’s in the trust fund, he can go bankrupt for all I care. I still get paid if I finish the job. That’s good enough for me.”

Aden shook his head. “I just can’t see why he thinks it’s worth it. There are God-knows-how-many-thousand Copies in existence, right now—running half the biggest corporations in the world, in case you hadn’t noticed—and this man wants to spend sixty thousand dollars proving that artificial life can go beyond bacteria?”

Maria groaned. “We’ve been through this before. The Autoverse is not Virtual Reality. Copies are not the human equivalent of A. lamberti. They’re a cheat, they’re a mess. They do what they’re meant to do, very efficiently. But there’s no… underlying logic to them. Every part of their body obeys a different set of ad hoc rules. Okay, it would be insane to try to model an entire human body on a molecular level—but if you’re interested in the way fundamental physics affects biology, Copies are irrelevant, because they have no fundamental physics. The behavior of a Copy’s neurons doesn’t arise from any deeper laws, it’s just a matter of Some “rules for neurons” which are based directly on what’s known about neurons in the human body. But in the human body, that behavior is a consequence of the laws of physics, acting on billions of molecules. With Copies, we’ve cheated, for the sake of efficiency. There are no molecules, and no laws of physics; we’ve just put in the net results—the biology—by hand.”

“And that offends your aesthetic sensibilities?”

“That’s not the point. Copies have their place—and when the time comes, I’d rather be a software mongrel than dead. All I’m saying is, they’re useless for telling you what kind of physics can support what kind of life.”

“A burning question of our time.”

Maria felt herself flush with anger, but she said evenly, “Maybe not. I just happen to find it interesting. And apparently Paul Durham does too. And maybe it’s too abstract a question to qualify as science… maybe working in the Autoverse is nothing but pure mathematics. Or philosophy. Or art. But you don’t seem to have any qualms about spending a year in Seoul, practicing your own useless artform at the Korean taxpayers’ expense.”

“It’s a private university.”

“Korean students’ expense, then.”

“I never said there was anything wrong with you taking the job—I just don’t want to see you get screwed if this man turns out to be lying.”

“What could he possibly have to gain by lying?”

“I don’t know—but I still don’t see what he has to gain if he’s telling the truth.” He shrugged. “But if you’re happy, I’m happy. Maybe it’ll all be okay. And I know, the way things are going, you can’t afford to be picky.”

Picky? Maria started laughing. Discussing this on Aden’s terms was ridiculous. Durham wasn’t stringing her along, wasting her time; he was absolutely serious—his notes proved that. Three hundred pages—months of work. He’d taken the plan as far as he could, short of learning the intricacies of the Autoverse himself.

And maybe she still didn’t understand his motives—but maybe there was nothing to be “understood.” When she’d been immersed in his notes, there’d been no mystery at all. On its own terms, Durham’s plan was… natural, obvious. An end in itself, requiring no dreary explanation rooted in the world of academic glory and monetary gain.

Aden said, “What’s so funny?”

“Never mind.”

He shifted in the chair, and looked at her oddly. “Well, at least you won’t have to spend all your time in Seoul looking for work, now. That would have been a bore.”

“I’m not going to Seoul.”

“You’re joking.”

She shook her head.

“What’s the problem? You can do this job anywhere, can’t you?”

“Probably. Yes. I just—”

Maria felt a twinge of uncertainty. He seemed genuinely hurt. He’d made it clear that he’d go without her, if he had to—but that was understandable. Composer-in-residence was his perfect job—and she had nothing to weigh against that, nothing to lose by accompanying him. He might have put his position more diplomatically, instead of making her feel like optional baggage—but that was neither proof that he was trying to drive her away, nor an unforgivable crime in itself. He was tactless sometimes. She could live with that.

“What’s wrong with you? You’d love it in Seoul. You know you would.”

She said, “I’d love it too much. There’d be too many distractions. This project is going to be hard work, the hardest thing I’ve ever done, and if I can’t give it all my attention, it’s going to be impossible.” It had started as an ad lib excuse, but it was true. She had six months, if not to build a world, at least to sketch one; if she didn’t eat, sleep and breathe it, it would never come together, it would never come to life.

Aden snorted. “That’s ludicrous! You don’t even have to write a program that runs. You said yourself, as long as you make a reasonable effort, whatever you hand over will be good enough. What’s Durham going to say? ‘Sorry, but I don’t think this slime mould would ever invent the wheel’?”

“Getting it right matters to me.”

Aden said nothing. Then, “If you want to stay behind because of your mother, why can’t you just say so?”

Maria was startled. “Because it’s not true.”

He stared at her angrily. “You know, I was going to offer to stay here with you. But you didn’t want to talk about it.”

Maria untangled that. “That’s what you came here to tell me? That if I planned to stay in Sydney because of Francesca, you’d turn down the job in Seoul?”

“Yes.” He said it as if it should have been obvious to her all along. “She’s dying. Do you think I’d walk off and leave you to cope with that alone? What kind of shit do you think I am?”

She’s not dying; she’s going to be scanned.

But she didn’t say that. “Francesca doesn’t care if I go or stay. I offered to move in with her, but she doesn’t want to be looked after by anyone. Let alone by me.”

“Then come to Seoul.”

“Why, exactly? So you won’t feel bad about leaving me? That’s what it all comes down to, isn’t it? Your peace of mind.”

Aden thought about that for a while. Then he said, “All right. Fuck you. Stay.”

He got to his feet and walked out of the room. Maria listened to him fumbling with his cycle, then opening the front door, slamming it closed.

She tidied up in the kitchen, checked the locks, switched off the lights. Then she went upstairs and lay on her bed, leaving the room in darkness, trying to picture the likely course of events over the next few weeks. Aden would phone before he left, trying to patch things up, but she could see how easy it would be, now, to break things off permanently. And now that it had reached that stage, it seemed like the obvious thing to do. She wasn’t upset, or relieved—just calm. It always made her feel that way: burning bridges, driving people away. Simplifying her life.

She’d left the terminal switched on after reading Durham’s ROM; the screen was blank, and supposedly pure black, but as her eyes adapted to the dark she could see it glowing a faint gray. Every now and then there was a brief flash at a random point on the screen—a pixel activated by background radiation, struck by a cosmic ray. She watched the flashes, like a slow rain falling on a window to another world, until she fell asleep.

11 (Remit not paucity) JANUARY 2051

Malcolm Carter presented as a tall, solid, vigorous-looking man in late middle age—and in fact he was fifty-eight, so his visitor’s body might easily have been styled directly on his real one. Peer remembered seeing photographs of Carter in the early thirties, when he rose to prominence as one of the first architect-programmers to concentrate on the needs of Copies, rather than catering to the human visitors who used virtual environments merely for work or entertainment. Visitors had ended up hiring him too, though—visitors like Kate who were on their way in. And Kate had moved in a similar orbit then, a young computer artist snatched out of obscurity in Oregon and adopted by the San Francisco glitterati at about the same time as Carter’s own ascent from a small Arizona software house. Peer wasn’t sure he would have recognized the man from those old magazine shots—but then, nobody continued to look the way they’d looked in the thirties, if they could possibly help it.

Carter shook hands with Peer, and nodded at Kate; Peer wondered, curious but not really jealous, if they were greeting each other a little more warmly in a private detour from the version of the meeting he was seeing. They were standing in a spacious reception area, the walls and high ceiling decorated with a motif of tiered concentric circles moulded into the cream-colored plaster, the floor tiled in black-and-white diamonds. This was Carter’s publicly listed VR address; anyone at all could call the number and “come here.” The room spawned separate versions for separate callers, though; Peer and Kate had taken steps to arrive together, but there had been no risk of them accidentally bumping into one of Carter’s—or Durham’s—wealthy clients.

Carter said, “I hope you don’t mind if we keep this brief and to the point. I don’t like to use inducers for more than twenty-four hours at a stretch.”

Peer said, “It’s good of you to make the time to see us at all.” He cursed himself silently; he was contemplating paying this man a substantial portion of his entire wealth—and trusting him with the fate of an autonomous version of his own consciousness. He had a right to an audience. Still, at a slow-down of sixty…

Carter—if it really was Carter, and not just a convincing mask—pointed out a door at the end of the room. “There’s a rough sketch of the city through there, if you want to take the tour later on; just call out for a guide if you need one. But I expect the city itself’s not your main concern. What you really want to know is, can I fit you safely into the cracks?”

Peer glanced at Kate. She remained silent. She was already convinced; this was all for his benefit.

Carter held out a hand toward the middle of the room. “See that fountain?” A ten-meter-wide marble wedding cake, topped with a winged cherub wrestling a serpent, duly appeared. Water cascaded down from a gushing wound in the cherub’s neck. Carter said, “It’s being computed by redundancies in the sketch of the city. I can extract the results, because I know exactly where to look for them—but nobody else would have a hope in hell of picking them out.”

Peer walked up to the fountain. Even as he approached, he noticed that the spray was intangible; when he dipped his hand in the water around the base he felt nothing, and the motion he made with his fingers left the foaming surface unchanged. They were spying on the calculations, not interacting with them; the fountain was a closed system.

Carter said, “In your case, of course, nobody will need to know the results. Except you—and you’ll know them because you’ll be them.”

Peer replied, almost without thinking, “Not me. My clone.”

“Whatever.” Carter clapped his hands, and a multicolored, three-dimensional lattice appeared, floating in the air above the fountain. “This is a schematic of part of the software running the sketch of the city. Each cube represents a process. Packets of data—those blips of colored light—flow between them.

“There’s nothing so crude as a subset of processes dedicated to the fountain. Every individual process—and every individual packet of data—is involved with some aspect of the city. But there are some slightly inefficient calculations going on here and there, and some ’redundant’ pieces of information being exchanged.” Pin-pricks in a smattering of the cubes, and some of the data, glowed bright blue. “One of the simplest tricks is to use a vector when only a direction is needed—when the magnitude of the vector is irrelevant. Perfectly reasonable operations on the vector, entirely justified in their own context, incidentally perform arithmetic on the magnitude. But that’s just one technique; there are dozens of others.” He clapped his hands again, and everything but the blue highlights vanished. The diagram re-formed, the scattered processes coming together into a compact grid. “The point is, the fountain gets computed along with the city, without any of the software explicitly stealing time for a parasitic task. Every line of every program makes sense in terms of computing the city.”

Peer said, “And if Durham runs your code through an optimizer which rescales all the unnecessary vectors, trims away all the inefficiencies… ?”

Carter shook his head. “I don’t believe he’d meddle with the code at all, but even if he does, optimizers can only track things so far. In the full version of the city, the results of your calculations will propagate so widely that it would take months for any program to deduce that the data’s not actually needed somewhere—that it ultimately makes no difference to the legitimate inhabitants.” He grinned. “Optimizing anything to do with Copies is a subtle business. You must have heard about the billionaire recluse who wanted to run as fast as possible—even though he never made contact with the outside world—so he fed his own code into an optimizer. After analyzing it for a year, the optimizer reported this program will produce no output, and spat out the optimized version—which did precisely nothing.”

Peer laughed, although he’d heard the joke before.

Carter said, “The fact is, the city is so complex, there’s so much going on, that even if it had all been left to chance, I wouldn’t be surprised if there were some quite sophisticated secondary computations taking place, purely by accident. I haven’t gone looking for them, though—it would bum up far too much processor time. And the same applies to anyone searching for you. It’s just not a practical proposition. Why would anyone spend millions of dollars scanning for something which can do no harm?”

Peer gazed up at the blue schematic skeptically. Carter came across as if he knew what he was talking about, but a few plausible-looking graphics proved nothing.

Carter seemed to read his mind. “If you have any doubts, take a look at the software I used.” A large, fat book appeared, floating in front of Peer. “This modifies program A to surreptitiously carry out program B, given A is sufficiently more algorithmically complex than B. What that means, exactly, is in the technical appendix. Try it out, show it to your favorite expert system… verify it any way you like.”

Peer took hold of the book, squeezed it down to credit-card size, and slipped it into the back pocket of his jeans. He said, “There’s no reason why you shouldn’t be able to do everything you claim: piggyback us onto the city, hide us from searches, protect us from optimization. But… why? What do you get out of this? What you’re asking for is nothing, compared to what Durham must be paying you. So why take the risk? Or do you screw all your clients as a matter of principle?”

Carter chose to seem amused, not offended. “The practice of skimming off a percentage of a construction project has a long, honorable tradition. All the more honorable if the client’s needs aren’t seriously compromised. In this case, there’s also some elegant programming involved—worth doing for its own sake. As for the money, I’m charging you enough to cover my costs.” He exchanged a look with Kate—for Peer’s benefit, or he wouldn’t have seen it. “But in the end, I’m only making the offer as a favor. So if you think I’m going to cheat you, you’re welcome to decline.”

Peer changed tack. “What if Durham is cheating his clients? You’re only screwing them out of a few QIPS—but what if Durham doesn’t plan to run the city at all, just vanish with the money? Have you ever seen his hardware? Have you used it?”

“No. But he never claimed—to me—that he had his own hardware. The version of the story I got is that the city’s going to run on the public networks. That’s bullshit, of course; the Copies funding him wouldn’t wear that for a second—it’s just a polite way of telling me that the hardware is none of my business. And as for vanishing with the money, from what I can deduce about his cash flow, he’ll be lucky to break even on the project. Which suggests to me that someone else entirely is handling the true financial arrangements; Durham is just a front man, and the real owner of the hardware will pay him for his troubles, once the whole thing is wrapped up.”

“The owner of what? This hypothetical ‘breakthrough machine’ that nobody’s laid eyes on?”

If he’s persuaded Sanderson and Repetto to pay him, then you can be sure he’s shown them something that he hasn’t shown me.”

Peer was about to protest, but Carter’s expression said: take it or leave it, believe what you like. I’ve done this much for my ex-lover, but the truth is, I don’t care if you’re convinced or not.

Carter excused himself. When he turned and walked away across the room, footsteps echoing in the cavernous space, Peer couldn’t believe he would have hung around for the fifteen real-time minutes it took to reach the exit. Not a busy man like that. In fact, he’d probably conducted two or three other meetings with Copies while he’d been talking to them, dropping in and out of the conversation, leaving a mask to animate his features in his absence.

Kate said, “What’s the worst that can happen? If Durham is a con man, if the city’s a hoax, what have we lost? All money can buy us is QIPS—and you’re the one who’s so sure that it doesn’t matter how slowly we run.”

Peer scowled, still staring at the exit Carter had used, surprised to find himself reluctant to drag his gaze away. The door meant nothing to him. He said, “Half the charm of this lies in stealing a free ride. Or bribing Carter to steal it for us. There’s not much… dignity in stowing away on a ship going nowhere.”

“You could choose not to care.”

“I don’t want to do that. I don’t pretend to be human, but I still have a… core personality. And I don’t want equanimity. Equanimity is death.”

“On the skyscraper—”

“On the skyscraper I rid myself of distractions. And it’s confined to that one context. When I emerge, I still have goals. I still have desires.” He turned to her, reached out and brushed her cheek with his fingers. “You could choose not to care about security. Or QIPS rates, weather control, the politics of computing—you could choose to view all the threatening noises of the outside world as so much flatulence. Then you wouldn’t need, or want, to do this at all.”

Kate left the body he was touching where it was, but took a step backward in another just like it. Peer let his hand drop to his side.

She said, “Once I’m part of this billionaires’ city, I’ll happily forget about the outside world. Once I have all that money and influence devoted to my survival.”

“Do you mean, that will be enough to satisfy you—or do you intend making a conscious decision to be satisfied?”

She smiled enigmatically—and Peer made a conscious decision to be moved by the sight. She said, “I don’t know yet. You’ll have to wait and see.”

Peer said nothing. He realized that, in spite of his doubts, he’d almost certainly follow her—and not just for the shock of creating a second version, not just for the sake of undermining his last anthropomorphic delusions. The truth was, he wanted to be with her. All of her. If he backed out and she went ahead, the knowledge that he’d passed up his one opportunity to have a version of himself accompany her would drive him mad. He wasn’t sure if this was greed or affection, jealousy or loyalty—but he knew he had to be a part of whatever she experienced in there.

It was an unsettling revelation. Peer took a snapshot of his state of mind.

Kate gestured toward the door which led to the sketch of the city.

Peer said, “Why bother with that? There’ll be plenty of time to explore the real thing.”

She looked at him oddly. “Don’t you want to satisfy your curiosity? Now—and forever, for the one who’ll stay behind?”

He thought about it, then shook his head. “One clone will see the finished city. One won’t. Both will share a past when they’d never even heard of the place. The clone outside, who never sees the city, will try to guess what it’s like. The clone inside will run other environments, and sometimes he won’t think about the city at all. When he does, sometimes he’ll mis-remember it. And sometimes he’ll dream about wildly distorted versions of what he’s seen.

“I define all those moments as part of me. So… what is there to be curious about?”

Kate said, “I love it when you go all doctrinaire on me.” She stepped forward and kissed him—then as he reached out to hold her, she slipped away into yet another body, leaving him embracing nothing but dead weight. “Now shut up and let’s go take a look.”

+ + +

Peer doubted that he’d ever know exactly why he’d died. No amount of agonized introspection, tortuous video-postcard interrogation of ex-friends, or even expert system analysis of his final scan file, had brought him any nearer to the truth. The gap was too wide to be bridged; the last four years of his corporeal life had been lost to him—and the events of the period seemed more like an ill-fated excursion into a parallel world than any mere episode of amnesia.

The coroner had returned an open finding. Rock-climbing accidents were rare, the best technology was almost foolproof—but David Hawthorne had scornfully eschewed all the mollycoddling refinements (including the black box implants which could have recorded the actions leading up to his death, if not the motives behind them). No pitons full of microchips, which could have performed ultrasound tomography of the cliff face and computed their own load-bearing capacity; no harness packed with intelligent crash balloons, which could have cushioned his sixty-meter fall onto jagged rocks; no robot climbing partner, which could have carried him twenty kilometers over rugged terrain with a broken spine and delivered him into intensive care as if he’d floated there on a cloud of morphine.

Peer could empathize, to a degree. What was the point of being scanned, only to remain enslaved by an obsolete respect for the body’s fragility? Having triumphed over mortality, how could he have gone on living as if nothing had changed? Every biological instinct, every commonsense idea about the nature of survival had been rendered absurd—and he hadn’t been able to resist the urge to dramatize the transformation.

That didn’t prove that he’d wanted to die.

But whether his death had been pure misfortune, unequivocal suicide, or the result of some insanely dangerous stunt not (consciously) intended to be fatal, the four-years-out-of-date David Hawthorne had awakened in the virtual slums to realize that, personally, he’d given the prospect about as much serious consideration as that of awakening in Purgatory. Whatever he’d come to believe in those missing years, whatever he’d imagined in his last few seconds of life on that limestone overhang, up until his final scan he’d always pictured his virtual resurrection as taking place in the distant future, when either he’d be seriously wealthy, or the cost of computing would have fallen so far that money would scarcely matter.

He’d been forty-six years old, in perfect health; a senior executive with Incite PLC—Europe’s twenty-fifth largest marketing firm—second-in-charge of the interactive targeted mail division. With care, he could have died at the age of a hundred and fifty, to become an instant member of the elite—perhaps, by then, in a cybernetic body barely distinguishable from the real thing.

But having paid for the right not to fear death, at some level he must have confused the kind of abstract, literary, morally-charged, beloved-of-fate immortality possessed by mythical heroes and virtuous believers in the afterlife, with the highly specific free-market version he’d actually signed up for.

And whatever the convoluted psychological explanation for his death, in financial terms the result was very simple. He’d died too soon.

In a real-time week—a few subjective hours—he had gone from a model of flesh and blood in the lavish virtual apartment he’d bought at the time of his first scan, to a disembodied consciousness observing his Bunker. Even that hadn’t been enough to let him cling to his role in the outside world. Full life insurance was not available to people who’d been scanned—let alone those who also indulged in dangerous recreations—and the coroner’s verdict had even ruled out payment from the only over-priced watered-down substitute policy he’d been able to obtain. At a slowdown of thirty, the lowest Bunker-to-real-time factor the income from his investments could provide, communication was difficult, and productive work was impossible. Even if he’d started burning up his capital to buy the exclusive use of a processor cluster, the time-rate difference would still have rendered him unemployable. Copies whose trust funds controlled massive shareholdings, deceased company directors who sat on the unofficial boards which met twice a year and made three or four leisurely decisions, could live with the time-dilated economics of slowdown. Hawthorne had died before achieving the necessary financial critical mass—let alone the kind of director-emeritus status where he could be paid for nothing but his name on the company letterhead.

As the reality of his situation sank in, he’d spiraled into the blackest depression. Any number of expensive, disabling diseases might have dragged him from upper-middle-class comfort into comparative poverty and isolation—but dying “poor” had an extra sting. In corporeal life, he’d happily gone along with the consensus: money as the deepest level of reality, ownership records as the definition of truth… while escaping most weekends to the manicured garden of the English countryside, camping beneath the clouds, clearing his head of the City’s byzantine fictions—reminding himself how artificial, how arbitrary, it all was. He’d never quite deluded himself that he could have lived off the land: “vanishing” into a forest mapped twice a day by EarthSat on a centimeter scale; surviving on the flesh of protected species, tearing the radio-tracking collars off foxes and badgers with his bare teeth; stoically enduring any rare diseases and parasitic infestations to which his childhood vaccinations and polyclonal T-cell boosts hadn’t granted him immunity. The truth was, he almost certainly would have starved, or gone insane—but that wasn’t the point. What mattered was the fact that his genes were scarcely different from those of his hunter-gatherer ancestors of ten thousand years before; that air was still breathable, and free; that sunshine still flooded the planet, still drove the food chain, still maintained a climate in which he could survive. It wasn’t physically impossible, it wasn’t biologically absurd, to imagine life without money.

Watching the screens of his Bunker, he’d looked back on that trite but comforting understanding with a dizzying sense of loss—because it was no longer in his power to distance himself, however briefly, from the mass hallucination of commerce-as-reality, no longer possible to wrench some half-self-mocking sense of dignity and independence out of his hypothetical ability to live naked in the woods. Money had ceased to be a convenient fiction to be viewed with appropriate irony—because the computerized financial transactions which flowed from his investments to the network’s QIPS providers now underpinned everything he thought, everything he perceived, everything he was.

Friendless, bodiless, the entire world he’d once inhabited transformed into nothing but a blur of scenery glimpsed through the window of a high-speed train, David Hawthorne had prepared to bale out.

It was Kate who had interrupted him. She’d been delegated to make a “welcoming call” by a slum-dwellers’ committee, which she’d only joined in the hope that they’d sponsor one of her projects. This was before she’d made the conscious decision not to desire an audience for any of her art, rendering its quota of computing time relative to any other process irrelevant.

Hawthorne’s only contact since his death had been brief recorded messages from ex-friends, ex-lovers, ex-relations and ex-colleagues, all more or less bidding him farewell, as if he’d embarked on a one-way voyage to a place beyond the reach of modern communications. There’d also been an offer of counseling from his scanning clinic’s Resurrection Trauma expert system—first ten subjective minutes absolutely free. When Kate had appeared on his communications screen, synched to his time rate and talking back, he’d poured out his soul to her.

She’d persuaded him to postpone baling out until he’d considered the alternatives. She hadn’t had to argue hard; the mere fact of her presence had already improved his outlook immeasurably. Thousands of Copies, she’d said, survived with slowdown factors of thirty, sixty, or worse—playing no part in human society, earning no money but the passive income from their trust funds, living at their own speed, defining their worth on their own terms. He had nothing to lose by trying it himself.

And if he couldn’t accept that kind of separatist existence? He always had the choice of suspending himself, in the hope that the economics of ontology would eventually shift in his favor—albeit at the risk of waking to find that he’d matched speeds with a world far stranger, far harder to relate to, than the present in fast motion.

For someone whose fondest hope had been to wake in a robot body and carry on living as if nothing had changed, the slums were a shock. Kate had shown him around the Slow Clubs—the meeting places for Copies willing to synch to the rate of the slowest person present. Not a billionaire in sight. At the Cabaret Andalou, the musicians presented as living saxophones and guitars, songs were visible, tangible, psychotropic radiation blasting from the mouths of the singers—and on a good night, a strong enough sense of camaraderie, telepathy, synergy, could by the mutual consent of the crowd take over, melting away (for a moment) all personal barriers, mental and mock-physical, reconstructing audience and performers into a single organism: one hundred eyes, two hundred limbs, one giant neural net resonating with the memories, perceptions and emotions of all the people it had been.

Kate had shown him some of the environments she’d bought—and some she’d built herself—where she lived and worked in solitude. An overgrown, oversized, small-town back garden in early summer, an enhanced and modified childhood memory, where she carved solid sculptures out of nothing but the ten-to-the-ten-thousandth possibilities of color, texture and form. A bleak gray stretch of shoreline under eternally threatening clouds, the sky dark oil on canvas, a painting come to life, where she went to calm herself when she chose not to make the conscious decision to be calm.

She’d helped him redesign his apartment, transforming it from a photorealist concrete box into a system of perceptions which could be as stable, or responsive, as he wished. Once, before sleep, he’d wrapped the structure around himself like a sleeping bag, shrinking and softening it until the kitchen cradled his head and the other rooms draped his body. He’d changed the topology so that every window looked in through another window, every wall abutted another wall; the whole thing closed in on itself in every direction, finite but borderless, universe-as-womb.

And Kate had introduced him to Daniel Lebesgue’s interactive philosophical plays: The Beholder, The Sane Man (his adaptation of Pirandello’s Enrico TV), and, of course, Solipsist Nation. Hawthorne had taken the role of John Beckett, a reluctant Copy obsessed with keeping track of the outside world—who ends up literally becoming an entire society and culture himself. The play’s software hadn’t enacted that fate upon Hawthorne—intended for visitors and Copies alike, it worked on the level of perceptions and metaphors, not neural reconstruction. Lebesgue’s ideas were mesmerizing, but imprecise, and even he had never tried to carry them through—so far as anyone knew. He’d vanished from sight in 2036; becoming a recluse, baling out, or suspending himself, nobody could say. His disciples wrote manifestos, and prescriptions for virtual Utopias; in the wider vernacular, though, to be “Solipsist Nation” simply meant to have ceased deferring to the outside world.

Three subjective weeks—almost four real-time years—after his resurrection, Hawthorne had stepped off the merry-go-round long enough to catch up with the news from outside. There’d been nothing especially dramatic or unexpected in the summaries—no shocking political upheavals, no stunning technological breakthroughs, no more nor less civil war or famine than in the past. The BBC’s headlines of the day: Five hundred people had died in storms in southeast England. The European Federation had cut its intake of environmental refugees by forty percent. Korean investors had gone ahead with a threatened embargo on US government bonds, as part of a trade war over biotechnology tariffs, and utilities had begun disconnecting power, water and communications services from federal buildings. Up-to-the-minute details notwithstanding, it had all seemed as familiar as some brand-name breakfast food: the same texture, the same taste, as he remembered from four, from eight, years before. With his eyes locked on the terminal in front of him, the oddly soothing generic images drawing him in, the three hallucinatory weeks of dancing saxophones and habitable paintings had receded into insignificance, as if they’d been nothing but a vivid dream. Or at least something on another channel, with no risk of being mistaken for news.

Kate had said, “You know, you can sit here forever, watch this forever, if that’s what you want. There are Copies—we call them Witnesses—who refine themselves into… systems… which do nothing but monitor the news, as thoroughly as their slowdown allows. No bodies, no fatigue, no distractions. Pure observers, watching history unfold.”

“That’s not what I want.”

He hadn’t taken his eyes off the screen, though. Inexplicably, he’d started to cry, softly, grieving for something that he couldn’t name. Not the world defined by the news systems; he’d never inhabited that place. Not the people who’d sent him their recorded farewells; they’d been useful at the time, but they meant nothing to him anymore.

“But?”

“But outside is still what’s real to me—even if I can’t be a part of it. Flesh and blood. Solid ground. Real sunlight. It’s still the only world that matters, in the end. I can’t pretend I don’t know that. Everything in here is just beautiful, inconsequential fiction.” Including you. Including me.

Kate had said, “You can change that.”

“Change what? Virtual Reality is Virtual Reality. I can’t transform it into something else.”

“You can change your perspective. Change your attitudes. Stop viewing your experiences here as less than real.”

“That’s easier said than done.”

“But it isn’t.”

She’d summoned up a control panel, shown him the software he could use: a program which would analyze his model-of-a-brain, identify his qualms and misgivings about turning his back on the world—and remove them.

“A do-it-yourself lobotomy.”

“Hardly. There’s no ‘physical’ excision. The program carries out trial-and-error adjustment of synaptic weights, until it finds the minimum possible alteration which achieves the desired goal. A few billion short-lived stripped-down versions of your brain will be tested and discarded along the way, but don’t let that bother you.”

“You’ve run this on yourself?”

She’d laughed. “Yes. Out of curiosity. But it found nothing to change in me. I’d already made up my mind. Even on the outside, I knew this was what I wanted.”

“So… I press a button and there’s someone new sitting here? One instant synthetic satisfied customer? I annihilate myself, just like that?”

“You’re the one who jumped off a cliff.”

“No. I’m the one who didn’t.”

“You won’t ‘annihilate yourself.’ You’ll only change as much as you have to. And you’ll still call yourself David Hawthorne. What more can you ask for? What more have you ever done?”

They’d talked it through for hours, debating the fine philosophical and moral points; the difference between “naturally” accepting his situation, and imposing acceptance upon himself. In the end, though, when he’d made the decision, it had seemed like just another part of the dream, just another inconsequential fiction. In that sense, the old David Hawthorne had been true to his beliefs—even as he rewired them out of existence.

Kate had been wrong about one thing. Despite the perfect continuity of his memories, he’d felt compelled to mark the transition by choosing a new name, plucking the whimsical monosyllable out of thin air.

The “minimum possible alteration”? Perhaps if he had ended up less radically Solipsist Nation, far more of his personality would have to have been distorted for him to have been convinced at all. A few bold necessary cuts had squared the circle, instead of a thousand finicky mutilations.

That first change, though, had cleared the way for many more, a long series of self-directed mutations. Peer (by choice) had no patience with nostalgia or sentimentality; if any part of his personality offended him, he struck it out. Some traits had (most likely) vanished forever: a horde of petty jealousies, vanities, misgivings and pointless obsessions; a tendency to irrational depression and guilt. Others came and went. Peer had acquired, removed and restored a variety of talents, mood predispositions and drives; cravings for knowledge, art and physical experience. In a few subjective days, he could change from an ascetic bodiless student of Sumerian archaeology, to a hedonistic gastronome delighting in nothing more than the preparation and consumption of lavishly simulated feasts, to a disciplined practitioner of Shotokan karate.

A core remained; certain values, certain emotional responses, certain aesthetic sensibilities had survived these transitions unscathed.

As had the will to survive itself.

Peer had once asked himself: Was that kernel of invariants—and the more-or-less unbroken thread of memory—enough? Had David Hawthorne, by another name, achieved the immortality he’d paid for? Or had he died somewhere along the way?

There was no answer. The most that could be said, at any moment, was that someone existed who knew—or believed—that they’d once been David Hawthorne.

And so Peer had made the conscious decision to let that be enough.

12 (Rip, tie, cut toy man) JUNE 2045

Paul switched on the terminal and made contact with his old organic self. The djinn looked tired and frayed; all the begging and bribery required to set up the latest stage of the experiment must have taken its toll. Paul felt more alive than he’d ever felt, in any incarnation; his stomach was knotted with something like fear, but the electric tingling of his skin felt more like the anticipation of triumph. His body was about to be mutilated, carved up beyond recognition—and yet he knew he would survive, suffer no harm, feel no pain.

Squeak. “Experiment three, trial zero. Baseline data. All computations performed by processor cluster number four six two, Hitachi Supercomputer Facility, Tokyo.”

“One. Two. Three.” It was nice to be told where he was, at last; Paul had never visited Japan before. “Four. Five. Six.” And on his own terms, he still hadn’t. The view out the window was Sydney, not Tokyo; why defer to the external geography, when it made no difference at all? “Seven. Eight. Nine. Ten.”

Squeak. “Trial number one. Model partitioned into five hundred sections, run on five hundred processor clusters, distributed globally.”

Paul counted. Five hundred clusters. Five only for the crudely modeled external world; all the rest were allocated to his body—and most to the brain. He lifted his hand to his eyes—and the information flow that granted him motor control and sight traversed tens of thousands of kilometers of optical cable. There was no (perceptible) delay; each part of him simply hibernated when necessary, waiting for the requisite feedback from around the world.

It was, of course, pure lunacy, computationally and economically; Paul guessed that he was costing at least a hundred times as much as usual—not quite five hundred, since each cluster’s capacity was only being partly used—and his slow-down factor had probably risen from seventeen to as much as fifty. Once, it had been hoped that devoting hundreds of computers to each Copy might improve the slowdown problem, not worsen it—but the bottlenecks in shifting data between processor clusters kept even the richest Copies from reducing the factor below seventeen. It didn’t matter how many supercomputers you owned, because splitting yourself between them wasted more time on communications than was saved by the additional computing power.

Squeak. “Trial number two. One thousand sections, one thousand clusters.”

Brain the size of a planetand here I am, counting to ten. Paul recalled the perennial—naive and paranoid—fear that all the networked computers of the world might one day spontaneously give birth to a global hypermind; but he was, almost certainly, the first planet-sized intelligence on Earth. He didn’t feel much like a digital Gaia, though. He felt exactly like an ordinary human being sitting in a room a few meters wide.

Squeak. “Trial number three. Model partitioned into fifty sections and twenty time sets, implemented on one thousand clusters.”

“One. Two. Three.” Paul struggled to imagine the outside world on his own terms, but it was almost impossible. Not only was he scattered across the globe, but widely separated machines were simultaneously computing different moments of his subjective time frame. Was the distance from Tokyo to New York now the length of his corpus callosum? Had the world shrunk to the size of his skull—and vanished from time altogether, except for the fifty computers which contributed at any one time to what he called “the present”?

Maybe not—although in the eyes of some hypothetical space traveler the whole planet was virtually frozen in time, and flat as a pancake. Relativity declared that this point of view was perfectly valid—but Paul’s was not. Relativity permitted continuous deformation, but no cutting and pasting. Why not? Because it had to allow for cause and effect. Influences had to be localized, traveling from point to point at a finite velocity; chop up space-time and rearrange it, and the causal structure would fall apart.

What if you were an observer, though, who had no causal structure? A self-aware pattern appearing by chance in the random twitches of a noise machine, your time coordinate dancing back and forth through causally respectable “real time”? Why should you be declared a second-class being, with no right to see the universe your way? Ultimately, what difference was there between so-called cause and effect, and any other internally consistent pattern?

Squeak. “Trial number four. Model partitioned into fifty sections and twenty time sets; sections and states randomly allocated to one thousand clusters.”

“One. Two. Three.”

Paul stopped counting, stretched his arms wide, stood up slowly. He wheeled around once, to examine the room, checking that it was still intact, still complete. Then he whispered, “This is dust. All dust. This room, this moment, is scattered across the planet, scattered across five hundred seconds or more—but it still holds itself together. Don’t you see what that means?”

The djinn reappeared, but Paul didn’t give him a chance to speak. The words flowed out of him, unstoppable. He understood.

“Imagine… a universe entirely without structure, without shape, without connections. A cloud of microscopic events, like fragments of space-time… except that there is no space or time. What characterizes one point in space, for one instant? Just the values of the fundamental particle fields, just a handful of numbers. Now, take away all notions of position, arrangement, order, and what’s left? A cloud of random numbers.

“That’s it. That’s all there is. The cosmos has no shape at all—no such thing as time or distance, no physical laws, no cause and effect.

“But… if the pattern that is me could pick itself out from all the other events taking place on this planet… why shouldn’t the pattern we think of as ’the universe’ assemble itself, find itself, in exactly the same way? If I can piece together my own coherent space and time from data scattered so widely that it might as well be part of some giant cloud of random numbers… then what makes you think that you’re not doing the very same thing?”

The djinn’s expression hovered between alarm and irritation.

Squeak. “Paul… what’s the point of all this? ‘Space-time is a construct; the universe is really nothing but a sea of disconnected events…’ Assertions like that are meaningless. You can believe it if you want to… but what difference would it make?”

What difference? We perceive—we inhabit—one arrangement of the set of events. But why should that arrangement be unique? There’s no reason to believe that the pattern we’ve found is the only coherent way of ordering the dust. There must be billions of other universes coexisting with us, made of the very same stuff—just differently arranged. If I can perceive events thousands of kilometers and hundreds of seconds apart to be side by side and simultaneous, there could be worlds, and creatures, built up from what we’d think of as points in space-time scattered all over the galaxy, all over the universe. We’re one possible solution to a giant cosmic anagram… but it would be ludicrous to believe that we’re the only one.”

Squeak. Durham snorted. “A cosmic anagram? So where are all the leftover letters? If any of this were true—and the primordial alphabet soup really is random—don’t you think it’s highly unlikely that we could structure the whole thing?”

Paul thought about it. “We haven’t structured the whole thing. The universe is random, at the quantum level. Macroscopically, the pattern seems to be perfect; microscopically, it decays into uncertainty. We’ve swept the residue of randomness down to the lowest level.”

Squeak, The djinn strived visibly for patience. “Paul… none of this could ever be tested. How would anyone ever observe a planet whose constituent parts were scattered across the universe, let alone communicate with its hypothetical inhabitants? What you’re saying might have a certain—purely mathematical—validity: grind the universe into fine enough dust, and maybe it could be rearranged in other ways that make as much sense as the original. If those rearranged worlds are inaccessible, though, it’s all angels on the heads of pins.”

“How can you say that? I’ve been rearranged! I’ve visited another world!”

Squeak. “If you did, it was an artificial world; created, not discovered.”

“Found, created… there’s no real difference.”

Squeak. “What are you claiming? Some influence from this other world flowed into the computers, changed the way the model ran?”

“Of course not! Your pattern hasn’t been violated; the computers did exactly what was expected of them. That doesn’t invalidate my perspective. Stop thinking of explanations, causes and effects; there are only patterns. The scattered events that formed my experience had an internal consistency every bit as real as the consistency in the actions of the computers. And perhaps the computers didn’t provide all of it.”

Squeak. “What do you mean?”

“The gaps, in experiment one. What filled them in? What was I made of, when the processors weren’t describing me? Well… it’s a big universe. Plenty of dust to be me, in between descriptions. Plenty of events—nothing to do with your computers, maybe nothing to do with your planet or your epoch—out of which to construct ten seconds of experience.”

Squeak, The djinn looked seriously worried now. “You’re a Copy in a virtual environment under computer control. Nothing more, nothing less. These experiments prove that your internal sense of space and time is invariant. That’s exactly what we always expected—remember? Come down to Earth. Your states are computed, your memories have to be what they would have been without manipulation. You haven’t visited any other worlds, you haven’t built yourself out of fragments of distant galaxies.”

Paul laughed. “Your stupidity is… surreal. What did you create me for, if you’re not even going to listen to what I have to say? I’ve had a glimpse of the truth behind… everything: space, time, the laws of physics. You can’t shrug that off by saying that what happened to me was inevitable.

Squeak. “Control and subject are still identical.”

“Of course they are! That’s the whole point! Like… gravity and acceleration in General Relativity—it all depends on what you can’t tell apart. This is a new Principle of Equivalence, a new symmetry between observers. Relativity threw out absolute space and time—but it didn’t go far enough. We have to throw out absolute cause and effect!”

Squeak. The djinn muttered, dismayed, “Elizabeth said this would happen. She said it was only a matter of time before you’d lose touch.”

Paul stared at him, jolted back to the mundane. “Elizabeth? You said you hadn’t even told her.”

Squeak. “Well, I have now. I didn’t tell you, because I didn’t think you’d want to hear her reaction.”

“Which was?”

Squeak. “I was up all night arguing with her. She wanted me to shut you down. She said I was… seriously disturbed, to even think about doing this.”

Paul was stung. “What would she know? Ignore her.”

Squeak. Durham frowned apologetically—an expression Paul recognized at once, and his guts turned to ice. “Maybe I should pause you, while I think things over. Elizabeth raised some… valid ethical questions. I think I should talk it through with her again.”

“Fuck that! I’m not here for you to put on ice every time you have a change of heart. And if Elizabeth wants to have a say in my life, she can damn well talk it through with me.”

Paul could see exactly what would happen. If he was paused, Durham wouldn’t restart him—he’d go back to the original scan file and start again from scratch, handling his prisoner differently, hoping to end up with a more cooperative subject. Maybe he wouldn’t even perform the first set of experiments at all.

The ones which had given him this insight.

The ones which had made him who he was.

Squeak. “I need time to think. It would only be temporary. I promise.”

No! You have no right!”

Durham hesitated. Paul felt numb, disbelieving. Some part of him refused to acknowledge any danger—refused to accept that it could be this easy to die. Being paused wouldn’t kill him, wouldn’t harm him, wouldn’t have the slightest effect. What would kill him would be not being restarted. He’d be passively annihilated, ignored out of existence. The fate that befell his own shit.

Durham reached offscreen.

13 (Remit not paucity) FEBRUARY 2051

Maria said, “Recalculate everything up to epoch five, then show me sunrise on Lambert. Latitude zero, longitude zero, altitude one.”

She waited, staring into the blank workspace, fighting the temptation to change her instructions and have the software display every stage of the simulation, which would have slowed things down considerably. After several minutes, a fissured dark plain appeared, raked with silver light. The unnamed sun—dazzling and swollen, and, so low in the sky, too white by far—turned a chain of extinct volcanoes on the horizon into black silhouettes like a row of pointed teeth. In the foreground, the surface looked glassy, inhospitable.

Maria raised her viewpoint to a thousand meters, then sent it skimming east. The terrain repeated itself, the eerily symmetric cones of dead volcanoes the only relief from the fractured igneous plains. This specific, detailed scenery was nothing more than a series of computerized “artist’s impressions,” manufactured on demand from purely statistical data about the planet’s topography; the simulation itself hadn’t dealt with anything so finicky as individual volcanoes. Touring the planet was a wasteful means of finding out anything—but it was hard to resist playing explorer, treating this world as if its secrets had to be deduced painstakingly from its appearance… even when the truth was the exact opposite. Reluctantly, Maria froze the image and went straight to the underlying numerical data. The atmosphere was much too thin, again. And this time, there was almost no aqua at all.

She backtracked through the simulation’s history to see when the aqua had been lost, but this version of Lambert had never possessed significant oceans—or ice caps, or atmospheric vapor. She’d made a slight change in the composition of the primordial gas-and-dust cloud, increasing the proportion of blue and yellow atoms, in the hope that this would ultimately lead to a denser atmosphere for Lambert. Instead, she’d caused more than half of the debris in the Kuiper belt to condense into a whole new stable outer planet. As a consequence, far fewer ice-rich comets from the belt had ended up striking Lambert, robbing it of its largest source of aqua by far—and much of its atmosphere. Gas released by volcanic eruptions provided a poor substitute; the pressure was far too low, and the chemistry was all wrong.

Maria was beginning to wish she’d kept her mouth shut. It had taken her almost an hour on the phone to persuade Durham that it was worth trying to give Lambert a proper astronomical context, and a geological history that stretched back to the birth of its sun.

“If we present this world as a fait accompli, and say: “Look, it can exist in the Autoverse”… the obvious response to that will be: “Yes, it can exist—if you put it there by hand—but that doesn’t mean it’s ever likely to have formed.” If we can demonstrate a range of starting conditions that lead to planetary systems with suitable worlds, that will be one less element of uncertainty to be used against us.”

Durham had eventually agreed, so she’d taken an off-the-shelf planetary-system modeling program—irreverently titled The Laplacian Casino—and adapted it to Autoverse chemistry and physics; not the deep physics of the Autoverse cellular automaton, but the macroscopic consequences of those rules. Mostly, that came down to specifying the properties of various Autoverse molecules: bond energies, melting and boiling points versus pressure, and so on. Aqua was not just water by another name, yellow atoms were not identical to nitrogen—and although some chemical reactions could be translated as if there was a one-to-one correspondence, in the giant fractionating still of a protostellar nebula subtle differences in relative densities and volatilities could have profound effects on the final composition of each of the planets.

There were also some fundamental differences. Since the Autoverse had no nuclear forces, the sun would be heated solely by gravitational energy—the velocity its molecules acquired as the diffuse primordial gas cloud fell in on itself. In the real universe, stars unable to ignite fusion reactions ended up as cold, short-lived brown dwarfs—but under Autoverse physics, gravitational heating could power a large enough star for billions of years. (Units of space and time were not strictly translatable—but everybody but the purists did it. If a red atom’s width was taken to be that of hydrogen, and one grid-spacing per clock-tick was taken as the speed of light, a more or less sensible correspondence emerged.) Similarly, although Planet Lambert would lack internal heating from radioisotope decay, its own gravitational heat of formation would be great enough to drive tectonic activity for almost as long as the sun shone.

Without nuclear fusion to synthesize the elements, their origin remained a mystery, and a convenient gas cloud with traces of all thirty-two—and the right mass and rotational velocity—had to be taken for granted. Maria would have liked to have explored the cloud’s possible origins, but she knew the project would never be finished if she kept lobbying Durham to expand the terms of reference. The point was to explore the potential diversity of Autoverse life, not to invent an entire cosmology.

Gravity in the Autoverse came as close as real-world gravity to the classical, Newtonian inverse-square law for the range of conditions that mattered, so all the usual real-world orbital dynamics applied. At extreme densities, the cellular automaton’s discrete nature would cause it to deviate wildly from Newton—and Einstein, and Chu—but Maria had no intention of peppering her universe with black holes, or other exotica.

In fact, gravity had been seen as an irrelevant side effect of Lambert’s original choice of automaton rules—since running an Autoverse large enough for it to make the slightest difference was blatantly impossible—and several people had tried to remove the redundancy, while leaving everything else intact. Nobody had succeeded, though; their “rationalized” versions had always failed to generate anything remotely like the rich chemistry of the original. A Peruvian mathematician, Ricardo Salazar, had eventually proved that they shouldn’t have bothered: the Autoverse rules were poised on the border between two radically different levels of algorithmic complexity, and any tinkering in the hope of improved efficiency was necessarily self-defeating. The presence or absence of gravity, in itself, had no bearing on Autoverse chemistry—but the roots of both phenomena in the simple automaton rules seemed to be inextricably entwined.

Maria was aiming for a star with four planets. Three small worlds, one giant. The seed-world, Lambert, second from the sun—with a decent-sized moon if possible. Whether or not tidal pools had been a driving force in real-world evolution, life’s bridge from sea to land (and even though the sun itself would cause small tides, regardless), it couldn’t hurt to make Lambert as generally Earth-like as possible, since Earth was still the only example to turn to for inspiration. With so much about terrestrial evolution still in dispute, the safest policy was to cover every factor which might have been significant. The gravitational effects of the other planets would ensure a reasonably complex set of Milankovitch cycles: minor orbital changes and axis wobbles, providing long-term climate variations, ice ages and interglacials. A belt of comets and other debris would complete the picture; not merely supplying an atmosphere, early on, but also offering the chance of occasional mass-extinctions for billions of years to come.

The trick was to ensure that all of these supposedly evolution-enhancing features coincided with a version of Lambert which could support the seed organism in the first place. Maria had half a dozen possible modifications to A. lamberti in mind, to render it self-sufficient, but she was waiting to see what kind of environments were available before making a final decision.

That still left unanswered the question of whether the seed organism—or life of any kind—could have arisen on Lambert, rather than being placed there by human hands. Max Lambert’s original reason for designing the Autoverse had been the hope of observing self-replicating molecular systems—primitive life—arising from simple chemical mixtures. The Autoverse was meant to provide a compromise between real-world chemistry—difficult and expensive to manipulate and monitor in test-tube experiments, and hideously slow to compute in faithful simulations—and the tantalizing abstractions of the earliest “artificial life”: computer viruses, genetic algorithms, self-replicating machines embedded in simple cellular automaton worlds; all trivially easy to compute, but unable to throw much light on the genesis of real-world molecular biology.

Lambert had spent a decade trying to find conditions which would lead to the spontaneous appearance of Autoverse life, without success. He’d constructed A. lamberti—a twelve-year project—to reassure himself that his goal wasn’t absurd; to demonstrate that a living organism could at least function in the Autoverse, however it had come to be there. A. lamberti had permanently side-tracked him; he’d never returned to his original research.

Maria had daydreamed about embarking on her own attempt at abiogenesis, but she’d never done anything about it. That kind of work was open-ended; in comparison, any problems with mutation in A. lamberti seemed utterly tractable and well-defined. And although, in a sense, it went to the heart of what Durham was trying to prove, she was glad he’d chosen to compromise; if he’d insisted on starting his “thought experiment” with a totally sterile world, the uncertainties in the transition from inanimate matter to the simplest Autoverse life would have overwhelmed every other aspect of the project.

She scrapped the desert Planet Lambert and returned to the primordial gas cloud. She popped up a gadget full of slider controls and adjusted the cloud’s composition, taking back half the increases she’d made in the proportions of blue and yellow. Planetology by trial and error. The starting conditions for real-world systems with Earth-like planets had been mapped out long ago, but nobody had ever done the equivalent for the Autoverse. Nobody had ever had a reason.

Maria felt a flicker of unease. Each time she stopped to remind herself that these worlds would never exist—not even in the sense that a culture of A. lamberti “existed”—the whole project seemed to shift perspective, to retreat into the distance like a mirage. The work itself was exhilarating, she couldn’t have asked for anything more, but each time she forced herself to put it all into context—not in the Autoverse, but in the real world—she found herself light-headed, disoriented. Durham’s reasons for the project were so much flimsier than the watertight internal logic of the thing itself; stepping back from the work was like stepping off a rock-solid planet and seeing it turn into nothing but a lightly tethered balloon.

She stood and walked over to the window, and parted the curtains. The street below was deserted; the concrete glowed in the hyperreal glare of the midday sun.

Durham was paying her good moneymoney that would kelp get Francesco scanned. That was reason enough to press on. And if the project was ultimately useless, at least it did no harm; it was better than working on some hedonistic VR resort or some interactive war game for psychotic children. She let the curtain fall back into place and returned to her desk.

The cloud floated in the middle of the workspace, roughly spherical, rendered visible in spite of the fact that its universe was empty of stars. That was a shame; it meant the future citizens of Lambert were destined to be alone. They’d have no prospect of ever encountering alien life—unless they built their own computers, and modeled other planetary systems, other biospheres.

Maria said, “Recalculate. Then show me sunrise again.”

She waited.

And this time—-false colors, by definition—the disk of the sun was bright cherry red, beneath a thick bank of clouds streaked orange and violet, spread across the sky—and the whole scene was repeated, stretched out before her, shimmering, inverted. Mirrored in the face of the waters.

+ + +

By a quarter to eight, Maria was thinking about logging off and grabbing some food. She was still on a high, but she could feel how close she was coming to the point where she’d be useless for the next thirty-six hours if she pushed herself any further.

She’d found a range of starting conditions for the cloud which consistently gave rise to hospitable versions of Lambert, along with all the astronomical criteria she’d been aiming for—except for the large satellite, which would have been a nice touch but wasn’t critical. Tomorrow, she could begin the task of providing A. lamberti with the means of surviving alone on this world, manufacturing its own nutrose from thin air, with the help of sunlight. Other workers had already designed a variety of energy-trapping pigment molecules; the “literal translation” of chlorophyll lacked the right photochemical properties, but a number of useful analogues had been found, and it was a matter of determining which could be integrated into the bacteria’s biochemistry with the fewest complications. Bringing photosynthesis to the Autoverse would be the hardest part of the project, but Maria felt confident; she’d studied Lambert’s notes, and she’d familiarized herself with the full range of techniques he’d developed for adapting biochemical processes to the quirks of Autoverse chemistry. And even if the pigment she chose, for the sake of expediency, wasn’t the most efficient molecule for the task, as long as the seed organism could survive and reproduce it would have the potential to stumble on a better solution itself, eventually.

The potential, if not the opportunity.

She was about to shut down The Laplacian Casino when a message appeared in the foreground of the workspace:

Juno: Statistical analysis of response times and error rates suggests that your link to the JSN is being monitored. Would you like to switch to a more heavily encrypted protocol?

Maria shook her head, amused. It had to be a bug in the software, not a bug on the line. Juno was a public-domain program (free, but all donations welcome) which she’d downloaded purely as a gesture of solidarity with the US privacy lobby. Federal laws there still made bug-detection software, and any half-decent encryption algorithms, illegal for personal use—lest the FBI be inconvenienced—so Maria had sent Juno’s authors a donation to help them fight the good fight. Actually installing the program had been a joke; the idea of anyone going to the trouble of listening in to her conversations with her mother, her tedious VR contract work, or her self-indulgent excursions into the Autoverse, was ludicrous.

Still, the joke had to be carried through. She popped up a word processor on the JSN—the terminal’s local one wouldn’t have shown up to an eavesdropper tapping the fiber—and typed:

Whoever you are, be warned: I’m about to display the Longford Mind-Erasing Fractal Basilisk, so

The doorbell rang. Maria checked the peephole camera’s view. There was a woman on the front step, nobody she knew. Early forties, conservatively dressed. The not-so-subtle give-away was clearly visible behind her: one compact two-seater Mitsubishi “Avalon” electric car. The New South Wales Police Department were probably the only people in the world who’d bought that model, before the Bankstown factory closed down in forty-six. Maria had often wondered why they didn’t give in and fit blue flashing lights to all their supposedly unmarked cars; acknowledging the situation would have been more dignified than carrying on as if nobody knew.

Dredging her memory for recent misdemeanors—but finding none—she hurried downstairs.

“Maria Deluca?”

“Yes.”

“I’m Detective-Sergeant Hayden. Computer Fraud Squad. I’d like to ask you a few questions, if that’s convenient.”

Maria rescanned for guilty secrets; still no trace—but she would have preferred a visitor from Homicide or Armed Robbery, someone who’d clearly come to the wrong house. She said, “Yes, of course. Come in.” Then, as she backed away from the door, “Ah—I nearly forgot, I suppose I should verify… ?”

Hayden, with a thin smile of blatantly insincere approval, let Maria plug her notepad into the socket of her Police Department badge. The notepad beeped cheerfully; the badge knew the private code which matched the current public key being broadcast by the Department.

Seated in the living room, Hayden got straight to the point. She displayed a picture on her notepad.

“Do you know this man?”

Maria cleared her throat. “Yes. His name’s Paul Durham. I’m… working for him. He’s given me some contract programming.” She felt no surprise; just the jolt of being brought down to earth. Of course the Fraud Squad were interested in Durham. Of course the whole fantasy of the last three months was about to unravel before her eyes. Aden had warned her. She’d known it herself. It was a dream contract, too good to be true.

An instant later, though, she backed away from that reaction, furious with herself. Durham had paid the money into the trust fund, hadn’t he? He’d met the costs of her new JSN account. He hadn’t cheated her. Too good to be true was idiot fatalism. Two consenting adults had kept all their promises to each other; the fact that no outsider would understand the transaction didn’t make it a crime. And after all he’d done for her, at the very least she owed him the benefit of the doubt.

Hayden said, “What kind of ‘contract programming’?”

Maria did her best to explain without taking all night. Hayden was—not surprisingly—reasonably computer literate, and even knew what a cellular automaton was, but either she hadn’t heard of the Autoverse, or she wanted to hear it all again from Maria.

“So you believe this man’s paying you thirty thousand dollars… to help him state his position on a purely theoretical question about artificial life?”

Maria tried not to sound defensive. “I’ve spent tens of thousands of dollars on the Autoverse, myself. It’s like a lot of other hobbies; it’s a world unto itself. People can get obsessive, extravagant. It’s no stranger than… building model airplanes. Or reenacting battles from the American Civil War.”

Hayden didn’t argue the point, but she seemed unmoved by the comparisons. “Did you know that Paul Durham sold insurance to Copies?”

“I knew he was an insurance salesman. He told me that himself. Just because he’s not a professional programmer doesn’t mean he can’t—”

“Did you know he was also trying to sell his clients shares in some kind of sanctuary? A place to go—or to send a clone—in case the political climate turned against them?”

Maria blinked. “No. What do you mean—a sanctuary? A privately owned supercomputer? He’s been trying to raise money, form a consortium… ?”

Hayden said flatly, “He’s certainly raising money—but I doubt he’ll ever raise enough to purchase the kind of hardware he’d need for the kind of service he’s offering.”

“So, what are you accusing him of doing? Embarking on a business venture which you don’t happen to believe will be successful?” Hayden said nothing. “Have you spoken to him about this? There might be a simple explanation for whatever you’ve been told. Some senile Copy might have taken his sales pitch for a perpetuity fund the wrong way.” Senile Copy? Well… some postdementia scan file might have proved resistant to the cognitive repair algorithms.

Hayden said, “Of course we’ve spoken to him. He’s refused to cooperate, he won’t discuss the matter. That’s why we’re hoping you’ll be able to assist us.”

Maria’s defiant optimism wavered. If Durham had nothing to hide, why would he refuse to defend himself?

She said, “I don’t see how I can help you. If you think he’s been misleading his clients, go talk to his clients. It’s their testimony you need, not mine.”

There was an awkward pause, then Hayden said, “The testimony of a Copy has no standing; legally, they’re just another kind of computer software.”

Maria opened her mouth, then realized that any excuse she offered would only make her sound more foolish. She salvaged some pride with the silent observation that the legal position of Copies was so farcical that any sane person could have trouble keeping it in mind.

Hayden continued. “Durham could be charged with defrauding the executors of the estates, by means of supplying misleading data to the software they use to advise them. There are precedents for that; it’s like publishing false prospectus information that causes automated share-buying programs to buy your stock. But there’s still the question of evidence. We can interview Copies as an informal source of information, to guide an investigation, but nothing they say will stand up in court.”

Maria recalled an episode of The Unclear Family where a similar problem had arisen. Babette and Larry Unclear had witnessed bank accounts being pilfered, when the relevant data trail had—inexplicably—taken solid form as an accusing tableau of ice-sculptures in their cyber-suburban backyard. She couldn’t recall exactly how the plot had turned out; ten-year-old Leroy had probably done something marginally illegal, but morally unimpeachable, to trick the thieves into giving themselves away to the authorities…

She said, “I don’t know what you expect me to tell you. Durham hasn’t defrauded me. And I don’t know anything about this scheme.”

“But you’re working on it with him.”

“I certainly am not!”

Hayden said drily, “You’re designing a planet for him. What do you think that’s for?”

Maria stared at her blankly for a second, then almost laughed. “I’m sorry, I can’t have explained things very well. I’m designing a planet that “could” exist in the Autoverse, in the broadest sense of the word. It’s a mathematical possibility. But it’s too large to be run on a real computer. It’s not some VR—”

Hayden cut her off. “I understand that perfectly. That doesn’t mean Durham’s clients would have grasped the distinction. Technical details about the Autoverse aren’t exactly general knowledge.”

True. Maria hesitated. But—

“It still makes no sense. For a start, these people would have advisers, researchers, who’d tell them that anyone promising them an Autoverse planet was full of shit. And why would Durham offer them an Autoverse planet—covered in primordial slime—when he could offer them a standard set of VR environments which would be a thousand times more attractive and a thousand times more plausible?”

“I believe he’s offering them both. He’s hired an architect in the US to work on the VR part.”

“But why both? Why not just VR? You couldn’t fit a single Copy into the Autoverse—and if you did, it would die on the spot. It would take fifty or sixty years of research to translate human biochemistry into Autoverse terms.”

“They wouldn’t know that.”

“They could find out in ten seconds flat. Forget about advisers; it would take one call to a knowledge miner, total cost five dollars. So why tell a lie that could be so easily uncovered? What’s the advantage—from a Copy’s point of view—of an Autoverse planet over patchwork VR?”

Hayden was unfazed. “You’re the Autoverse expert. So you tell me.”

“I don’t know.” Maria stood up. She was beginning to feel claustrophobic; she hated having strangers in the house. “Can I get you something to drink? Tea? Coffee?”

“No. But you go ahead—”

Maria shook her head and sat down again; she had a feeling that if she went into the kitchen, she wouldn’t want to return.

She couldn’t see why Durham would refuse to talk to the police, unless he was involved in something dubious enough to have him thrown out of his job, at the very least. Fuck him. He might not have intended to cheat her, but he’d screwed her nonetheless. She wouldn’t get a cent for the work she’d completed; other creditors would have no call on the trust fund if Durham merely went bankrupt—but if the money was the pro-ceeds of crime…

Lorenzo the Magnificent. Yeah.

The worst of it was, for all she knew, Hayden believed she was a willing accomplice. And if Durham intended to remain silent, she’d have to clear her own name.

How?

First, she had to find out about the scam, and untangle her role in it.

She said, “What exactly is he promising these Copies?”

“A refuge. A place where they’ll be safe from any kind of backlash—because they won’t be connected to the outside world. No telecommunications; nothing to trace. He feeds them a long spiel about the coming dark age, when the unwashed masses will no longer put up with being lorded over by rich immortals—and evil socialist governments will confiscate all the supercomputers for weather control.”

Hayden seemed to find the prospect laughable. Maria suspended judgement; what mattered was how Durham’s clients felt, and she could imagine Operation Butterfly making a lot of Copies feel threatened. “So they send their clones in, and slam the door, in case the originals don’t make it through the purges. But then what? How long is this “dark age” supposed to last?”

Hayden shrugged. “Who knows? Hundreds of years? Presumably Durham himself—or some trustworthy successor, several generations later—will decide when it’s safe to come out. The two Copies whose executors filed complaints didn’t wait to hear the whole scenario; they threw him out before he could get down to details like that.”

“He must have approached other Copies.”

“Of course. No one else has come forward, but we have a tentative list of names. All with estates incorporated overseas, unfortunately; I haven’t been able to interview any of them, yet—we’re still working on the jurisdictional red tape. But a few have made it clear already, through their lawyers, that they won’t be willing to discuss the matter—which presumably means that they’ve swallowed Durham’s line, and now they don’t want to hear a word against him.”

Maria struggled to imagine it: No communications. Cut off from reality, indefinitely. A few “Solipsist Nation” Copies might relish the prospect—but most of them had too little money to be the targets of an elaborate scam. And even if Durham’s richest, most paranoid clients seriously believed that the world was on the verge of turning against them… what if things went so badly wrong, outside, that links were never restored? The humans guarding the sanctuary could die out—or just walk away. How could any but the most radically separatist of Copies face the risk of being stranded inside a hidden computer, buried in the middle of a desert somewhere, with no means of discovering for themselves when civilization was worth rejoining—and no means of initiating contact in any case?

Radioisotope power sources could run for thousands of years; multiply redundant hardware of the highest standard could last almost as long, in theory. All these Copies would have, to remember reality by, would be the information they’d brought in with them at the start. If it turned into a one-way trip, they’d be like interstellar colonists, carrying a snapshot of Earth culture off into the void.

Except that interstellar colonists would merely face a growing radio time lag, not absolute silence. And whatever they were leaving behind, at least they’d have something to look forward to: a new world to explore.

A new world—and the possibility of new life.

So what better cure could there be for claustrophobia than the promise of dragging an entire planet into the refuge, seeded with the potential for developing its own exotic life?

Maria didn’t know whether to be outraged or impressed. If she was right, she had to admire Durham’s sheer audacity. When he had asked for a package of results which would persuade “the skeptics” about the prospects for an Autoverse biosphere, he hadn’t been thinking of academics in the artificial life scene. He’d wanted to convince his clients that, even in total isolation, they’d have everything reality could ever offer the human race—including a kind of “space exploration,” complete with the chance of alien contact. And these would be genuine aliens; not the stylish designer creatures from VR games, constructs of nothing but the human psyche; not the slick, unconvincing biomorphs of the high-level phenotype-selection models, the Darwinian equivalent of Platonic ideals. Life which had come the whole tortuous way, molecule by molecule, just like the real thing. Or, almost the whole way; with a biogenesis still poorly understood, Durham had had enough sense to start with “hand-made” microbes—otherwise his clients might never have believed that the planet would bear life at all.

Maria explained the idea, tentatively. “He’d have to have convinced these Copies that running the Autoverse is much faster than modeling real biochemistry—which it is—without being too specific about the actual figures. And I still think it’s a crazy risk to take; anyone could easily find out the truth.”

Hayden thought it over. “Would it matter if they did? If the point of this world is mainly psychological—a place to “escape to” if the worst happens, and reality becomes permanently inaccessible—then it wouldn’t matter how slowly it ran. Once they’d given up hope of reestablishing contact, slowdown would become irrelevant.”

“Yes, but there’s slow—and there’s physically impossible. Sure, they could take in a crude sketch of the planet—which is what Durham’s asked me to provide—but they wouldn’t have a fraction of the memory needed to bring it to life. And even if they found a way around that, it could take a billion years of Autoverse time before the seed organism turned into anything more exciting than blue-green algae. Multiply that by a slow-down of a trillion… I think you get the picture.”

“Flat batteries?”

“Flat universe.”

Hayden said, “Still… if they don’t want to think too seriously about the prospect of ending up permanently trapped, they might not want to look too closely at any of this. Thanks to you, Durham will have a thick pile of impressive technical details that he can wave in their faces, convincing enough to take the edge off their fear of cabin fever. Maybe that’s all they want. The only part that matters, if everything goes smoothly, is the conventional VR—good enough to keep them amused for a couple of real-time centuries—and that checks out perfectly.”

Maria thought this sounded too glib by far, but she let it pass. “What about the hardware? How does that check out?”

It doesn’t. There’ll never be any hardware. Durham will vanish long before he has to produce it.”

“Vanish with what? Money handed over with no questions asked—no safeguards, no guarantees?”

Hayden smiled knowingly. “Money handed over, mostly, for legitimate purposes. He’s commissioned a VR city. He’s commissioned an Autoverse planet. He’s entitled to take a percentage of the fees—there’s no crime in that, so long as it’s disclosed. For the first few months, everything he does will be scrupulously honest. Then at some point, he’ll ask his backers to pay for a consultants’ report—say, a study of suitably robust hardware configurations. Tenders will be called for. Some of them will be genuine—but the most attractive ones will be forged. Later, Durham will claim to have received the report, the “consultants” will be paid… and he’ll never be seen again.”

Maria said, “You’re guessing. You have no idea what his plans are.”

“We don’t know the specifics—but it will be something along those lines.”

Maria slumped back in her chair. “So, what now? What do I do? Call Durham and tell him the whole thing’s off?”

“Absolutely not! Keep working as if nothing had happened—but try to make contact with him more often. Find excuses to talk to him. See if you can gain his trust. See if you can get him to talk about his work. His clients. The refuge.”

Maria was indignant. “I don’t remember volunteering to be your informant.”

Hayden said coolly, “It’s up to you, but if you’re not willing to cooperate, that makes our job very difficult…”

“There’s a difference between cooperation and playing unpaid spy!”

Hayden almost smiled. “If you’re worried about money, you’ll have a far better chance of being paid if you help us to convict Durham.”

“Why? What am I meant to do—try suing him after he’s already gone bankrupt repaying the people he’s cheated?”

“You won’t have to sue him. The court is almost certain to award you compensation as one of the victims—especially if you’ve helped bring the case to trial. There’s a fund, revenue from fines. It doesn’t matter whether Durham can pay you himself.”

Maria digested that. The truth was, it still stank. What she wanted to do was cut her losses and walk away from the whole mess. Pretend it had never happened.

And then what? Go crawling back to Aden for money? There were still no jobs around; she couldn’t afford to write off three months’ work. A few thousand dollars wouldn’t get Francesca scanned—but the lack of it could force her to sell the house sooner than she wanted to.

She said, “What if I make him suspicious? If I suddenly start asking all these questions…”

“Just keep it natural. Anyone in your position would be curious; it’s a strange job he’s given you—he must expect questions. And I know you went along with what he told you at the start, but that doesn’t mean you can’t have given it more thought and decided that there are a few things that still puzzle you.”

Maria said, “All right, I’ll do it.” Had she ever had a choice? “But don’t expect him to tell me the truth. He’s already lied to me; he’s not going to change his story now.”

“Maybe not. But you might be surprised. He might be desperate to have someone to take into his confidence—someone to boast to. Or he might just drop a few oblique hints. Anything’s possible, as long as you keep talking to him.”

When Hayden had left, Maria sat in the living room, too agitated to do anything but run through the whole exchange again in her head. An hour before, she’d been exhausted, but triumphant; now she just felt weary and stupid. Keep working as if nothing had happened! The thought of tackling photo-synthesis in A. lamberti—for the sake, now, of ingratiating herself with the Fraud Squad—was so bizarre it made her giddy.

It was a pity Durham hadn’t been honest with her, and invited her in on the scam. If she’d known all along that she was meant to be helping to screw rich Copies out of their petty cash, at least the work would have had the real-world foundation she’d always felt was missing.

She finally went upstairs, without having eaten. Her connection to the JSN had been logged off automatically, but the message from Juno, locally generated, still hovered in the workspace. As she gestured to the terminal to switch itself off, she wondered if she should have asked Hayden: Is it you who’s been tapping my phone line?

14 (Remit not paucity) FEBRUARY 2051

Seated in his library, Thomas viewed the final report in his knowledge miner’s selection from the last real-time week of news. A journalist in a fur-lined coat appeared to address the camera, standing in light snow in front of the US Supreme Court building—although she was more likely to have been seated in a warm studio, watching a software puppet mime to her words.

“Today’s five-to-one majority decision means that the controversial Californian statute will remain in force. Authorities taking possession of computer storage media to check for simulations of the brain, body or personality of a suspected felon, dead or alive, are not violating the Fourth Amendment rights of either the next of kin or the owners of the computer hardware. Chief Justice Andrea Steiner stressed that the ruling does not affect the status of Copies themselves, one way or another. The software, she said, can be confiscated and examined—but it will not stand trial.”

The terminal blinked back to a menu. Thomas stretched his arms above his head, acutely conscious for a moment of the disparity between his frail appearance and the easy strength he felt in his limbs. He had become his young self again, after all. Become him in the fleshwhether or not he chose to face him in the mirror. But the thought led nowhere.

Thomas had been following the saga of the Californian legislation from the start. He hoped Sanderson and her colleagues knew what they were doing; if their efforts backfired, it could have unpleasant ramifications for Copies everywhere. Thomas’s own public opinion model had shrugged its stochastic shoulders and declared that the effects of the law could go either way, depending on the steps taken to follow through—and several other factors, most of which would be difficult to anticipate, or manipulate.

Clearly, the aim was to shock apathetic US voters into supporting human rights for Copies—lest the alternative be de facto kidnap, mind pillage, and possibly even execution, all without trial. The computer-literate would understand just how useless the law would be in practice—but they’d already been largely won over. The Unclear Family rated highest with the demographics least likely to grasp the technical realities—a storehouse of good will that had yet to be fully exploited. Thomas could see the possibilities. Resurrected blue-collar worker Larry Unclear could turn out to have been under suspicion of murder at the time of his death. Flashback: Misunderstanding in bar leads to heated, highly visible, argument between Larry and guest-star X. Comic escalation to full-scale brawl. Taking advantage of the confusion, guest-star Y smashes a bottle over the skull of guest-star X—while Larry, with his usual endearing ineffectuality, has ended up comatose under a table. The new law could see him dragged from his home and family in the dead of night for a Kafkaesque virtual interrogation, in which his guilty dreams of being responsible are taken to be memories of actually committing the crime… while guest-star Y, still a living human, receives a civilized trial, lies through his teeth, and is acquitted. Son Leroy could save the day somehow, at the last minute, as usual.

Thomas closed his eyes and buried his face in his hands. Most of the room ceased being computed; he pictured himself adrift in Durham’s sea of random numbers, carrying the chair and a fragment of floor with him, the only objects granted solidity by his touch.

He said, “I’m not in any danger.” The room flickered half-way back into existence, subtly modified the sound of his words, then dissolved into static again. Who did he believe would accuse him? There was no one left to care about Anna’s death. He’d outlived them all.

But as long as the knowledge of what he’d done continued to exist, inside him, he could never be certain that it wouldn’t be revealed.

For months after the crime, he’d dreamed that Anna had come to his apartment. He’d wake, sweating and shouting, staring into the darkness of his room, waiting for her to show herself. Waiting for her to tear the skin of normality from the world around him, to reveal the proof of his damnation: blood, fire, insanity.

Then he’d started rising from his bed when the nightmare woke him, walking naked into the shadows, daring her to be there. Willing it. He’d enter every room in the apartment, most of them so dark that he had to feel his way with an outstretched hand, waiting for her fingers suddenly to mesh with his.

Night after night, she failed to appear. And gradually, her absence became a horror in itself; vertiginous, icy. The shadows were empty, the darkness was indifferent. Nothing lay beneath the surface of the world. He could have slaughtered a hundred thousand people, and the night would still have failed to conjure up a single apparition to confront him.

He wondered if this understanding would drive him mad.

It didn’t.

After that, his dreams had changed; there were no more walking corpses. Instead, he dreamed of marching into Hamburg police station and making a full confession.

Thomas stroked the scar on the inside of his right forearm, where he’d scraped himself on the brickwork outside the window of Anna’s room, making his clumsy escape. No one, not even Ilse, had ever asked him to account for it; he’d invented a plausible explanation, but the lie had remained untold.

He knew he could have his memories of the crime erased. Edited out of his original scan file, his current brain model, his emergency snapshots. No other evidence remained. It was ludicrous to imagine that anyone would ever have the slightest reason—let alone the legal right, let alone the power—to seize and examine the data which comprised him… but if it eased his paranoid fears, why not? Why not neutralize his unease at the technical possibility of his mind being read like a book—or a ROM chip—by turning the metaphor, or near-literal truth, to his own advantage? Why not rewrite the last incriminating version of his past? Other Copies exploited what they’d become with inane sybaritic excesses. Why not indulge himself in some peace of mind?

Why not? Because it would rob him of his identity. For sixty-five years, the tug on his thoughts of that one night in Hamburg had been as constant as gravity; everything he’d done since had been shaped by its influence. To tear out the entire tangled strand of his psyche—render half of his remaining memories incomprehensible—would be to leave himself a baffled stranger in his own life.

Of course, any sense of loss, or disorientation, could be dealt with, too, subtracted out… but where would the process of amputation end? Who would remain to enjoy the untroubled conscience he’d manufactured? Who’d sleep the sleep of the just in his bed?

Memory editing wasn’t the only option. Algorithms existed which could transport him smoothly and swiftly into a state of enlightened acceptance: rehabilitated, healed, at peace with himself and his entire uncensored past. He wouldn’t need to forget anything; his absurd fear of incrimination by mind-reading would surely vanish, along with his other neuroses-of-guilt.

But he wasn’t prepared to swallow that fate, either—however blessed he might have felt once the transformation was complete. He wasn’t sure that there was any meaningful distinction between redemption and the delusion of redemption… but some part of his personality—though he cursed it as masochistic and sentimental—baulked at the prospect of instant grace.

Anna’s killer was dead! He’d burnt the man’s corpse! What more could he do, to put the crime behind him?

On his “deathbed,” as his illness had progressed—as he’d flirted giddily every morning with the prospect of ordering his final scan—he’d felt certain that witnessing the fate of his body would be dramatic enough to purge him of his stale, mechanical, relentless guilt. Anna was dead; nothing could change that. A lifetime of remorse hadn’t brought her back. Thomas had never believed that he’d “earned” the right to be free of her—but he’d come to realize that he had nothing left to offer the little tin metronome in his skull but an extravagant ritual of atonement: the death of the murderer himself.

But the murderer had never really died. The corpse consigned to the furnace had been nothing but shed skin. Two days before being scanned, Thomas had lost his nerve and countermanded his earlier instructions: that his flesh-and-blood self be allowed to regain consciousness after the scan.

So the dying human had never woken, never known that he was facing death. And there had been no separate, mortal Thomas Riemann to carry the burden of guilt into the flames.

+ + +

Thomas had met Anna in Hamburg in the summer of 1983, in a railway station cafe. He was in town to run errands for his father. She was on her way to West Berlin, for a concert. Nick Cave and the Bad Seeds.

The cafe was crowded, they shared a table. Anna’s appearance wasn’t striking—dark-haired, green-eyed, her face round and flat. Thomas would never have looked twice at her if they’d passed in the street—but she soon made an impression.

She looked him over appraisingly, then said, “I’d kill for a shirt like that. You have expensive tastes. What do you do to support them?”

Thomas lied carefully. “I was a student. Engineering. Up until a few months ago. It was hopeless, though; I was failing everything.”

“So what do you do now?”

He looked doleful. “My father owns a merchant bank. I went into engineering to try to get away from the family business, but—”

She wasn’t sympathetic at all. “But you screwed up, and now he’s stuck with you?”

“And vice versa.”

“Is he very rich?”

“Yes.”

“And you hate him?”

“Of course.”

She smiled sweetly. “Why don’t I kidnap him for you? You give me all the inside information, and we’ll split the ransom money, fifty-fifty.”

“You kidnap bankers for a living, do you?”

“Not exclusively.”

“I think you work in a record store.”

“You’re wrong.”

“Or a second-hand clothes shop.”

“You’re getting colder.”

“Who are you meeting in Berlin?”

“Just some friends.”

When her train was announced, he asked her for her number. She wrote it on the sleeve of his shirt.

For the next few months whenever he was traveling north, be phoned her. Three times, she made excuses. He almost gave up, but he kept recalling the mocking expression on her face, and he knew he wanted to see her again.

Early in November, she finally said, “Drop round, if you like. I’m not doing anything.”

He’d planned to take her to a nightclub, but she had a child with her, a baby just a few months old. “He’s not mine. I’m looking after him for a friend.” They watched TV, then had sex on the sofa. Climbing off him, Anna said, “You’re really quite sweet.” She kissed him on the cheek, then vanished into the bedroom, locking him out. Thomas fell asleep watching an old John Wayne movie. Two teenage girls with smeared mascara pounded on the door around two in the morning and Anna sold them a plastic sachet of white powder.

Thomas, still on the couch, asked her if the powder was heroin, or cocaine.

“Heroin.”

“Do you use that shit?”

“No.” She regarded him with mild amusement; she didn’t care if he believed her or not.

He woke again at half past five. Anna had gone. The baby was still in his crib, screaming. Thomas changed him and fed him; Anna had shown him where everything was. He wanted a shower, but there was no hot water. He shaved, and left in time for his meeting, telling himself Anna would be back soon. All morning, and all through lunch, he could smell the sour odor of the child’s skin on his hands, and he wondered if the smiling property developers could smell it too.

He phoned from the hotel, paying for the night he hadn’t spent there, knowing that his father would scrutinize his expenses. Anna was home; he’d woken her. Someone nearby grunted with displeasure. Thomas didn’t mention the child.

The next time, he came on a Saturday afternoon, with no need to be anywhere else in a hurry. They met at the Alsterpavillon, drank their coffee looking down on the buffoons in rowboats on the Binnenalster, then went shopping on Jungfernstieg. Thomas paid for the clothes Anna chose, authentic gothic designer trash that looked far worse than the cheapest imitation; it seemed she didn’t really want to dress like him, after all. They walked arm-in-arm from shop to shop, and in the entrance to the most expensive boutique, they stopped and kissed for several minutes, blocking the way of customers trying to get past, then went in and spent a lot of money.

Later, in a nightclub with a bad live band who dressed like the Beatles and did Sex Pistols covers, they ran into Martin, a tall wiry blond youth who Anna introduced as a friend. Martin was all vicious back-slapping amiability, trying so hard to be intimidating that he was almost comical. They all staggered back to Anna’s flat together, and sat on the floor listening to records. When Anna went to the toilet, Martin drew a knife and told Thomas he intended to kill him. He was very drunk. Thomas stood up, kicked him once in the face, breaking his nose, then took away the knife and dragged him moaning out into the hall. Thomas turned him on his side so he wouldn’t choke on the blood, then locked the door.

Anna came out of the bathroom. Thomas told her what had happened. She went out and checked on Martin, and put a pillow under his head.

While Anna was undressing him, Thomas said, “On TV once, I saw an English soldier who’d just come back from Northern Ireland. And he said, ‘It was hell there, but at least it was real. At least I’ve lived now.’” Thomas laughed sadly. “The poor fool had it all upside down. Slaughtering people is real—and living an ordinary life is some kind of dream, some kind of delusion? Poor fucked-up kid.”

He searched Anna for needle marks, but he couldn’t find a single one.

Back in his office in Frankfurt, alone in his apartment, at the dinner table in his parents’ home, Thomas thought about Anna, in images and scents. The memories never distracted him; he could carry on a conversation, or keep reading a mortgage schedule, while she played in his head like wallpaper music.

His father cornered him at Easter. “You should think about getting married. It makes no difference to me, but there are social advantages you’re going to need sooner or later. And think how happy it would make your mother.”

Thomas said, “I’m twenty-four years old.”

“I was engaged when I was twenty-four.”

“Maybe I’m gay. Or perhaps I have an incurable venereal disease.”

“I don’t see why either should be an obstacle.”

Thomas saw Anna every second weekend. He bought her whatever she asked for. Sometimes she had the child with her. The boy was called Erik.

Thomas asked her, “Who’s the mother? Have I met her?”

She said, “You don’t want to.”

He worried about her sometimes—afraid she’d get herself arrested, or beaten up by junkies or rivals—but she seemed to be able to take care of herself. He could have hired private detectives to uncover the mysteries of her life, and bodyguards to watch over her, but he knew he had no right. He could have bought her an apartment, set her up with investments—but she never suggested anything of the kind, and he suspected she’d be deeply insulted if he made the offer. His gifts were lavish, but he knew she could have lived without them. They were using each other. She was, he told himself, as independent as he was.

He wouldn’t have said he loved her. He didn’t ache when they were apart; he just felt pleasantly numb, and looked forward to the next time he’d see her. He was jealous, but not obsessive, and she kept her other lovers out of the way; he rarely had to acknowledge their existence. He never saw Martin again.

Anna traveled with him to New York. They fell asleep in the middle of a Broadway show, saw the Pixies play at the Mudd Club, climbed the stairs to the top of Manhattan Chase.

Thomas turned twenty-five. His father promoted him. His mother said, “Look at all your gray hairs.”

In the spring, Erik disappeared. Anna said casually, “His mother’s gone, she’s moved away.”

Thomas was hurt; he’d liked having the boy around. He said, “You know, I used to think he might be yours.”

She was baffled. “Why? I told you he wasn’t. Why would I have lied?”

Thomas had trouble sleeping. He kept trying to picture the future. When his father died, would he still be seeing Anna, once a fortnight in Hamburg, while she dealt heroin and fucked pimps and junkies? The thought made him sick. Not because he didn’t want everything to stay the same, but because he knew that it couldn’t.

The Saturday in June was, almost, the second anniversary of the day they’d met. They went to a flea market in the afternoon, and he bought her cheap jewelery. She said, “Anything nicer would be asking for trouble.”

They ate junk food, went dancing. They ended up back at Anna’s flat at half past two. They danced around the tiny living room, propping each other up, more tired than drunk.

Thomas said, “God, you’re beautiful.” Marry me.

Anna said, “I’m going to ask you for something I’ve never asked for before. I’ve been trying to work up the courage all day.”

“You can ask for anything.” Marry me.

“I have a friend, with a lot of cash. Almost two hundred thousand marks. He needs someone who can—”

Thomas stepped back from her, then struck her hard across the face. He was horrified. He’d never hit her before; the thought had never even occurred to him. She started punching him in the chest and face; he stood there and let her do it for a while, then grabbed both her hands by the wrists.

She caught her breath. “Let go of me.”

“I’m sorry.”

“Then let go of me.”

He didn’t. He said, “I’m not a money-laundering facility for your friends.”

She looked at him pityingly. “Oh, what have I done? Offended your high moral principles? All I did was ask. You might have made yourself useful. Never mind. I should have known it was too much to expect.”

He pushed his face close to hers. “Where are you going to be, in ten years’ time? In prison? At the bottom of the Elbe?”

“Fuck off.”

“Where? Tell me?”

She said, “I can think of worse fates. I could end up playing happy families with a middle-aged banker.”

Thomas threw her toward the wall. Her feet slipped from under her before she hit it; her head struck the bricks as she was going down.

He crouched beside her, disbelieving. There was a wide gash in the back of her head. She was breathing. He patted her cheeks, then tried to open her eyes; they’d rolled up into her skull. She’d ended up almost sitting on the floor, legs sprawled in front of her, head lolling against the wall. Blood pooled around her.

He said, “Think fast. Think fast.”

He knelt over her, one knee to either side, took her face in his hands, then closed his eyes. He brought her head forward, then slammed it back against the wall. Five times. Then he held his fingers near her nostrils, without opening his eyes. He felt no exhalation.

He backed away from her, turned away and opened his eyes, then walked around the flat, wiping things he might have touched with his handkerchief. Avoiding looking at her. He was crying and shaking, but he couldn’t think why.

There was blood on his hands, his shirt, his trousers, his shoes. He found a garbage bag, put all his clothes in it, then washed the blood from his skin. There was a black spot in the center of his vision, but he worked around it. He put the garbage bag in his suitcase, and put on fresh clothes: blue jeans and a black T-shirt. He went through the flat, packing away everything that belonged to him. He almost took Anna’s address book, but when he checked he saw that he wasn’t in it. He looked for diaries, but found none.

Dozens of people had seen them together, month after month. Anna’s neighbors, Anna’s friends. Dozens of people had seen them leave the nightclub. He wasn’t sure how many of her friends knew what he did, where he was from. He’d never told any of them more than his first name, he’d always lied about the rest—but Anna might have told them everything she knew.

Having been seen with her alive was bad enough; he couldn’t risk being seen walking out the front door the night she was killed.

The flat was two flights up. The bathroom window opened onto an alley. Thomas threw the suitcase down; it landed with a soft thud. He thought of jumping—almost believing that he could land unhurt, or almost believing that he wouldn’t care—but there was a gray clarity underneath those delusions, and an engine in his skull a billion years old which only wanted to survive.

He climbed up into the window frame, into the gap left by the sliding half-pane, one foot either side of the track. There was no ledge, as such, just the double brickwork of the wall itself. He had to crouch to fit, but he found he could keep his balance by pushing his left hand up against the top of the frame, jamming himself in place.

He turned sideways, then reached across the outside wall, and into the frame of the bathroom window of the neighboring flat. He could hear traffic, and music somewhere, but no lights showed from within the flat, and the alley below was deserted. The two windows were scarcely a meter apart, but the second one was closed, halving its width. With one hand on each edge, he shifted his right foot to the neighbor’s window. Then, gripping the intervening wall tightly between his forearms, he moved his left foot across. Finally, securing himself by pressing up with his right hand, he let go of the first frame completely.

He shuffled across the one-brick’s-width ledge, fighting an impulse to mutter Ave Marias. Pray for us sinners? He realized that he’d stopped weeping. A drain pipe ran close to the far side of the window. He imagined tearing his palms open on jagged rusty metal, but the pipe was smooth; it took all his strength to hold himself in place, gripping it with hands and knees. When he touched the ground with his feet, his legs gave way. But not for long.

He hid in a public toilet for three hours, staring up at one corner of the room. The lights, the tiles, could have belonged to a prison or an asylum. He found himself disconnected, from the world, the past; his time breaking up into moments, shocks of awareness, shimmering droplets of mercury, beads of sweat.

This isn’t me. This is something else that believes it’s me. And it’s wrong, wrong, wrong.

Nobody disturbed him. At six o’clock he walked out into the morning light, and caught a train home.

15 (Remit not paucity) APRIL 2051

Durham’s north Sydney flat was small, and very sparsely furnished; not at all what Maria had expected. The combined living room and kitchen was all she’d seen, but it was clear from the outside that there wasn’t space for much more. Durham was on the sixteenth floor, but the building was hemmed in on all sides by ugly late-twenties office towers, blue and pink ersatz-marble monstrosities; no expensive harbor views here. For someone who was ripping off gullible millionaires—or even someone who merely sold them insurance—Durham didn’t seem to have much to show for it. Maria thought it unlikely that the place had been set up entirely for her benefit, to fit the story he’d told her: to demonstrate the frugal lifestyle which supposedly enabled him to pay her out of his own pocket. He’d invited her out of the blue; she would never have had a reason to insist on seeing where he lived.

She put her notepad down on the scratched dining table, and turned it so that Durham could read the graphs. “These are the latest results for the two most promising species. A. lithophila has the higher mutation rate, per generation, but it reproduces much more slowly, and it’s more vulnerable to climate change. A. hydrophila is more prolific, with a stabler genome. It’s not intrinsically hardier; it’s just better protected by the ocean.”

Durham said, “What’s your gut feeling?”

“What’s yours?”

A. litho evolves into a few promising species—which all get wiped out by one major crisis. A. hydro slowly builds up a huge stock of survival-neutral mutations, some of which turn out to be useful on land. The first few hundred thousand species which blow out of the sea don’t make it—but it doesn’t matter, there are always more. Or am I just being swayed too much by terrestrial preconceptions?”

“The people you’re trying to convince will almost certainly think the same way.”

Durham laughed. “It wouldn’t hurt to be right, as well as persuasive. If they’re not mutually exclusive ambitions.”

Maria didn’t reply. She stared down at the notepad; she couldn’t look Durham in the eye. Talking to him by phone, with software filters, had been bearable. And the work itself had been an end in itself; immersed in the elaborate game of Autoverse biochemistry, she’d found it all too easy to carry on, as if it made no difference what it was for. But she’d done next to nothing to make Durham more likely to take her into his confidence. That was why she’d agreed to this meeting—and why she had to take advantage of it.

The trouble was, now that she was here, she was so ill at ease that she could barely discuss the most neutral technicalities without her voice faltering. If he started spouting lies about his hopes of debating the skeptics of the artificial life mafia in some future issue of Cellular Automaton World, she’d probably start screaming. Or, more likely, throw up on the bare linoleum floor.

He said, “By the way, I signed the release on your fee this morning—I’ve authorized the trust fund to pay you in full. The work’s been going so well, it seemed only fair.”

Maria glanced up at him, startled. He looked perfectly sincere, but she couldn’t help wondering—not for the first time—if he knew that she’d been approached by Hayden, knew exactly what she’d been told. She felt her cheeks flush. She’d spent too many years using phones and filters; she couldn’t keep anything from showing on her face.

She said, “Thank you. But aren’t you afraid I might take the first plane to the Bahamas? There’s still a lot of work to be done.”

“I think I can trust you.”

There wasn’t a trace of irony in his voice—but there really didn’t need to be.

He said, “Speaking of trust… I think your phone may be bugged. I’m sorry; I should have told you that sooner.”

Maria stared at him. “How did you know?”

“Know? You mean, it is? You’ve had definite signs?”

“I’m not sure. But how… ?”

“Mine is. Bugged. So it makes sense that yours would be, too.”

Maria was bewildered. What was he going to do—announce that the Fraud Squad were watching him? If he came right out and said it, she didn’t think she could dissemble any longer. She’d have to confess that she already knew—and then she’d have to tell him everything Hayden had said.

Taking the pressure off completely. Ending the farce for good. She had no talent for these stupid games; the sooner they could both stop lying to each other, the better.

She said, “And who exactly do you think is doing it?”

Durham paused to think it over, as if he hadn’t seriously considered the question before. “Some corporate espionage unit? Some national security organization? There’s really no way of telling. I know very little about the intelligence community; your guess would be as good as mine.”

“Then why do you think they’re—?”

Durham said blithely, “If I was developing a computer, say, thirty orders of magnitude more powerful than any processor cluster in existence, don’t you think people like that might take an interest?”

Maria almost choked. “Ah. Yes.”

“But of course I’m not, and eventually they’ll convince themselves of that, and leave us both alone. So there’s absolutely nothing to worry about.”

“Right.”

Durham grinned at her. “Presumably, they think that just because I’ve commissioned an Autoverse planet, there’s a chance that I might possess the means to actually run it. They’ve searched this place a couple of times; I don’t know what they expected to find. A little black box, sitting in a comer of one of the rooms? Hidden under a pot plant, quietly cracking military codes, raking in a fortune on the stock market—and simulating a universe or two on the side, just to keep from getting bored. Any five-year-old could tell them how ludicrous that is. Maybe they think I’ve found a way to shrink individual processors to the size of an atom. That would just about do it.”

So much for an end to the lying. He wasn’t going to make this easy for her. All right. Maria forced the words out evenly: “And any five-year-old could tell you that if anyone searched your flat, it was the Fraud Squad.”

Durham was still giving nothing away. “Why do you say that?”

“Because I know they’re watching you. They’ve spoken to me. They’ve told me exactly what you’re doing.” Maria faced him squarely now. She was tense at the prospect of a confrontation, but she had nothing to be ashamed of; he was the one who’d set out to deceive her from the start.

He said, “Don’t you think the Fraud Squad would need to get a warrant, and search the flat in my presence?”

“Then maybe it hasn’t been searched at all. That’s not the point.”

He nodded slightly, as if conceding some minor breach of etiquette. “No, it’s not. You want to know why I lied to you.”

Maria said, “I know why. Please don’t treat me like an idiot.” Her bitterness surprised her, she’d had to conceal it for so long. “I was hardly going to agree to be your… accomplice—“

Durham raised one hand from the tabletop, a half conciliatory, half impatient gesture. Maria fell silent, more from astonishment at how calmly he seemed to be taking all this than any desire to give him a chance to defend himself.

He said, “I lied because I didn’t know if you’d believe the truth or not. I think you might have, but I couldn’t be sure. And I couldn’t risk it. I’m sorry.”

“Of course I would have believed the truth! It would have made a lot more sense than the bullshit you fed me! But, yes, I can see why you couldn’t risk it.”

Durham still showed no sign of contrition. “Do you know what it is that I’m offering my backers? The ones who’ve been funding your work?”

“A sanctuary. A privately owned computer somewhere.”

“That’s almost true. Depending on what you take those words to mean.”

Maria laughed cynically. “Oh, yes? Which words do you have trouble with? ‘Privately owned’?”

“No. ‘Computer.’ And, ‘somewhere.’”

“Now you’re just being childish.” She reached out and picked up her notepad, slid her chair back and rose to her feet. Trying to think of a parting shot, it struck her that the most frustrating thing was that the bastard had paid her. He’d lied to her, he’d made her an accomplice—but he hadn’t actually swindled her.

Durham looked up at her calmly. He said, “I’ve committed no crime. My backers know exactly what they’re paying for. The Fraud Squad, like the intelligence agencies, are jumping to absurd conclusions. I’ve told them the whole truth. They’ve chosen not to believe me.”

Maria stood by the table, one hand on the back of the chair. “They said you refused to discuss the matter.”

“Well, that’s a lie. Although what I had to say certainly wasn’t what they wanted to hear.”

“What did you have to say?”

Durham gave her a searching look. “If I try to explain, will you listen? Will you sit down and listen, to the end?”

“I might.”

“Because if you don’t want to hear the whole story, you might as well leave right now. Not every Copy took me up on the offer—but the only ones who went to the police were the ones who refused to hear me out.”

Maria said, exasperated, “What do you care what I think, now? You’ve extracted all the Autoverse technobabble from me you could possibly need. And I know nothing more about your scam than the police do; they’ll have no reason to ask me to testify against you, if all I can say in court is ‘Detective Hayden told me this, Detective Hayden told me that.’ So why don’t you quit while you’re ahead?”

Durham said simply, “Because you don’t understand anything. And I owe you an explanation.”

Maria looked toward the door, but she didn’t take her hand off the back of the chair. The work had been an end in itself—but she was still curious to know precisely what Durham had intended to do with the fruits of her labor.

She said, “How was I going to spend the afternoon, anyway? Modeling the survival of Autobacterium hydrophila in sea spray?” She sat. “Go ahead. I’m listening.”

Durham said, “Almost six years ago—loosely speaking—a man I know made a Copy of himself. When the Copy woke up, it panicked, and tried to bale out. But the original had sabotaged the software; baling out was impossible.”

“That’s illegal.”

“I know.”

“So who was this man?”

“His name was Paul Durham.”

“You? You were the original?”

“Oh, no. I was the Copy.

16 (Toy man, picture it) JUNE 2045

Paul felt a hand gripping his forearm. He tried to shake it off, but his arm barely moved, and a terrible aching started up in his shoulder. He opened his eyes, then closed them again in pain. He tried again. On the fifth or sixth attempt, he managed to see a face through washed-out brightness and tears.

Elizabeth.

She raised a cup to his lips. He took a sip, spluttered and choked, but then managed to force some of the thin sweet liquid down.

She said, “You’re going to be fine. Just take it easy.”

“Why are you here?” He coughed, shook his head, wished he hadn’t. He was touched, but confused. Why had his original lied—claiming that she wanted to shut him down—when in fact she was sympathetic enough to go through the arduous process of visiting him?

He was lying on something like a dentist’s couch, in an unfamiliar room. He was in a hospital gown; there was a drip in his right arm, and a catheter in his urethra. He glanced up to see an interface helmet, a bulky hemisphere of magnetic axon current inducers, suspended from a gantry, not far above his head. He thought: fair enough, to construct a simulated meeting place that looked like the room that her real body must be in. Putting him in the couch, though, and giving him all the symptoms of a waking visitor, seemed a little extreme.

He tapped the couch with his left hand. “What’s the message? You want me to know exactly what you’re going through? Okay. I’m grateful. And it’s good to see you.” He shuddered with relief, and delayed shock. “Fantastic, to tell the truth.” He laughed weakly. “I honestly thought he was going to wipe me out. The man’s a complete lunatic. Believe me, you’re talking to his better half.”

Elizabeth was perched on a stool beside him. She said, “Paul. Try to listen carefully to what I’m going to say. You’ll start to reintegrate the memories gradually, on your own, but it’ll help if I talk you through it all first. To start with, you’re not a Copy. You’re flesh and blood.”

Paul coughed, tasting acid. Durham had let her do something unspeakable to the model of his digestive system.

“I’m flesh and blood? What kind of sadistic joke is that? Do you have any idea how hard it’s been, coming to terms with the truth?”

She said patiently, “It’s not a joke. I know you don’t remember yet, but… after you made the scan that was going to run as Copy number five, you finally told me what you were doing. And I persuaded you not to run it—until you’d tried another experiment: putting yourself in its place. Finding out, firsthand, what it would be forced to go through.

“And you agreed. You entered the virtual environment which the Copy would have inhabited—with your memories since the day of the scan suppressed, so you had no way of knowing that you were only a visitor.”

“I—?”

“You’re not the Copy. Do you understand? All you’ve been doing is visiting the environment you’d prepared for Copy number five. And now you’re out of it. You’re back in the real world.”

Her face betrayed no hint of deception—but software could smooth that out. He said, “I don’t believe you. How can I be the original? I spoke to the original. What am I supposed to believe? He was the Copy? Thinking he was the original?”

“Of course not. That would hardly have spared the Copy, would it? The fifth scan was never run. I controlled the puppet that played your ‘original’—software provided the vocabulary signature and body language, but I pulled the strings. You briefed me, beforehand, on what to have it say and do. You’ll remember that, soon enough.”

“But… the experiments?”

“The experiments were a sham. They could hardly have been performed on a visitor, on a physical brain—could they?”

Paul shook his head, and whispered, “Abulafia.”

No interface window appeared.

He gripped the couch and closed his eyes, then laughed. “You say I agreed to this? What kind of masochist would do that? I’m going out of my mind. I don’t know what I am.”

Elizabeth took hold of his arm again. “You’re disoriented—but that won’t last long. And you know why you agreed. You were sick of Copies baling out on you. You had to come to terms with their experience. Spending a few days believing you were a Copy would make or break the project: you’d either end up psychologically prepared, at last, to give rise to a Copy who’d be able to cope with its fate—or you’d gain enough sympathy for their plight to stop creating them.

“The plan was to tell you everything while you were still inside, after the third experiment. But when you went weird on me in there, I panicked. All I could think of was having the puppet playing your original tell you that it was going to pause you. I wasn’t trying to frighten you. I didn’t think you’d take it so badly.”

A technician came into the room and removed the drip and catheter. Paul propped himself up and looked out through the windows of the room’s swing doors; he could see half a dozen people in the corridor. He bellowed wordlessly at the top of his lungs; they all turned to stare in his direction. The technician said mildly, “Your penis might sting for an hour or two.”

Paul slumped back onto the couch and turned to Elizabeth. “You wouldn’t pay for reactive crowds. I wouldn’t pay for reactive crowds. It looks like you’re telling the truth.”

+ + +

People, glorious people: thousands of strangers, meeting his eyes with suspicion or puzzlement, stepping out of his way on the street—or, more often, clearly, consciously refusing to. The freedom of the city was so sweet. He walked the streets of Sydney for a full day, rediscovering every ugly shopping arcade, every piss-stinking litter-strewn park and alley, until, with aching feet, he squeezed his way home through the evening rush hour, to watch the real-time news.

There was no room for doubt: he was not in a virtual environment. Nobody in the world could have had reason to spend so much money, simply to deceive him.

When Elizabeth asked if his memories were back, he nodded and said of course. She didn’t grill him on the details. In fact, having gone over her story so many times in his head, he could almost imagine the stages: his qualms after the fifth scan; repeatedly putting off running the model; confessing to Elizabeth about the project; accepting her challenge to experience for himself just what his Copies were suffering.

And if the suppressed memories hadn’t actually reintegrated, well, he’d checked the literature, and there was a two point five percent risk of that happening; electronically censoring access to memories could sometimes permanently weaken the neural connections in which they were encoded.

He even had an account from the database service which showed that he’d consulted the very same articles before.

He reread and replayed the news reports that he’d accessed from inside—and found no discrepancies. He flicked through encyclopedic databases—spot-checking random facts of history, geography, astronomy—and although he was surprised now and then by details which he’d never come across before, there were no startling contradictions. The continents hadn’t moved. Stars and planets hadn’t vanished. The same wars had been lost and won.

Everything was consistent. Everything was explicable.

And yet he couldn’t stop wondering about the fate of a Copy who was shut down and never run again. A normal human death was one thing—woven into a much vaster tapestry, it was a process which made perfect sense. From the internal point of view of a Copy whose model was simply halted, though, there was no explanation whatsoever for its demise—just an edge where the pattern abruptly came to an end.

But if the insight he’d gained from the experiments was true (whether or not they’d ever really happened)—if a Copy could assemble itself from dust scattered across the world, and bridge the gaps in its existence with dust from across the universe… then why should it ever come to an inconsistent end? Why shouldn’t the pattern keep on finding itself?

Or find a larger pattern into which it could merge?

The dust theory implied a countless number of alternative worlds: billions of different possible histories spelled out from the same primordial alphabet soup. One history in which Durham did run Copy number five—and one in which he didn’t, but was persuaded to take its place as a visitor, instead.

But if the visitor had been perfectly deceived, and had experienced everything the Copy did… what set the two of them apart? So long as the flesh-and-blood man had no way of knowing the truth, it was meaningless to talk about “two different people” in “two different worlds.” The two patterns of thoughts and perceptions had effectively merged into one.

If the Copy had been allowed to keep on running after the visitor had learned that he was flesh and blood, their two paths would have diverged again. But the Copy had been shut down; it had no future at all in its original world, no separate life to live.

So the two subjective histories remained as one. Paul had been a visitor believing he was a Copy. And he’d also been the Copy itself. The patterns had merged seamlessly; there could be no way of saying that one history was true and the other false. Both explanations were equally valid.

Once, preparing to be scanned, he’d had two futures.

Now he had two pasts.

+ + +

Paul woke in darkness, confused for a moment, then pulled his cramped left arm out from under the pillow and glanced at his watch. Low power infrared sensors in the watch face detected his gaze, and flashed up the time—followed by a reminder: due at landau 7 a.m. It was barely after five, but it hardly seemed worth going back to sleep.

Memories of the night before came back to him. Elizabeth had finally confronted him, asking what decision he’d reached: to abandon his life’s work, or to forge ahead, now that he knew, firsthand, what was involved.

His answer seemed to have disappointed her. He didn’t expect to see her again.

How could he give up? He knew he could never be sure that he’d discovered the truth—but that didn’t mean that nobody else could.

If he made a Copy, ran it for a few virtual days, then terminated it abruptly… then at least that Copy would know if its own pattern of experience continued.

And if another Paul Durham in one of the countless billions of alternative worlds could provide a future for the terminated Copy—a pattern into which it could merge—then perhaps that flesh-and-blood Durham would repeat the whole process

And so on, again and again.

And although the seams would always be perfect, the “explanation” for the flesh-and-blood human believing that he had a second past as a Copy would necessarily grow ever more “contrived,” less convincing… and the dust theory would become ever more compelling.

Paul lay in bed in the darkness, waiting for sunrise, staring into the future down this corridor of mirrors.

One thing nagged at him. He could have sworn he’d had a dream, just before he woke: an elaborate fable, conveying some kind of insight. That’s all he knew—or thought he knew. The details hovered maddeningly on the verge of recollection.

His dreams were evanescent, though, and he didn’t expect to remember anything more.

17 (Remit not paucity) APRIL 2051

Maria shifted in her seat to try to get her circulation flowing, then realized it wasn’t enough. She stood up and limped around the room, bending down to massage her cramped right calf.

She said, “And you claim you’re the twenty-third?” She was almost afraid to sound too skeptical; not because she believed that Durham would take offence, but because the story was so strangely entrancing that she wasn’t sure she wanted to deflate it, yet. One hint of mockery and the floodgates would open. “You’re the twenty-third flesh-and-blood Paul Durham whose past includes all those who came before?”

Durham said, “I may be wrong about the exact number. I may have counted this last version more than once; if I’m capable of believing in twenty-three incarnations, some of them might be false. The whole nature of the delusions I suffered contributes to the uncertainty.”

“Contributes. Isn’t that a bit of an understatement?”

Durham was unflappable. “I’m cured now. The nanosurgery worked. The doctors pronounced me sane, and I have no reason to question their judgement. They’ve scanned my brain; it’s functioning impeccably. I’ve seen the data, before and after. Activity in the prefrontal cortex—”

“But don’t you see how absurd that is? You acknowledge that you were deluded. You insist that you’re cured now. But you claim that your delusions weren’t delusions—”

Durham said patiently, “I’ve admitted from the outset: my condition explains everything. I believed—because I was mentally ill—that I was the twenty-third-generation Copy of another Paul Durham, from another world.”

Because you were mentally ill! End of story.”

“No. Because I’m certifiably rational now—and the logic of the dust theory makes as much sense to me as ever. And it makes no difference whether my memories are true, false, or both.”

Maria groaned. “Logic of the dust theory! It’s not a theory. It can’t be tested.”

“Can’t be tested by whom?”

“By anyone! I mean… even assuming that everything you believe is the truth: you’ve ‘been through’ twenty-three separate experiments, and you still don’t know what you’ve proved or disproved! As you say: your condition accounts for everything. Haven’t you heard of Occam’s razor: once you have a perfectly simple explanation for something, you don’t go looking for ever more complicated ways of explaining the very same thing? No dust theory is required.” Her words reverberated in the near-empty room. She said, “I need some fresh air.”

Durham said firmly, “After twenty-three ambiguous results, I know how to get it right this time. A Copy plus a virtual environment is a patchwork, a mess. A system like that isn’t rich enough, detailed enough, or consistent enough, to be self-sustaining. If it was, when I was shut down, the entire VR world I was in would have persisted. That never happened. Instead—every time—I found a flesh-and-blood human with a reason to believe he shared my past. That explained my pattern of experience far better than VR—even to the point of insanity.

“What I have to do now is construct a consistent pattern which can only have one past.”

Maria took a few deep breaths. It was almost too much to bear: Durham’s sad flat, his cosmic visions, his relentless, mechanical logic, grinding away trying to make sense of the legacy of his disease. The doctors had cured him, he was sane. He just didn’t want to disown his delusional past—so he’d invented a flawlessly logical, utterly irrefutable, reason to hang on to it.

If he’d really told the cops all this, why were they still hounding him? They should have seen that he was harmless and left him alone—and left his moronic clients to fend for themselves. The man wasn’t even a danger to himself. And if he could ever harness a fraction of the energy and intelligence he’d put into this “project” and direct it towards something worthwhile—

Durham said, “Do you know what a Garden-of-Eden configuration is?”

Maria was caught blank for a second, then she said, “Yes, of course. In cellular automaton theory, it’s a state of the system that can’t be the result of any previous state. No other pattern of cells can give rise to it. If you want a Garden-of-Eden configuration, you have to start with it—you have to put it in by hand as the system’s first state.”

Durham grinned at her as if she’d just conceded the whole argument. She said, “What?”

“Isn’t it obvious? A cellular automaton isn’t like patchwork VR; it’s every bit as consistent as a physical universe. There’s no jumble of ad hoc high-level laws; one set of rules applies to every cell. Right?”

“Yes, but—”

“So if I set up a cellular automaton in a Garden-of-Eden configuration, run it through a few trillion clock ticks, then shut it down… the pattern will continue to find itself in the dust—separate from this version of me, separate from this world, but still flowing unambiguously from that initial state. A state which can’t be explained by the rules of the automaton. A state which must have been constructed in another world—exactly as I remember it.

“The whole problem, so far, has been that my memories are always entirely explicable within the new world. I shut myself down as a Copy—and find myself in a flesh-and-blood body with flesh-and-blood memories which the laws of physics could have produced from earlier states of a flesh-and-blood brain. This world can explain me only as a man whose delusions are unlikely beyond belief—but there’s no denying that I do have a complete extra history, here, that’s not literally, physically impossible. So whatever I prefer to believe, I have to concede that the outcome of the experiment is still ambiguous. I could, still, be wrong.

“But a cellular automaton can’t provide an ‘extra history’ for a Garden-of-Eden configuration! It’s mathematically impossible! If I find myself inside a cellular automaton universe, and I can track my past back to a Garden-of-Eden configuration, that will be conclusive proof that I did seed the whole universe in a previous incarnation. The dust theory will be vindicated. And I’ll finally know—beyond any doubt—that I haven’t merely been insane all along.”

Maria felt punch-drunk. At one level, she knew she should stop humoring him, stop treating his ideas seriously. On another, it seemed that if Durham was so wrong, she should be able to point out the reasons why. She shouldn’t have to call him a madman and refuse to listen to another word.

She said, “Find yourself in a cellular automaton world? You don’t mean the Autoverse—?”

“Of course not. There’s no prospect of translating a human into Autoverse biochemistry.”

“Then what?”

“There’s a cellular automaton called TVC. After Turing, von Neumann and Chiang. Chiang completed it around twenty-ten; it’s a souped-up, more elegant version of von Neumann’s work from the nineteen fifties.”

Maria nodded uncertainly; she’d heard of all this, but it wasn’t her field. She did know that John von Neumann and his students had developed a two-dimensional cellular automaton, a simple universe in which you could embed an elaborate pattern of cells—a rather Lego-like “machine”—which acted as both a universal constructor and a universal computer. Given the right program—a string of cells to be interpreted as coded instructions rather than part of the machine—it could carry out any computation, and build anything at all. Including another copy of itself—which could build another copy, and so on. Little self-replicating toy computers could blossom into existence without end.

She said, “Chiang’s version was three-dimensional, wasn’t it?”

“Much better. N-dimensional. Four, five, six, whatever you like. That leaves plenty of room for data within easy reach. In two dimensions, the original von Neumann machine had to reach farther and farther—and wait longer and longer—for each successive bit of data. In a six-dimensional TVC automaton, you can have a three-dimensional grid of computers, which keeps on growing indefinitely—each with its own three-dimensional memory, which can also grow without bound.”

Maria said numbly, “Where are you supposed to fit into all of this? If you think translating human biochemistry into Autoverse terms is difficult, how are you going to map yourself into a six-dimensional world designed solely to support von Neumann machines?”

“The TVC universe is one big, ever-expanding processor cluster. It runs a Copy of me—”

“I thought the whole point was to do away with Copies!”

“-- in a VR environment which lets me interact with the TVC level. Yes, I’ll be a patchwork Copy, as always—there’s no alternative to that—but I’ll also be linked to the cellular automaton itself. I’ll witness its operation, I’ll experience its laws. By observing it, I’ll make it a part of what has to be explained.

“And when the simulated TVC universe being run on the physical computer is suddenly shut down, the best explanation for what I’ve witnessed will be a continuation of that universe—an extension made out of dust.”

Maria could almost see it: a vast lattice of computers, a seed of order in a sea of a random noise, extending itself from moment to moment by sheer force of internal logic, “accreting” the necessary building blocks from the chaos of non-space-time by the very act of defining space and time.

Visualizing wasn’t believing, though.

She said, “What makes you so sure? Why not another deluded psychiatric patient, who believes he was—briefly—a Copy being run on a TVC automaton being run on a processor cluster in another world?”

“You’re the one who invoked Occam’s razor. Wouldn’t you say that a self-contained TVC universe is a simpler explanation, by far?”

“No. It’s about the most bizarre thing I can imagine.”

“It’s a lot less bizarre than yet another version of this universe, containing yet another version of me, with yet another set of convenient delusions.”

“How many of your clients believed all this? How many think they’re coming along for the ride?”

“Fifteen. And there’s a sixteenth who, I think, is tempted.”

“They paid—?”

“About two million each.” He snorted. “It’s quite funny, the significance the police have attached to that. Some large sums of money have changed hands, for reasons more complex than usual—so they assume I must be doing something illegal. I mean, billionaires have been known to make donations larger than that to the Church of the God Who Makes No Difference.” He added hastily, “None of mine.”

Maria was having some trouble with the scale of things herself. “You found fifteen Copies willing to part with two million dollars after hearing this bullshit? Anyone that gullible deserves to lose their money.”

Durham took no offence. “If you were a Copy, you’d believe the dust theory, too. You’d feel the truth of it in your nonexistent bones. Some of these people carried out the same experiments as I did—computing themselves in randomized fragments—but others didn’t need to. They already knew that they could scatter themselves across real time and real space, and they’d still find themselves. Every Copy proves the dust theory to itself a million times a day.”

It suddenly occurred to Maria that Durham might have invented all of this for her sake, alone—while telling his clients exactly what Hayden had assumed: some fraudulent but utterly non-metaphysical tale of a hidden supercomputer. But she couldn’t see what he had to gain by confusing her… and too many details made too much sense, now. If his clients had accepted the whole mad vision, the problem of making them believe in a nonexistent supercomputer vanished. Or at least changed from a question of evidence to a question of faith. She said, “So you promised to fit a snapshot of each of your ‘backers’ into the Garden-of-Eden configuration, plus the software to run them on the TVC?”

Durham said proudly, “All that and more. The major world libraries; not quite the full holdings, but tens of millions of files—text, audio, visual, interactive—on every conceivable subject. Databases too numerous to list—including all the mapped genomes. Software: expert systems, knowledge miners, metaprogrammers. Thousands of off-the-shelf VR environments: deserts, jungles, coral reefs, Mars and the moon. And I’ve commissioned Malcolm Carter, no less, to create a major city to act as a central meeting place: Permutation City, capital of the TVC universe.

“And, of course, there’ll be your contribution: the seed for an alien world. Humanity is going to find other life in this universe, eventually. How can we give up hope of doing the same? Sure, we’ll have our own software descendants, and recreated Earth animals, and no doubt novel, wholly artificial creatures as well. We won’t be alone. But we still need a chance to confront the Other. We mustn’t leave that possibility behind. And what could be more alien than Autoverse life?”

Maria’s skin crawled. Durham’s logic was impeccable; an endlessly expanding TVC universe, with new computing power being manufactured out of nothing in all directions, “would” eventually be big enough to run an Autoverse planet—or even a whole planetary system. The packed version of Planet Lambert—the compressed description, with its topographic summaries in place of actual mountains and rivers—would easily fit into the memory of a real-world computer. Then Durham’s Copy could simply wait for the TVC grid to be big enough—or pause himself, to avoid waiting—and have the whole thing unfold.

Durham said, “I’ve been working on the software which will run the first moments of the TVC universe on a real-world computer. I can probably finish that myself. But I can’t complete the Autoverse work without you, Maria.”

She laughed sharply. “You want me to keep working for you? You lie to me. You get me visited by the Fraud Squad. You confess to a history of mental illness. You tell me you’re the twenty-third incarnation of a retailing millionaire from a parallel world—”

“Whatever you think about the dust theory—and whatever you think about my psychological health—I can prove to you that I’m not a criminal. My backers will vouch for that; they all know exactly what their money’s being used for. None of them are victims of fraud.”

“I accept that. I just—”

“Then accept the payment. Finish the work. Whatever the police have told you, you have every right to the money, and I have every right to give it to you. Nobody’s going to take you to court, nobody’s going to throw you into prison.”

Maria was flustered. “Just, hold on. Will you give me a chance to think?” Durham’s sheer reasonableness was beginning to be as exhausting as the impassioned rhetoric of any obvious fanatic. And so much ground had shifted in the last half-hour that she hadn’t had a chance to even start to reappraise her own situation: legally, financially… and morally.

She said, “Why don’t your backers tell the police all this? If they can confirm your story for me, why can’t they do the same for the cops? By refusing to talk, they’re just fueling suspicion.”

Durham agreed. “Tell me about it. It makes everything ten times harder—but I’m just going to have to keep on living with that. Do you think they’d risk the truth becoming public knowledge? There have already been some embarrassing leaks—but so far we’ve been able to muddy the water by putting out our own misinformation. Copies with de facto control of billion-dollar business empires would much rather have people linking them to some dubious salesman and his breakthrough supercomputer—and have the rumors fizzle out from lack of substantiation—than let the world know that they plan to send a clone into an artificial universe which runs without hardware. The share markets can get nervous enough when people start wondering if a certain board of directors have all taken up playing virtual Caligula in their spare time. If word got out that a Copy in a position of power had done something which might be construed as a sign that they no longer felt obliged to give a shit about their corporate responsibilities, their personal wealth, or the continued existence of Planet Earth… “

Maria walked over to the window. It was open, but the air outside was still; standing by the insect screen she might as well have been standing by a solid brick wall. People were arguing loudly in the flat above; she’d only just noticed.

When Durham had first approached her, she’d wondered, half seriously, if she’d be taking advantage of a man who’d taken leave of his senses. Now, she couldn’t just shrug that off as a hypocritical insult to a fellow eccentric. This wasn’t a matter of an artificial life fanatic with more money than sense. An ex-psychiatric patient was planning to spend thirty million dollars of other people’s money to “prove” his own sanity—and lead the clones of his followers into a cybernetic paradise which would last for about twenty seconds. Taking a cut seemed just a tiny bit like doing the catering for the Jonestown massacre.

Durham said, “If you don’t agree to finish the biosphere seed, who would I get to replace you? There’s nobody else who could even begin to grasp what’s involved.”

Maria eyed him sharply. “Don’t start flattering me. And don’t kid yourself about the seed, either. You asked for a package of persuasive data, and that’s all you’ll be getting—even if I finish the work. If you’re counting on Planet Lambert’s inhabitants rising up on their hind legs and talking to you… I can’t guarantee that happening if you ran the whole thing a billion times. You should have simulated real-world biochemistry. At least it’s been shown that intelligent life can arise within that system… and you’d supposedly have the computing power to do it.”

Durham said reasonably, “A. lamberti seemed simpler, surer. Any real-world organism—modeled subatomically—would be too big a program to test out in advance on any physical computer. And it’d be too late to change my mind and try another approach if I failed to get it to work—stuck in the TVC universe, with plenty of books and journals, but no pool of expertise.”

Maria felt a deep chill pass through her; every time she thought she’d accepted just how seriously Durham took this lunacy, he gave an answer like that which drove it home to her anew.

She said, “Well, Autoverse life might turn out just as useless. You might have A. hydrophila spewing out useless mutations, generation after generation, with nothing you can do to fix it.”

Durham seemed about to reply, but then stopped himself. Maria felt the chill return, at first without knowing why. A second later, she glared at him, outraged, as furious as if he’d come right out and asked her.

“I will not be there to fix it for you!”

Durham had the grace to look cowed, momentarily—but instead of denying that the thought had ever crossed his mind, he said, “If you don’t believe in the dust theory, what difference would it make if there’s a scan file of you in the Garden-of-Eden data?”

“I don’t want a Copy of me waking up and living for a few subjective seconds, knowing that it’s going to die!”

“Who said anything about waking it? Running a Copy on a simulated TVC grid is a computer-intensive operation. We can’t afford to wake more than one Copy while we’re still running on a physical computer. Mine. As far as you’re concerned, your scan file would never even be used to build a Copy; the data would just sit there, completely inert. And you could sit outside at a terminal, overseeing the whole operation, making sure I kept my word.”

Maria was scandalized—although it took her a second to weave through Durham’s infuriating logic to find a target.

“And you—certain that I’d eventually wake—would happily take me on board under false pretences?”

Durham seemed genuinely baffled by the accusation. “False pretences? I’ve given you all the facts, and I’ve argued my case as hard as I can; it’s not my fault if you don’t believe me. Am I supposed to feel guilty for being right?”

Maria started to reply, but then the point seemed too ridiculous to pursue. She said, “Never mind. You won’t get a chance to feel anything about it, because I’m certainly not offering you a scan file.”

Durham bowed his head. “It’s your decision.”

Maria hugged herself. She was actually trembling slightly. She thought: I’m afraid of exploiting him? If what he’s doing really is legitimate… finish the job, take the money. His Copy’s going to spend a few seconds believing it’s headed for Copy Heaven—and that’s going to happen whatever I do. The fifteen clones will just sleep through it all, as if they’d never been made. That’s no Jonestown.

Durham said, “The fee would be six hundred thousand dollars.”

Maria said, “I don’t care if it’s six hundred million.” She’d meant to shout, but her words faded out into a whisper.

Six hundred thousand dollars would be enough to save Francesca’s life.

18 (Remit not paucity) MAY 2051

Peer seemed to be making love with Kate, but he had his doubts. He lay on the soft dry grass of a boundless meadow, in mild sunshine. Kate’s hair was longer than usual, tickling his skin wherever she kissed him, brushing against him with an erotic precision which seemed unlikely to have been left to chance. Insect chirps and birdsong were heard. Peer could recall David Hawthorne screwing a long-suffering lover in a field, once. They’d been driving back to London from her father’s funeral in Yorkshire; it had seemed like a good idea at the time. This was different. No twigs, no stones, no animal shit. No damp earth, no grass stains, no itching.

The perfect meadow itself was no reason for suspicion; neither of them were verisimilitude freaks, masochistic re-creators of the irritating details of real environments. Good sex was, equally, a matter of choice. But Peer still found himself wondering if Kate really had agreed to the act. She hadn’t actually made love to him for months—however many times he’d recycled the memories of the last occasion—and he couldn’t rule out the possibility that he’d merely decided to fool himself into believing that she’d finally relented. He’d never gone quite so far before—so far as he presently knew—but he had a vague memory of resolving to do a thorough job of concealing the evidence if he ever did.

He could clearly remember Kate beginning to flirt as they’d toured Carter’s city, and then reaching out and starting to undress him as they stood in the exit doorway. He’d shut down all limits on her access to his body while she’d been unbuttoning his shirt—and he’d bellowed with shock and delight when, in the middle of their physically plausible foreplay, an invisible second Kate, twenty times his size, had picked him up in one hand, raised him to her mouth, and licked his body from toes to forehead like a sweet-toothed giant taking the icing off a man-shaped cake.

None of this struck him as especially unlikely; if Kate had decided to make love again, it was the kind of thing he could imagine her doing. That in itself proved nothing. He could have scripted this fantasy to fit everything he knew about her—or chosen the scenario, and then rewritten his “knowledge” of her to accommodate the action. In either case, software could have laid down a trail of false memories: a plausible transition from their meeting with Carter—which he felt certain had actually happened—to this moment. All memories of having planned the deception would have been temporarily suppressed.

Kate stopped moving. She shook her head, spattering his face and chest with sweat, and said, “Are you here where you seem to be, or off somewhere else?”

“I was about to ask you the same question.”

She smiled wickedly. “Ah. Then maybe this body you hope is me only asked you first to put your mind at ease.”

In the sky above her right shoulder, Peer could see a stray cloud taking on a new shape, a whimsical sculpture parodying the bodies on the grass below.

He said, “And then admitted as much?”

Kate nodded, and started slowly rising. “Of course. For the very same reason. How many levels of bluffing will it take before you get bored and say: Fuck it, I don’t care?”

She lifted herself until they were almost apart. He closed his eyes and violated the geometry, licking the sweat from between her shoulder blades without moving a muscle. She responded by sticking her tongue in both of his ears simultaneously. He laughed and opened his eyes.

The cloud above had darkened. Kate lowered herself onto him again, trembling very slightly.

She said, “Don’t you find it ironic?”

“What?”

“Trans-humans taking pleasure by stimulating copies of the neural pathways which used to be responsible for the continuation of the species. Out of all the possibilities, we cling to that.”

Peer said, “No, I don’t find it ironic. I had my irony glands removed. It was either that, or castration.”

She smiled down at him. “I love you, you know. But would I tell you that? Or would you be stupid enough to pretend that I had?”

Warm, sweet rain began to fall.

He said, “I don’t care, I don’t care, I don’t care.”

+ + +

Peer sat on the lowest of the four wooden steps leading up to the back porch of his homestead, glancing down now and then at his bare feet and thin brown arms. Ten-year-old farm boy at dusk. Kate had made both the environment and the body for him, and be liked the tranquil mood of the piece. There was no invented family, no role to play; this was a painting, not a drama. One place, one moment, lasting as long as he chose to inhabit it. The scenery wasn’t quite photorealist—there were subtle distortions of form, color and texture which made it impossible to forget that he was inhabiting a work of art—but there were no sledgehammer techniques: no visible brushstrokes, no Van Gogh lighting effects.

Violating the whole aesthetic, an interface window hovered in front of him, a meter above the chicken-feed-scattered dirt. The cloning utility insisted on following an elaborate confirmation sequence; Peer kept saying, “Please skip to the final question, I know exactly what I’m doing”—but icons in legal wigs and gowns kept popping up in front of the window and declaring solemnly, “You must read this warning carefully. Your brain model will be directly examined for evidence of complete understanding before we proceed to the next stage.”

It was a thousand times more trouble than baling out—he knew that for certain, having almost done it—but then, baling out entailed fewer legal complications for the people outside. Peer’s estate was controlled by an executor, who’d signed a contract obliging her to act according to “any duly authenticated communications—including, but not limited to, visual and/or auditory simulations of a human being appearing to proffer instructions or advice.” What duly authenticated meant revolved around a ninety-nine-digit code key which had been “hardwired” into Peer’s model-of-a-brain when his Copy was generated from his scan file. He could summon it up consciously if he had to, in some unlikely emergency, but normally he made use of it by a simple act of will. He’d record a video postcard, wish it to be duly authenticated—and it was done. Unless the key was stolen—plucked right out of the computer memory which contained the data representing his brain—Peer was the only software on the planet capable of encrypting instructions to his executor in a form compatible with her own matching key. It was the closest thing he had to a legal identity.

By law, any clone which a Copy made of itself had to be given a new key. It was up to the initial Copy, prior to the cloning, to divide up the worldly assets between the two future selves—or rather, divide them up between the executor’s two portfolios.

Peer fought his way through the process of assuring the cloning utility that he really had meant what he’d told it from the start: The clone would require no assets of its own. Peer would run it on sufferance, paying for its running time himself. He didn’t plan on keeping it conscious for more than a minute or two; just long enough to reassure himself that he was doing the right thing.

He almost wished that Kate was with him, now. She’d offered to be here, but he’d turned her down. He would have been glad of her support, but this had to be done in private.

Finally, the utility said, “This is your last chance to cancel. Are you sure you wish to proceed?”

Peer closed his eyes. When I see my original, sitting on the porch, I’ll know who I am, and accept it.

He said, “Yes, I’m sure.”

Peer felt no change. He opened his eyes. His newly made twin stood on the ground where the interface window had been, staring at him, wide-eyed. Peer shivered. He recognized the boy as himself, and not just intellectually—Kate’s piece included adjustments to every part of his brain which dealt with his body image, so he’d be no more shocked by catching a glimpse of himself in a mirror than he was by the way his limbs felt as he walked. But the effect wasn’t so much to see through the “disguise” of the ten-year-old body, as to find himself thinking of the clone—and himself—as if the two of them really were that young. How could he send this child into exile?

Peer brushed the absurd notion aside. “Well?”

The clone seemed dazed. “I—”

Peer prompted him. “You know what I want to hear. Are you ready for this? Are you happy with your fate? Did I make the right decision? You’re the one who knows, now.”

“But I don’t know.” He looked at Peer pleadingly, as if hoping for guidance. “Why am I doing this? Remind me.”

Peer was taken aback, but some disorientation was only to be expected. His own voice sounded “normal” to him—thanks to the neural adjustments—but the clone still sounded like a frightened child. He said gently, “Kate. We want to be with her. Both of her—”

The clone nodded fervently. “Of course.” He laughed nervously. “And of course I’m ready. Everything’s fine.” His eyes darted around the yard, as if he was searching for an escape route.

Peer felt his chest tighten. He said evenly, “You don’t have to go ahead if you don’t want to. You know that. You can bale out right now, if that’s what you’d prefer.”

The clone looked more alarmed than ever. “I don’t want that! I want to stow away with Kate.” He hesitated, then added, “She’ll be happier in there, more secure. And I do want to be with her; I want to know that side of her.”

“Then what’s wrong?”

The clone sank to his knees in the dirt For a second, Peer thought he was sobbing, then he realized that the noise was laughter.

The clone recovered his composure and said, “Nothing’s wrong—but how do you expect me to take it? The two of us, cut off from everything else. Not just the real world, but all the other Copies.”

Peer said, “If you get lonely, you can always generate new people. You’ll have access to ontogenesis software—and no reason to care about the slowdown.”

The clone started laughing again. Tears streamed down his face. Hugging himself, he tumbled sideways onto the ground. Peer looked on, bemused. The clone said, “Here I am trying to steel myself for the wedding, and already you’re threatening me with children.”

Suddenly, he reached out and grabbed Peer by one ankle, then dragged him off the step. Peer hit the ground on his arse with a jarring thud. His first instinct was to freeze the clone’s power to interact with him, but he stopped himself. He was in no danger—and if his twin wanted to burn off some aggression on his brother-creator, he could take it. They were evenly matched, after all.

Two minutes later, Peer was lying with his face in the dirt and his arms pinned behind his back. The clone kneeled over him, breathless but triumphant.

Peer said, “All right, you win. Now get off me—or I’ll double my height, put on forty kilograms, and get up and flatten you.”

The clone said, “Do you know what we should do?”

“Shake hands and say goodbye.”

“Toss a coin.”

“For what?”

The clone laughed. “What do you think?”

“You said you were happy to go.”

“I am. But so should you be. I say we toss a coin. If I win, we swap key numbers.”

“That’s illegal!”

Illegal!” The clone was contemptuous. “Listen to the Solipsist Nation Copy invoke the laws of the world! It’s easily done. The software exists. All you have to do is agree.”

Talking was difficult; Peer spat out sand, but there was a seed of some kind caught between his teeth which he couldn’t dislodge. He felt a curious reluctance to “cheat,” though—to remove the seed from his mouth, or the clone from his back. It had been so long since he’d been forced to endure the slightest discomfort that the novelty seemed to outweigh the inconvenience.

He said, “All right. I’ll do it”

And if he lost? But why should he fear that? Five minutes ago, he’d been prepared to give rise to—to become—the clone who’d stow away.

They created the coin together, the only way to ensure that it was subject to no hidden influences. The reality editor they jointly invoked offered a standard object ready-made for their purpose, which they decorated as a one-pound coin. The physics of flipping a real coin wouldn’t come into it; any Copy could easily calculate and execute a flick of the thumb leading to a predetermined outcome. The result would be controlled by a random number generator deep in the hidden layers of the operating system.

Peer said, “I toss, you call”—at exactly the same time as the clone. He laughed. The clone smiled faintly. Peer was about to defer, then decided to wait. A few seconds later, he said, alone, “All right, you toss.”

As the coin went up, Peer thought about encasing it in a second object, an invisibly thin shell under his control alone—but the long list of attributes of the fair coin probably included crying foul if its true faces were concealed. He shouted “Heads!” just before the thing hit the dirt.

The two of them fell to their hands and knees, almost bumping heads. A hen approached; Peer shooed it away with a backward kick.

President Kinnock, in profile, glinted in the dust.

The clone met his eyes. Peer did his best not to look relieved—short of severing ties with his body. He tried to read the clone’s expression, and failed; all he saw was a reflection of his own growing numbness. Pirandello had said it was impossible to feel any real emotion while staring into a mirror. Peer decided to take that as a good sign. They were still one person, after all—and that was the whole point.

The clone rose to his feet dusting off his knees and elbows. Peer took a hologram-embossed library card from the back pocket of his jeans and handed it over; it was an icon for a copy of all the environments, customized utilities, bodies, memories and other data he’d accumulated since his resurrection.

The clone said, “Don’t worry about me—or Kate. We’ll look after each other. We’ll be happy.” As he spoke, he morphed smoothly into an older body.

Peer said, “Ditto.” He reached up and shook the young man’s hand. Then he summoned one of his control windows and froze the clone, leaving the motionless body visible as an icon for the snapshot file. He shrunk it to a height of a few centimeters, flattened it into a two-dimensional postcard, and wrote on the back: to malcolm carter.

Then he walked down the road a kilometer to one of Kate’s little touches, a postbox marked us mail, and dropped the postcard in.

19 (Remit not paucity) JUNE 2051

The anaesthetist said, “Count backward from ten.”

Maria said, “Ten.”

She dreamed of arriving on Francesca’s doorstep with a suitcase full of money. As she walked down the hall behind her mother, the case fell open, and hundred-dollar bills fluttered out and filled the air like confetti.

Francesca turned to her, radiant with health. She said tenderly, “You shouldn’t have, my darling. But I understand. You can’t take it with you.”

Maria laughed. “You can’t take it with you.”

Her father was in the living room, dressed for his wedding day, although not as young. He beamed and held out his arms to Maria. His parents, and Francesca’s parents, stood behind him—and as Maria approached, she saw from on high that behind her grandparents were cousins and aunts, great-grandparents and great-aunts, row after row of relatives and ancestors, stretching back into the depths of the house, laughing and chattering. The money had brought them all back to life. How could she have been so selfish as to think of denying them this grand reunion?

Maria threaded her way through the crowd, greeting people she’d never known existed. Handsome, dark-eyed seventh cousins kissed her hand and whispered compliments in a beautiful dialect she didn’t understand. Veiled widows in elegant black dresses stood arm-in-arm with their resurrected husbands. Children weaved between the adults’ legs, stealing food by the handful and cramming it into their mouths on the run.

The clinic’s neurologist turned out to be a distant relative. Maria cupped her hands to the woman’s ear and shouted over the noise of the party: “Have I been scanned yet? Will my Copy remember any of this?” The neurologist explained that the scan only captured memories laid down permanently as changes in synaptic strengths; the fleeting electrochemistry of this dream would be lost forever. She added cryptically, “Lost to whoever’s not having it.”

Maria felt herself waking. Suddenly afraid that she might be the Copy, she struggled to remain in the dream—as if she could force her way back through the crowd, back through the plot, and leave by a different exit. But the scene grew vague and unconvincing; she could feel the heavy presence of her waking body: her aching shoulders, her swollen tongue.

She opened her eyes. She was alone in the Landau Clinic’s cheerfully decorated recovery room; she’d been wheeled through for a patient’s-eye view before being given the anesthetic, so she’d know exactly what to expect. It took a few seconds for the truths of the dream to fade, though. Her father was dead. Her grandparents were dead. There’d been no grand reunion. There never would be.

As for the Copy… her scan file didn’t even exist, yet; the raw tomographic data would take hours to be processed into a high-resolution anatomical map. And she could still change her mind and keep the results out of Durham’s hands altogether. He’d paid the clinic for the scan, but if she refused to hand over the file there’d be nothing he could do about it.

The recovery room was softly lit, lined with odorless blue and orange flowers. Maria closed her eyes. If Durham’s logic meant anything, raw tomographic data could probably process itself, find itself conscious, as easily as any Copy who’d been chopped up and run at random. There was no need for a finished scan file.

No need even to be scanned; the very same data surely existed, scattered about the universe, whether or not it was ever plucked from her brain and assembled in what she thought of as one place.

In fact, if Durham was right—if the events he believed would take place in his TVC universe could find themselves in the dust—then those events would happen, regardless. It could make no difference what anyone did in this world. The whole Garden-of-Eden project was superfluous. Every permutation of the dust which was capable of perceiving itself, making sense of itself, would do just that. And all she would have achieved by refusing to be scanned would have been to deny the Maria of that permutation a history which seemed to overlap with her own particular life. While a third woman—in another world, another permutation—would have taken her place in that role.

Maria opened her eyes. She’d just recalled the first thing she’d meant to do on waking. Every scanner was programmed to recognize—in real time, before all the arduous data processing that followed—the magnetic resonance spectrum of four or five special dyes, which could be used for alignment and identification. The scanning technician had obligingly loaned her a “number three” marker pen—and instructed the scanner to blind itself to that particular dye.

She pulled her hands out from under the sheets. Her left palm still read: you are not the copy.

She licked her fingers and started rubbing the unnecessary words away.

+ + +

Maria arrived at the north Sydney flat around half past twelve. Two terminals were set up side by side on Durham’s kitchen table; other than that, the place was as bare as it had been the last time she’d called.

Although it wasn’t, technically, necessary, Maria had insisted that she and Durham be in the same physical location throughout what he called the “launch”—the running of the first moments of the TVC universe as software on a real computer, the act which would supposedly seed an independent, self-sustaining universe, taking up where the version relying on real-world hardware left off. At least this way she could monitor the keys he pressed and the words he spoke, without having to wonder if she was being shown what was really going on at that level. She had no idea what she was guarding against—but Durham was a highly intelligent man with some very strange beliefs, and she had no reason to feel confident that he’d revealed the full extent of his delusions. His clients had confirmed part of his story—and they would have had the resources to check much more of it than she had—but Durham might still have lied to them about what was going on inside his head.

She wanted to trust him, she wanted to believe that she’d finally reached the truth—but it was hard to put any limits on how wrong she might yet be. She felt she’d known him too long to seriously fear for her physical safety—but the possibility remained that everything she thought she’d understood about the man would turn out, once again, to have been utterly misconceived. If he came away from the kitchen sink brandishing a carving knife, calmly announcing his intention to sacrifice her to the Spirit of the New Moon, she’d have no right to feel betrayed, or surprised. She couldn’t expect to live off the proceeds of insanity, and also take for granted the usual parameters of civilized behavior.

The flesh-and-blood Durham was only half the problem. Once the program simulating a TVC cellular automaton was started, the plan was that neither she nor Durham would intervene at all. Any external tinkering would violate the automaton’s rules—the fundamental laws of the new universe—making a mockery of the whole endeavor. Only Durham’s Copy, being run on the simulated TVC computers, could act in harmony with those laws. They would always have the option of aborting the project, pulling the plug—but in every other respect, the Copy would be in control.

(Of course, aborting the simulation if something went wrong would not—in Durham’s eyes—prevent the spawning of an independent universe beyond their control… but it might leave them with enough unspent computer time for a second attempt.)

With her hands tied once the universe was running, her only way to influence what did or didn’t happen was through the Garden-of-Eden configuration—which included all the programs the TVC lattice would initially run. Maria had written part of this internal launch software herself; Durham had written, or commissioned, the rest, but she’d checked it all personally. And she’d built in a safeguard: all the Copies but Durham’s would be blocked from running until the TVC processors had solved a suitably intractable mathematical equation. Maria had estimated that the world’s combined computing resources couldn’t have cracked the problem in under a decade; thirty million dollars’ worth, minus overheads, wouldn’t come close. That was no obstacle in the eyes of Durham and his followers; the ever-growing resources of the burgeoning TVC universe would make light work of it, solving the equation within a week or two of the launch. But short of any such universe coming into existence—and so long as the test wasn’t circumvented—there was no chance of a second Maria Deluca, or anyone else, waking. It was her guarantee that there’d be no virtual Jonestown. Just one lone prophet flickering in and out of existence.

Durham made instant coffee. Maria surveyed the spartan room. She said, “This isn’t good enough, you know. We should have two hundred people wearing headsets, and a giant screen taking up an entire wall. Like one of the old NASA missions.”

Durham spoke over the sound of boiling water. “Don’t worry; we’ll be using more computing power per second than NASA used for the entire Apollo program.”

Computing power. One more thing to worry about. Maria logged on to the QIPS exchange; the rate was up slightly since she’d last checked, but so far there was no sign of what she dreaded. In the event that Operation Butterfly entered the market again, today of all days, the Garden of Eden would be frozen out, postponed until the QIPS rate returned to normal levels. That wouldn’t make the slightest difference to Durham or his followers—even if the launch program was thrown off the network halfway through, and only completed days, or weeks, later. Real time was irrelevant. Maria could appreciate the logic of that—but the thought of a delay, or an unexpected slowdown, still made her sick with anxiety. Every legal opinion she’d obtained had made it clear that neither she nor Durham were likely to face prosecution—and if charges were brought against them, a conviction was highly improbable… and even if that happened, an appeal would almost certainly succeed. Nonetheless, every day she’d spent working with Durham as a knowing “accomplice” had made her feel more vulnerable to the whims of the authorities. Hayden had treated her icily when she’d confessed to having abandoned her laughable “undercover” role. The risk of harassment would hardly vanish the moment the project was completed—but the relief would still be considerable.

She was beginning to regret having honored her promise not to try to record Durham’s clients’ statements assuring her that they were fully informed participants in the scheme. The authenticated messages she’d viewed—on public terminals—might not have been the equivalent of human testimony, but having them stored away on a chip somewhere would have made her feel a lot more secure. Regardless of the legal status of the Copies, she couldn’t imagine being prosecuted for fraud if she could show that the de facto “victims of the crime” knew exactly what they were paying for.

Durham set her coffee down on the table. Maria mumbled thanks as he sat beside her. He said, “No last-minute qualms? You can still back out if you want to.”

She kept her eyes on the screen, the flickering pie chart of the QIPS exchange. “Don’t tempt me.” As if she’d seriously consider blowing her one real chance to have Francesca scanned—after all the work, all the anxiety—for no better reason than a laughable, microscopic fear that this artificial universe might actually blossom into self-contained existence.

Durham’s terminal beeped. Maria glanced at his screen; a message box said PRIORITY COMMUNICATION. She looked away as he viewed the text.

“Speaking of last-minute qualms, Riemann’s changed his mind. He wants in.”

Maria said irritably, “Well, tell him it’s too late. Tell him he’s missed the boat.” She wasn’t serious; from what she knew of the project’s finances, Durham had been set to barely break even by the end of the day. The price of one more ticket would transform his fortunes completely.

He said, “Relax—it will take half an hour at the most to fit him in. And his fee will cover much more than the increase in data; we’ll be able to run the whole launch a bit longer.”

Maria had to pause to let that sink in. Then she said, “You’re going to blow most of two million ecus on stretching out something that—”

Durham smiled. “That what? That would have worked anyway?”

“That you believe would have worked anyway!”

“The longer I get to see my Copy observing the TVC universe, the happier I’ll be. I don’t know what it will take to anchor the automaton rules—but if ten watertight experiments sounds good, then eleven sounds better.”

Maria pushed her chair back and walked away from her terminal. Durham tapped at his keyboard, first invoking the programs which would recompute the Garden-of-Eden configuration to include the new passenger and his luggage—then directing the windfall from Riemann straight into the project’s JSN account.

She said, “What’s wrong with you? Two million ecus is more than two million dollars! You could have lived on that for the rest of your life!”

Durham kept typing, passing Riemann’s documents through a series of legal checks. “I’ll get by.”

“Given it to a charity, then!”

Durham frowned, but said patiently, “I gather that Thomas Riemann gives generously to famine relief and crop research every year. He chose to spend this money on a place in my sanctuary; it’s hardly my role to channel his funds into whatever you or I decide is the worthiest cause.” He glanced at her and added, mock-solemnly, “That’s called fraud, Ms. Deluca. You can go to prison for that.”

Maria was unmoved. “You could have kept something for yourself. For this life, this world. I don’t imagine any of your clients expected you to do all this for nothing.”

Durham finished at the terminal and turned to her. “I don’t expect you to understand. You treat the whole project as a joke—and that’s fine. But you can hardly expect me to run it on that basis.”

Maria didn’t even know what she was angry about anymore: the delayed launch, the obscene waste of money—or just Durham sitting there making perfect sense to himself, as always.

She said, “The project is a joke. Three hundred million people are living in refugee camps, and you’re offering sanctuary to sixteen billionaires! What do they need protection from? There’s never going to be an anti-Copy revolution! They’re never going to be shut down! You know as well as I do that they’ll just sit there getting richer for the next ten thousand years!”

“Possibly.”

“So you are a fraud then, aren’t you? Even if your ’sanctu-ary’ really does come into existence—even if you prove your precious theory right—what have your backers gained? You’ve sent their clones into solitary confinement, that’s all. You might as well have put them in a black box at the bottom of a mineshaft.”

Durham said mildly, “That’s not quite true. You talk about Copies surviving ten thousand years. What about ten billion? A hundred billion?”

She scowled. “Nothing’s going to last that long. Haven’t you heard? They’ve found enough dark matter to reverse the expansion of the universe in less than forty billion years—”

“Exactly. This universe isn’t going to last.”

Maria nodded sarcastically, and tried to say something belittling, but the words stuck in her throat.

Durham continued blithely, “The TVC universe will never collapse. Never. A hundred billion years, a hundred trillion; it makes no difference, it will always be expanding.”

Maria said weakly, “Entropy—”

“Is not a problem. Actually, ‘expanding’ is the wrong word; the TVC universe grows like a crystal, it doesn’t stretch like a balloon. Think about it. Stretching ordinary space increases entropy; everything becomes more spread out, more disordered. Building more of a TVC cellular automaton just gives you more room for data, more computing power, more order. Ordinary matter would eventually decay, but these computers aren’t made out of matter. There’s nothing in the cellular automaton’s rules to prevent them from lasting for ever.”

Maria wasn’t sure what she’d imagined before; Durham’s universe—being made of the same “dust” as the real one, merely rearranged—suffering the same fate? She couldn’t have given the question much thought, because that verdict was nonsensical. The rearrangement was in time as well as space; Durham’s universe could take a point of space-time from just before the Big Crunch, and follow it with another from ten million years b.c. And even if there was only a limited total amount of “dust” to work with, there was no reason why it couldn’t be reused in different combinations, again and again. The fate of the TVC automaton would only have to make internal sense—and the thing would have no reason, ever, to come to an end.

She said, “So you promised these people… immortality?”

“Of course.”

“Literal immortality? Outliving the universe?”

Durham feigned innocence, but he was clearly savoring the shock he’d given her. “That’s what the word means. Not dying after a very long time. Just not dying, period.”

Maria leaned back against the wall, arms folded, trying to cast aside the feeling that the whole conversation was as insubstantial as anything Durham had hallucinated in the Blacktown psychiatric ward. She thought: When Francesca’s been scanned I’m going to take a holiday. Visit Aden in Seoul, if I have to. Anything to get away from this city, this man.

She said, “Ideas like that are powerful things. One of these days you’re going to hurt someone.”

Durham looked wounded himself, at that. He said, “All I’ve tried to do is be honest. I know: I lied to you, at first—and I’m sorry. I had no right to do that. But what was I supposed to do with the truth? Keep it locked up in my head? Hide it from the world? Give no one else the chance to believe, or disbelieve?” He fixed his eyes on her, calm and sane as ever; she looked away.

He said, “When I first came out of hospital, I wanted to publish everything. And I tried… but nobody reputable was interested—and publishing in the junk-science journals would have been nothing but an admission that it was all bullshit. So what else could I do, except look for private backers?”

Maria said, “I understand. Forget it. You’ve done what you thought you had to—I don’t blame you for that.” The cliches nearly made her gag, but all she could think about was shutting him up. She was sick of being reminded that the ideas which were nothing but a means to an end, for her—the ideas she could turn her back on forever, in eight hours’ time—were this man’s entire life.

He looked at her searchingly, as if genuinely seeking guidance. “If you’d believed everything I believe, would you have kept it all to yourself? Would you have lived out your life pretending to the world that you’d merely been insane?”

Maria was saved from answering by a beep from Durham’s terminal. The Garden-of-Eden configuration had been recomputed; Thomas Riemann’s snapshot was now built into their cellular automaton equivalent of the Big Bang.

Durham swung his chair around to face the screen. He said cheerfully, “All aboard the ship of fools!”

Maria took her place beside him. She reached over and tentatively touched his shoulder. Without looking at her, he reached up and squeezed her hand gently, then removed it.

Following a long cellular automaton tradition, the program which would bootstrap the TVC universe into existence was called FIAT. Durham hit a key, and a starburst icon appeared on both of their screens.

He turned to Maria. “You do the honors.”

She was about to object, but then it didn’t seem worth arguing. She’d done half the work, but this was Durham’s creation, whoever cut the ribbon.

She prodded the icon; it exploded like a cheap flashy fire-work, leaving a pincushion of red and green trails glowing on the screen.

“Very tacky.”

Durham grinned. “I thought you’d like it.”

The decorative flourish faded, and a shimmering blue-white cube appeared: a representation of the TVC universe. The Garden-of-Eden state had contained a billion ready-made processors, a thousand along each edge of the cube—but that precise census was already out of date. Maria could just make out the individual machines, like tiny crystals; each speck comprised sixty million automaton cells—not counting the memory array, which stretched into the three extra dimensions, hidden in this view. The data preloaded into most of the processors was measured in terabytes: scan files, libraries, databases; the seed for Planet Lambert—and its sun, and its three barren sibling planets. Everything had been assembled, if not on one physical computer—the TVC automaton was probably spread over fif-teen or twenty processor clusters—at least as one logical whole. One pattern.

Durham reduced the clock rate until the blue-white shimmer slowed to a stroboscopic flickering, then a steady alternation of distinct colors. The outermost processors were building copies of themselves; in this view, blue coded for complete, working processors, and white coded for half-finished machines. Each layer of blue grew a layer of white, which abruptly turned blue, and so on. The skin of this universe came with instructions to build one more layer exactly like itself (including a copy of the same instructions), and then wait for further commands to be passed out from the hub.

Durham zoomed in by a factor of two hundred, slowed down the clock rate further, and then changed the representation to show individual automaton cells as color-coded symbols. The processors were transformed from featureless blue or white boxes into elaborate, multicolored, three-dimensional mazes, rectilinear filigree alive with sparks of light.

In the throes of reproduction, each processor could be seen sprouting hundreds of pairs of fine red and green “construction wires,” which grew straight out into the surrounding empty space—until they all reached the same predetermined length, abruptly turned a tight one-hundred-and-eighty degrees, and then started growing back in the opposite direction. Glowing with elaborate moving striations, the wires zig-zagged back and forth between the surface of the mother computer and an unmarked boundary plane—until between them, they’d filled in the region completely, like some strange electronic silk weaving itself into a solid cocoon.

In close-up, the wires resolved into long lines of cells marked with arrowheads, some rendered in the brighter hues which represented “activated” states. Glowing stripes built from the binary code of bright and dim moved down the wire from arrow to arrow: the data of the blueprint for the daughter machine being shuffled out from the central memory.

With the clock rate slowed still further, the process could be followed in detail. Wherever a pulse of brightness reached the end of a construction wire, the transparent “Vacuum” of the null state was transformed into an “embryonic” cell, shown as a nondescript gray cube. Subsequent data told the new cell what to become—each pulse, or absence of a pulse, converting it into a slightly more specialized transition state, zeroing in on the particular final state required. The construction wires grew out from the mother computer using this principle, extending themselves by building more of themselves at their tips.

Having filled the entire region which the daughter machine would occupy, they then worked backward, retracting one step at a time; unweaving their zig-zag cocoon, and leaving behind whatever the blueprint required. The whole process looked grotesquely inefficient—far more time was spent on extending and retracting the wires themselves than on creating the cells of the daughter machine—but it kept the rules of the automaton as simple as possible.

Durham said, “This all looks fine to me. Okay to proceed?”

“Sure.” Maria had grown mesmerized; she’d forgotten her urgency, forgotten herself. “Crank it up.” At any speed where they could keep track of events at the level of individual processors—let alone individual cells—nothing useful would ever get done. Durham let the clock rate revert to the maximum they could afford, and the grid became a blur.

In contrast, the next stage would be painfully slow. Durham made coffee and sandwiches. All the overheads of running a Copy on a system of computers which was, itself, a simulation, addled up to a slowdown of about two hundred and fifty. More than four real-time minutes to a subjective second. There was no question of two-way communication—the TVC universe was hermetic, no data which hadn’t been present from the outset could affect it in any way—but they could still spy on what was happening. Every hour, they could witness another fourteen seconds of what the Copy of Durham had done.

Maria spot-checked at other levels, starting with the software running directly on the TVC grid. The “machine language” of the TVC computers was about as arcane and ridiculous as that of any hypothetical Turing machine, six-dimensional or not, but it had been simple enough to instruct a metaprogrammer to write—and rigorously validate—a program which allowed them to simulate conventional modern computers. So the processor clusters in Tokyo or Dallas or Seoul were simulating a cellular automaton containing a lattice of bizarre immaterial computers… which in turn were simulating the logic (if not the physics) of the processor clusters themselves. From there on up, everything happened in exactly the same way as it did on a real machine—only much more slowly.

Maria munched cheese and lettuce between thick slices of white bread. It was a Tuesday afternoon; most of the flats around them were silent, and the street below was lifeless. The neighboring office blocks had no tenants, just a few furtive squatters; where the sun penetrated the nearest building at just the right angle, Maria could see clothes hung out to dry on lines stretched between office partitions.

Durham put on music, a twentieth-century opera called Einstein on the Beach. He didn’t own a sound system, but he called up the piece from a library he’d bought for the Garden of Eden, and had a background task play it through his terminal’s speakers.

Maria asked, “What will you do with yourself when this is over?”

Durham replied without hesitation. “Finish the whole set of fifty experiments. Start Planet Lambert unfolding. Celebrate for about a week. Stroll down the main street of Permutation City. Wait for your little locking device to disengage. Wake up my passengers in their own private worlds—and hope that some of them are willing to talk to me, now and then. Start catching up on Dostoyevsky. In the original—”

“Yeah, very funny. I said you, not him.”

I’d like to think of us as inseparable.”

“Seriously.”

He shrugged. “What will you do?”

Maria put her empty plate down, and stretched. “Oh… sleep in until noon, for a week. Lie in bed wondering exactly how I’m going to break the news to my mother that she can now afford to be scanned—without making it sound like I’m telling her what to do.”

“Perish the thought.”

Maria said simply, “She’s dying. And she can save herself—without hurting anyone. Without stealing food from the mouths of the next generation, or whatever it is she thinks makes being scanned such a crime. Do you really think she—honestly—doesn’t want to stay alive? Or wouldn’t want to, if she could think it through clearly, without all the guilt and moralizing bullshit her generation saddled her with?”

Durham wasn’t taking sides. “I don’t know her, I can’t answer that.”

“She was a child of the nineties. Her kindergarten teachers probably told her that the pinnacle of her existence would be fertilizing a rainforest when she died.” Maria thought it over. “And the beauty of it is… she can still do that. Scan her, put her through a meat grinder… scatter the results over the Daintree.”

“You’re a sick woman.”

“I’ll have the money soon. I can afford to joke.”

Their terminals chimed simultaneously; the first fourteen seconds of life inside were ready to be viewed. Maria felt the food she’d just swallowed harden into a lump like a closed fist in her gut. Durham told the program to proceed.

The Copy sat in a simple, stylized control room, surrounded by floating interface windows. One window showed a representation of a small part of the TVC lattice. The Copy couldn’t take the same God’s-eye view of the lattice as they had; the software they’d used could only function on a level right outside his universe. There was no simple way he could discover the state of any given automaton cell; instead, a system of construction and sensor wires (all joined to specialized processors) had been built around a small region in the center of the lattice. Durham had christened this apparatus “the Chamber.” What went on deep inside the Chamber could be deduced, indirectly, from the data which ended up flowing down the sensor wires. It wasn’t as complicated as working out what had happened in a particle accelerator collision, based on the information registered by surrounding detectors—but the principle was the same, and so was the purpose. The Copy had to conduct experiments to test his own fundamental “laws of physics”—the TVC automaton’s rules. And the (simulated) modern computers running his VR environment had a (simulated) link to the Chamber, like the real-world computers linked to any real-world accelerator.

The Copy said, “Setting up the first experiment.” He deftly typed a sequence of code letters on his keyboard. Durham had rehearsed the whole thing before his scan, until he could perform each of fifty experiments in ten seconds flat, but Maria was still astonished that the Copy—who had woken abruptly to find himself seated in the control room, without any preliminaries, any chance to grow accustomed to his identity, and his fate—had had the presence of mind to leap straight into the task. She’d entertained visions of this first version of Durham to wake inside a computer finally realizing that “the other twenty-three times” were nothing at all like the real experience—and telling his original about it in no uncertain terms. But there didn’t seem to be much chance of that; the Copy just sat there typing as if his life depended on it.

The experimental setups could have been automated. The checking of the results could have been automated, too. The Copy could have spent two minutes sitting and watching a flashing green sign which said everything is just what you would have expected, don’t worry about the messy details. There was no such thing as a set of perceptions for the Copy which could prove that he inhabited a cellular automaton which obeyed all the rules which he hoped were being obeyed. It was all down to Occam’s razor in the end—and hoping that the simplest explanation for perceiving a diplay showing the correct results was that the correct results were actually occurring.

Maria stared into the screen, over the Copy’s shoulder, at the interface window within. When he typed the last code letter, the assembly of cells he’d constructed in the Chamber became unstable and started creating new cells in the surrounding “vacuum,” setting off a cascade which eventually impinged on the sensor wires. Disconcertingly, the Copy watched both a simulation—on his own terms—of what ought to be happening in the Chamber, and then a moment later a reconstruction of the “actual” events, based on the sensor data.

Both evidently matched the results of the simulations which the original Durham had committed to memory. The Copy clapped his hands together loudly in obvious jubilation, bellowed something incoherent, then said, “Setting up the sec—”

Maria was becoming giddy with all the levels of reality they were transecting—but she was determined to appear as blasé as ever. She said, “What did you do, wake him up with a brain full of amphetamines?”

Durham replied in the same spirit. “No, he’s high on life. If you’ve only got two minutes of it, you might as well enjoy it.”

They waited, passing the time checking software more or less at random, displaying everything from firing patterns in the Copy’s model brain to statistics on the performance of the TVC computers. Intuitively, the elaborate hierarchy of simulations within simulations seemed vulnerable, unstable—every level multiplying the potential for disaster. But if the setup resembled a house of cards, it was a simulated house of cards: perfectly balanced in a universe free of vibrations and breezes. Maria was satisfied that the architecture at every level was flawless—so long as the level beneath held up. It would take a glitch in the real-world hardware to bring the whole thing tumbling down. That was rare, though not impossible.

They viewed the second installment of the Copy at work, then took a coffee break. Einstein on the Beach was still playing, repetitive and hypnotic. Maria couldn’t relax; she was too wired on caffeine and nervous energy. She was relieved that everything was running smoothly—no software problems, no Operation Butterfly, no sign of either version of Durham going weird on her. At the same time, there was something deeply unsettling about the prospect of the whole thing unwinding, exactly as predicted, for the next six hours—and then simply coming to an end. She’d have the money for Francesca, then, and that justified everything… but the absolute futility of what they were doing still kept striking her anew—in between bouts of worrying over such absurdities as whether or not she could have made a better job of A. hydrophila’s response to dehydration. Durham would let her publish all the Autoverse work, so that hadn’t been a complete waste of time—and she could keep on refining it for as long as she liked before unleashing it on the skeptics… but she could already imagine the—bizarre—regret she’d feel because the improvements had come too late to be incorporated into the “genuine” Planet Lambert: the one they were currently flushing down a multi-million-dollar drain.

She said, “It’s a pity none of your passengers’ originals have bodies. Having paid for all this, they should be here, watching.”

Durham agreed. “Some of them may be here in spirit; I’ve granted them all the same viewing access to the simulation that we have. And their auditors will receive a verified log of everything—proof that they got what they paid for. But you’re right. This isn’t much of a celebration; you should be clinking glasses and sharing caviar with the others.”

She laughed, offended. “Others? I’m not one of your victims—I’m just the confidence artist’s accomplice, remember? And I’m not here to celebrate; I’m only here to make sure your doppelgänger doesn’t hot-wire the software and wake me up.”

Durham was amused. “Why would he try to wake you so soon? Do you think he’s going to become unbearably lonely in the space of two minutes?”

“I have no idea what he might do, or why. That’s the whole problem. He’s just as fucked up as you are.”

Durham said nothing. Maria wished she could take back the words. What was the point of needling him and mocking him, again and again—did she think she could ever bring him down to Earth? It was all a matter of pride; she couldn’t let a second go by without reminding him that she hadn’t been seduced by his ideas. Computer junkie, artificial life freak; she still had her feet planted firmly in the real world. His vision of an Autoverse biosphere had impressed her—when she’d thought he’d understood that it could never be anything but a thought experiment. And all the work he’d done on the TVC universe was ingenious—however ultimately pointless it was. In a way, she even admired his stubborn refusal to give in to common sense and accept his delusions for what they were.

She just couldn’t bear the thought that he harbored the faintest hope that he’d persuaded her to take the “dust hypothesis” seriously.

+ + +

At three minutes past ten, the money ran out—all but enough to pay for the final tidying-up. The TVC automaton was shut down between clock ticks; the processors and memory which had been allocated to the massive simulation were freed for other users—the memory, as always, wiped to uniform zeroes first for the sake of security. The whole elaborate structure was dissolved in a matter of nanoseconds.

Night had turned the windows of the flat to mirrors. No lights showed in the empty office towers; if there’d been cooking fires from the squatters, they’d been extinguished long ago. Maria felt disconnected, adrift in time; the trip north across the harbor bridge in sunlight seemed like a distant memory, a dream.

The individual components of the Garden of Eden were still held in mass storage. Maria deleted her scan file, carefully checking the audit records to be sure that the data hadn’t been read more often than it should have been. The numbers checked out; that was no guarantee, but it was reassuring.

Durham deleted everything else.

The recordings of the spy software remained, and they viewed the last brief scene of the Copy at work—and then replayed the whole two-minute recording.

Maria watched with a growing sense of shame. The individual fragments had barely affected her, but viewed without interruption, the Copy took on the air of a deranged sect leader driving a bus full of frozen billionaires straight toward the edge of a cliff—accelerating euphorically in the sure and certain knowledge that the thing would fly, carrying them all off into a land beyond the sunset. She clung to her rationalizations: the Copy’s limited separate identity, his joyful demise.

When the replay stopped in mid-experiment, Durham closed his eyes and let his head hang forward. He wept silently. Maria looked away.

He said, “I’m sorry. I’m embarrassing you.”

She turned back to him; he was smiling, and sniffling. She wanted to embrace him; the urge was half sisterly, half sexual. He was pale and unshaven, obviously drained—but there was more life in his eyes than ever, as if the fulfilment of his obsession had liberated him from his past so completely that he faced the world now like a newborn child.

He said, “Champagne?”

Maria hardened her heart. She still had no reason to trust him. She said, “Let me check my bank balance first; I might not have anything to celebrate.” Durham giggled, as if the very idea that he might have cheated her was preposterous. She ignored him, and used the terminal. The six hundred thousand dollars he’d promised had been deposited.

She stared at the digits on the screen for a while, numb with the strange truth that the simple pattern of data they represented, sanctified as “wealth,” could travel out into the living, breathing, decaying world… and return, enriched beyond measure: imprinted with everything which made Francesca human.

She said, “One glass. I’m cycling.”

+ + +

They emptied the bottle. Durham paced around the flat, growing increasingly hyperactive. “Twenty-three Copies! Twenty-three lives! Imagine how my successor must be feeling, right now! He has the proof, he knows he was right. All I have is the knowledge that I gave him that chance—and even that’s too much to bear.” He wept again, stopped abruptly. He turned and gazed at Maria imploringly. “I did it all to myself, but it was still madness, still torture. Do you think I knew, when I started out, how much pain and confusion there’d be? Do you think I knew what it would do to me? I should have listened to Elizabeth—but there is no Elizabeth here. I’m not alive. Do you think I’m alive! If a Copy’s not human, what am I? Twenty-three times removed?”

Maria tried to let it wash over her. She couldn’t feel simple compassion—she was too tainted, too culpable—so she tried to feel nothing at all. Durham had systematically pursued his beliefs as far as they could take him; he’d either be cured by that, or ready for another round of nanosurgery. Nothing she could do now would make any difference. She started to tell herself that by helping with the project—without ever conceding its premise—she might have helped him exorcize his delusions… but that wasn’t the point. She’d done it all for the money. For Francesca. And for herself. To spare herself the pain of Francesca dying. How dare the woman think of refusing? Copies, like funerals, were for the benefit of the survivors.

Durham suddenly went quiet. He sat down beside her, disheveled and contrite; she wasn’t sure if he’d become sober, or just moved on to a new phase. It was half past two; the opera had finished playing hours ago, the flat was silent.

He said, “I’ve been ranting. I’m sorry.”

The two swivel chairs they’d been sitting on all day were the only furniture in the room besides the table; there was no sofa she could sleep on, and the floor looked cold and hard. Maria thought about heading home; she could catch a train, and collect her cycle later.

She stood; then, barely thinking about it, leaned down and kissed him on the forehead.

She said, “Goodbye.”

Before she could straighten up, he put a hand on her cheek. His fingers were cool. She hesitated, then kissed him on the mouth—then almost recoiled, angry with herself. I feel guilty, I feel sorry for him, I only want to make up for that somehow. Then he met her eyes. He wasn’t drunk any more. She believed he understood everything she was feeling—the whole knot of confusion and shame—and all he wanted to do was smooth it away.

They kissed again. She was sure.

They undressed each other on the way to the bedroom. He said, “Tell me what you want, tell me what you like. I haven’t done this for a long time.”

“How long?”

“Several lives ago.”

He was skilled with his tongue, and persistent. She almost came—but before it could happen, everything broke down into isolated sensations: pleasant but meaningless, faintly absurd. She closed her eyes and willed it, but it was like trying to cry for no reason. When she pushed him gently away, he didn’t complain, or apologize, or ask stupid questions; she appreciated that.

They rested, and she explored his body. He was probably the oldest man she’d ever seen naked; certainly the oldest she’d ever touched. Fifty. He was… loose, rather than flabby; muscle had wasted rather than turned to fat. It was almost impossible to imagine Aden—twenty-four years old, and hard as a statue—ever succumbing to the same process. But he would. And her own body had already begun.

She slithered around and took his penis in her mouth, trying to psych herself past the comic strangeness of the act, trying to grow drunk on the stench of it, working with her tongue and teeth until he begged her to stop. They rearranged their bodies clumsily so they were side by side; he entered her and came at once. He cried out, bellowing in obvious pain, not histrionic delight. He gritted his teeth and turned ashen as he withdrew; she held his shoulders until he could explain. “My… left testicle went into spasm. It just… happens sometimes. It feels like it’s being crushed in a vise.” He laughed and blinked away the tears. She kissed him and ran a finger around his groin.

“That’s awful. Does it still hurt?”

“Yes. Don’t stop.”

Afterward, she found she didn’t want to touch him; his skin turned clammy as their sweat dried, and when he seemed to fall asleep, she disentangled herself from his embrace and shifted to the edge of the bed.

She didn’t know what she’d done: complicated everything, set herself up for yet another stage in their convoluted relationship—or simply marked the end of it, bidden him farewell? An hour of disastrous sex hadn’t resolved anything: she still felt guilty for taking the money, “taking advantage” of him.

What would she do, if he wanted to see her again? She couldn’t face the prospect of spending the next six months listening to him fantasize about the grand future which lay ahead for his homemade universe. She’d taken some pride in the fact that she’d never once humored him, never pretended for a moment to have accepted his theories—and she’d never met a nominally sane person who could disagree with her so graciously. But there’d be something dishonest about trying to forge a lasting friendship between them, in the face of her skepticism. And if she ever succeeded in disillusioning him… she’d probably feel guilty about that, as well.

The long day was catching up with her, it was too hard to think it through. Decisions would have to wait until morning.

Light from the kitchen spilled through the doorway onto her face; she called out softly to the house controller, to no effect, so she got up and switched off the light manually. She heard Durham stir as she felt her way back into the room. She paused in the doorway, suddenly reluctant to approach him.

He said, “I don’t know what you think, but I didn’t plan this.”

She laughed. What did he think he’d done? Seduced her? “Neither did I. All I ever wanted from you was your money.”

He was silent for a moment, but she could see his eyes and teeth flashing in the dark, and he seemed to be smiling.

He said, “That’s all right. All I wanted from you was your soul.”

20 (Can’t you time trip?)

Resting between descents, Peer looked up and finally realized what had been puzzling him. The clouds above the skyscraper were motionless; not merely stationary with respect to the ground, but frozen in every detail. The wispiest tendrils at the edges, presumably vulnerable to the slightest breeze, remained undisturbed for as long as he studied them. The shape of every cloud seemed flawlessly natural—but all the dynamism implicit in the wind-wrought forms, compelling at a glance, was pure illusion. Nothing in the sky was changing.

For a moment, he was simply bemused by this whimsical detail. Then he remembered why he’d chosen it.

Kate had vanished. She’d lied; she hadn’t cloned herself at all. She’d moved to Carter’s city, leaving no other version behind.

Leaving him—or half of him—alone.

The revelation didn’t bother him. On the skyscraper, nothing ever did. He clung to the wall, recuperating happily, and marveled at what he’d done to heal the pain. Back in cloud time, before he’d always been descending.

He’d set up the environment as usual—the city, the sky, the building—but frozen the clouds, as much to simplify things as to serve as a convenient reminder.

Then he’d mapped out a series of cues for memory and mood changes over fifteen subjective minutes. He’d merely sketched the progression, like a naive musician humming a melody to a transcriber; the software he’d used had computed the actual sequence of brain states. Moment would follow moment “naturally”; his model-of-a-brain would not be forced to do anything, but would simply follow its internal logic. By fine-tuning that logic in advance and loading the right memories, the desired sequence of mental events would unfold: from A to B to C to… A.

Peer looked over his shoulder at the ground, which never grew closer, and smiled. He’d dreamed of doing this before, but he’d never had the courage. Losing Kate forever—while knowing that he was with her—must have finally persuaded him that he had nothing to gain by putting it off any longer.

The scheme wouldn’t slip his mind completely—he could vaguely remember experiencing exactly the same revelation several times before—but his short-term memory had been selectively impaired to limit the clarity of this recursive false history, and once he was distracted, a series of free associations would eventually lead him back to exactly the state of mind he’d been in at the cycle’s beginning. His body—with respect to every visible cue in the environment—would also be back where it had started. The ground and sky were static, and every story of the building was identical, so his perceptions would be the same. And every muscle and joint in his body would have recovered perfectly, as always.

Peer laughed at his cloud-self’s ingenuity, and started to descend again. It was an elegant situation, and he was glad he’d finally had a cloud-reason to make it happen.

There was one detail, though, which he couldn’t focus on, one choice he’d made back in cloud time which he seemed to have decided to obscure from himself completely.

Had he programmed his exoself to let him run through the cycle a predetermined number of times? ABCABCABC… and then some great booming DBF breaking through the sky like the fist of God—or a tendril of cumulus actually moving—putting an end to his perpetual motion? A grappling hook could tear him from the side of the building, or some subtle change in the environment could nudge his thoughts out of their perfectly circular orbit. Either way, experiencing one uninterrupted cycle would be the same as experiencing a thousand, so if there was an alarm clock ticking away at all, his next cycle—subjectively—would be the one when the buzzer went off.

And if there was no clock? He might have left his fate in external hands. A chance communication from another Copy, or some event in the world itself, could be the trigger which would release him.

Or he might have chosen absolute solipsism. Grinding through the cycle whatever else happened, until his executor embezzled his estate, terrorists nuked the supercomputers, civilization crumbled, the sun went out.

Peer stopped and shook his head to flick sweat out of his eyes. The sense of déjá vu the action triggered was, presumably, purely synthetic; it told him nothing about the number of times he’d actually repeated the gesture. It suddenly struck him as unlikely that he’d done anything as inelegant as running the cycle more than once. His subjective time closed up in a loop, rolled in on itself; there was no need to follow the last moment with an external repetition of the first. Whatever happened—externally—“afterward,” the loop was subjectively seamless and complete. He could have shut himself down completely, after computing a single cycle, and it would have made no difference.

The breeze picked up, cooling his skin. Peer had never felt so tranquil; so physically at ease, so mentally at peace. Losing Kate must have been traumatic, but he’d put that behind him. Once and for ever.

He continued his descent.

21 (Remit not paucity) JUNE 2051

Maria woke from a dream of giving birth. A midwife had urged her, “Keep pushing! Keep pushing!” She’d screamed through gritted teeth, but done as she was told. The “child” had turned out to be nothing but a blood-stained statue, carved from smooth, dark wood.

Her head was throbbing. The room was in darkness. She’d taken off her wristwatch, but she doubted that she’d been asleep for long; if she had, the bed would have seemed unfamiliar, she would have needed time to remember where she was, and why. Instead, the night’s events had come back to her instantly. It was long after midnight, but it wasn’t a new day yet.

She sensed Durham’s absence before reaching across the bed to confirm it, then she lay still for a while and listened. All she heard was distant coughing, coming from another flat. No lights were on; she would have seen the spill.

The smell hit her as she stepped out of the bedroom. Shit and vomit, with a sickly sweet edge. She had visions of Durham reacting badly to a day of stress and a night of champagne, and she almost turned around and went back to the bedroom, to open the window and bury her face in a pillow.

The bathroom door was half-closed, but there were no sound effects suggesting that he was still in there; not a moan. Her eyes began to water. She couldn’t quite believe that she’d slept through all the noise.

She called out, “Paul? Are you all right?” There was no reply. If he was lying unconscious in a pool of vomit, alcohol had nothing to do with it; he had to be seriously ill. Food poisoning? She pushed open the door and turned on the light.

He was in the shower recess. She backed out of the room quickly, but details kept registering long after she’d retreated. Coils of intestine. Bloodred shit. He looked like he’d been kneeling, and then sprawled sideways. At first, she was certain that she’d seen the knife, red against the white tiles—but then she wondered if in fact she’d seen nothing but the Rorschach blot of a random blood stain.

Maria’s legs started to give way. She made it to one of the chairs. She sat there, light-headed, fighting to remain conscious; she’d never fainted in her life, but for a time it was all she could do to keep herself from blacking out.

The first thing she felt clearly was a sense of astonishment at her own stupidity, as if she’d just marched, with her eyes wide open, straight into a brick wall. Durham had believed that his Copy had achieved immortalityand proved the dust hypothesis. The whole purpose of his own life had been fulfilled by the project’s completion. What had she expected him to do, after that? Carry on selling insurance?

It was Durham she’d heard screaming through gritted teeth, shaping her dream.

And it was Durham who’d kept pushing, Durham who looked like he’d tried to give birth.

She called for an ambulance. “He’s cut his abdomen open with a knife. The wound is very deep. I didn’t look closely, but I think he’s dead.” She found that she could speak calmly to the emergency services switchboard puppet; if she’d had to say the same things to a human being, she knew she would have fallen apart.

When she hung up, her teeth started chattering, and she kept emitting brief sounds of distress which didn’t seem to belong to her. She wanted to get dressed before the ambulance and police arrived, but she didn’t have the strength to move—and the thought of even caring if she was discovered naked began to seem petty beyond belief. Then something broke through her paralysis, and she rose to her feet and staggered around the room, picking up the clothes they’d scattered on the floor just hours before.

She found herself fully dressed, slumped in a corner of the living room, reciting a litany of excuses in her head. She’d never humored him. She’d argued against his insane beliefs at every opportunity. How could she have saved him? By walking out on the project? That would have changed nothing. By trying to get him committed? His doctors had already pronounced him cured.

The worst thing she’d done was stand by and let him shut down his own Copy.

And there was still a chance

She sprang to her feet, rushed over to the nearest terminal, and logged back on to the project’s JSN account.

But Durham’s scan file was gone, deleted as meticulously, as irreversibly, as her own. The audit records showed no sign that the data had been preserved elsewhere; like her own file, it had even been flagged explicitly for exclusion from the JSN’s automatic hourly backups. The only place the data had been reproduced had been inside the Garden-of-Eden configuration itself—and every trace of that structure had been obliterated.

She sat at the terminal, replaying the file which showed Durham’s Copy conducting his experiments: testing the laws of his universe, rushing joyfully toward… what? The unheralded, inexplicable annihilation of everything he was in the process of establishing as the basis for his own existence?

And now his corpse lay in the bathroom, dead by his own hands, on his own terms; victim of his own seamless logic.

Maria buried her face in her hands. She wanted to believe that the two deaths were not the same. She wanted to believe that Durham had been right, all along. What had the JSN computers in Tokyo and Seoul meant to the Copy? No experiment performed within the TVC universe could ever have proved or disproved the existence of those machines. They were as irrelevant—to him—as Francesca’s ludicrous God Who Makes No Difference.

So how could they have destroyed him? How could he be dead?

There were quick, heavy footsteps outside, then a pounding on the door. Maria went to open it.

She wanted to believe, but she couldn’t.

22 (Remit not paucity) JUNE 2051

Thomas prepared himself to witness a death.

The flesh-and-blood Riemann was the man who’d killed Anna—not the Copy who’d inherited the killer’s memories. And the flesh-and-blood Riemann should have had the opportunity to reflect on that, before dying. He should have had a chance to accept his guilt, to accept his mortality. And to absolve his successor.

That hadn’t been allowed to happen.

But it wasn’t too late, even now. A software clone could still do it for him—believing itself to be flesh and blood. Revealing what the mortal, human self would have done, if only it had known that it was dying.

Thomas had found a suitable picture in a photo album—old chemical hardcopy images which he’d had digitized and restored soon after the onset of his final illness. Christmas, 1985: his mother, his father, his sister Karin and himself, gathered outside the family home, dazzled by the winter sunshine. Karin, gentle and shy, had died of lymphoma before the turn of the century. His parents had both survived into their nineties, showing every sign of achieving immortality by sheer force of will—but they’d died before scanning technology was perfected, having scorned Thomas’s suggestion of cryonic preservation. “I have no intention,” his father had explained curtly, “of doing to myself what nouveau riche Americans have done to their pets.” The young man in the photograph didn’t look much like the image Thomas would have conjured up by closing his eyes and struggling to remember—but the expression on his face, captured in transition from haunted to smug, rang true. Half afraid that the camera would reveal his secret; half daring it to try.

Thomas had kept copies of his deathbed scan file—off-line, in vaults in Geneva and New York—with no explicit purpose in mind, other than the vaguest notion that if something went irreparably wrong with his model, and the source of the prolem—a slow virus, a subtle programming error—rendered all of his snapshots suspect, starting life again with no memories since 2045 would be better than nothing.

Having assembled the necessary elements, he’d scripted the whole scenario in advance and let it run—without observing the results. Then he’d frozen the clone and sent it to Durham at the last possible moment—without giving himself a chance to back out, or, worse, to decide that he’d botched the first attempt, and to try again.

Now he was ready to discover what he’d done, to view the fait accompli. Seated in the library—with the drinks cabinet locked—he gestured to the terminal to begin.

The old man in the bed looked much worse than Thomas had expected: sunken-eyed, jaundiced and nearly bald. (So much for the honesty of his own appearance, the “minimal” changes he’d made to render himself presentable.) His chest was furrowed with scars, criss-crossed by a grid of electrodes; his skull was capped with a similar mesh. A pump suspended beside the bed fed a needle in his right arm. The clone was sedated by a crudely modeled synthetic opiate flowing into his crudely modeled bloodstream, just as Thomas’s original had been sedated by the real thing, from the time of the scan until his death three days later.

In this replay, though, the narcotic was scheduled to undergo a sudden drop in concentration—for no physically plausible reason, but none was required. A graph in a corner of the screen plotted the decline.

Thomas watched, sick with anxiety, feverish with hope. This—at last—was the ritual which he’d always believed might have cured him.

The old man attained consciousness, without opening his eyes; the EEG waveforms meant nothing to Thomas, but the software monitoring the simulation had flagged the event with a subtitle. Further text followed:

The anesthetic still hasn’t taken. Can’t they get anything right? [Garbled verbalization.] The scan can’t be over. I can’t be the Copy yet. The Copy will wake with a clear head, seated in the library, premodified to feel no disorientation. So why am I awake?

The old man opened his eyes.

Thomas shouted, “Freeze!” He was sweating, and nauseous, but he made no move to banish the unnecessary symptoms. He wanted catharsis, didn’t he? Wasn’t that the whole point? The subtitles gave only a crude hint of what the clone was experiencing. Much greater clarity was available; the recording included traces from key neural pathways. If he wanted to, he could read the clone’s mind.

He said, “Let me know what he’s thinking, what he’s going through.” Nothing happened. He clenched his fists and whispered, “Restart.”

The library vanished; he was flat on his back in the hospital bed, staring up at the ceiling, dazed. He looked down and saw the cluster of monitors beside him, the wires on his chest. The motion of his eyes and head was wrong—intelligible, but distressingly out of synch with his intentions. He felt fearful and disoriented—but he wasn’t sure how much of that was his own reaction and how much belonged to the clone. Thomas shook his own head in panic, and the library—and his body—returned.

He stopped the playback, and reconsidered.

He could break free any time he wanted to. He was only an observer. There was nothing to fear.

Fighting down a sense of suffocation, he closed his eyes and surrendered to the recording.

+ + +

He looked around the room groggily. He wasn’t the Copy—that much was certain. And this wasn’t any part of the Landau Clinic; as a VIP shareholder and future client, he’d toured the building too many times to be wrong about that. If the scan had been postponed for some reason, he ought to be back home—or on his way. Unless something had gone wrong requiring medical attention which the Landau was unable to provide?

The room was deserted, and the door was closed. He called out hoarsely, “Nurse!” He was too weak to shout.

The room controller replied, “No staff are available to attend to you, at present. Can I be of assistance?”

“Can you tell me where I am?”

“You’re in Room 307 of Valhalla.”

“Valhalla?” He knew he’d done business with the place, but he couldn’t remember why.

The room controller said helpfully, “Valhalla is the Health Dynamics Corporation of America’s Frankfurt Hospice.”

His bowels loosened with fright; they were already empty. [Thomas squirmed in sympathy, but kept himself from breaking free.] Valhalla was the meat-rack he’d hired to take care of his comatose body until it expired, after the scan—with the legal minimum of medical attention, with no heroic measures to prolong life.

He had been scanned—but they’d fucked up.

They’d let him wake.

It was a shock, but he came to terms with it rapidly. There was no reason to panic. He’d be out of here and scanned again in six hours flat—and whoever was responsible would be out on the street even faster. He tried to raise himself into a sitting position, but he was too dizzy from the lingering effects of the drug infusion to coordinate the action. He slumped back onto the pillows, caught his breath, and forced himself to speak calmly.

“I want to talk to the director.”

“I’m sorry, the director is not available.”

“Then, the most senior member of staff you can find.”

“No staff are available to attend to you, at present.”

Sweat trickled into his eyes. There was no point screaming about lawsuits to this machine. In fact… it might be prudent not to scream about lawsuits to anyone. A place like this would be perfectly capable of responding by simply drugging him back into a coma.

What he needed to do was let someone outside know about the situation.

He said, “I’d like to make a phone call. Can you connect me to the net?”

“I have no authority to do that.”

“I can give you an account number linked to my voiceprint, and authorize you to charge me for the service.”

“I have no authority to accept your account number.”

“Then… make a call, reversing all charges, to Rudolf Dieterle, of Dieterle, Hollingworth and Partners.”

“I have no authority to make such a call.”

He laughed, disbelieving. “Are you physically capable of connecting me to the net at all?”

“I have no authority to disclose my technical specifications.”

Any insult would have been a waste of breath. He lifted his head and surveyed the room. There was no furniture; no drawers, no table, no visitor’s chair. Just the monitors to one side of his bed, mounted on stainless steel trolleys. And no terminal, no communications equipment of any kind—not even a wall-mounted audio handset.

He probed the needle in his forearm, just below the inside of the elbow. A tight, seamless rubber sleeve, several centimeters wide, covered the entry point; it seemed to take forever to get his fingernails under the edge—and once he’d succeeded, it was no help. The sleeve was too tight to be dragged down his arm, and too elastic to be rolled up like a shirt sleeve. How did anyone, ever, take the thing off? He tugged at the drip tube itself; held in place by the sleeve, it showed no sign of yielding. The other end vanished inside the drug pump.

[Thomas began to wonder if the immovable needle, on top of the Kafkaesque room controller, would make the clone suspicious—but it seemed that the possibility of some future self waking the scan file a second time was too convoluted an explanation to occur to him in the middle of a crisis like this.]

He’d have to take the pump with him. That was a nuisance—but if he was going to march through the building wrapped in a sheet, looking for a terminal, it could hardly make him more conspicuous than he would have been anyway.

He started to peel the electrodes from his chest when a pulse of numbing warmth swept through his right arm. The pump beeped twice; he turned to see a green LED glowing brightly in the middle of the box, a light he hadn’t noticed before.

The wave of paralysis spread out from his shoulder before he could react—crimp the tube? He tried to roll himself out of the bed but if his body responded at all, he couldn’t feel it.

His eyes fluttered closed. He struggled to remain conscious—and succeeded. [The script guaranteed the clone several minutes of lucidity—which had nothing to do with the opiate’s true pharmacological effects.]

There’d be a computer log of his EEG. Someone would be alerted, soon, to the fact that he’d been awake… and they’d understand that the only humane thing to do would be to revive him.

But someone should have been alerted the moment he woke.

It was far more likely that he’d be left to die.

[Thomas felt ill. This was sadistic, insane.

It was too late for squeamishness, though. Everything he was witnessing had already happened.]

His body was numb, but his mind was crystalline. Without the blur of visceral distractions, his fear seemed purer, sharper than anything he’d ever experienced.

He tried to dredge up the familiar, comforting truths: The Copy would survive, it would live his life for him. This body was always destined to perish; he’d accepted that long ago. Death was the irreversible dissolution of the personality; this wasn’t death, it was a shedding of skin. There was nothing to fear.

Unless he was wrong about death. Wrong about everything.

He lay paralyzed, in darkness. Wishing for sleep; terrified of sleep. Wishing for anything that might distract him; afraid of wasting his last precious minutes, afraid of not being prepared.

Prepared? What could that mean? Extinction required no preparation. He wasn’t making any deathbed pleas to a God he’d stopped believing in at the age of twelve. He wasn’t about to cast aside seventy years of freedom and sanity, to return to his infantile faith. Approach the Kingdom of Heaven as a child, or you won’t get in? That very line was one which had helped him see through the crude mechanics of entrapment; the translation was all too obvious (even to a child): This bullshit would insult any adult’s intelligencebut swallow it anyway, or you’ll burn forever.

He was still afraid, though. The hooks had gone in deep.

The irony was, he had finally come to his senses and abandoned the whole insane idea of having himself woken, intentionally. To confront his mortality! To purge his Copy of guilt! What a pathetic fucking joke that would have been. And now the supposed beneficiary of the fatuous gesture would never even know that it had happened, anyway, by accident.

The blackness in his skull seemed to open out, an invisible view expanding into an invisible vista. Any sense of being in the hospice bed, merely numb and sightless, was gone now; he was lost on a plain of darkness.

What could he have told the Copy, anyway? The miserable truth? I’m dying in fear. I killed Anna for no reason but selfishness and cowardiceand now, in spite of everything, I’m still afraid that there might be an afterlife. A God. Judgement. I’ve regressed far enough to start wondering if every childish superstition I ever held might yet turn out to be truebut not far enough to embrace the possibility of repentance.

Or some anodyne lie? I’m dying in peace, I’ve found forgiveness, I’ve laid all my ghosts to rest. And you’re free, now, to live your own life. The sins of the father will not be visited upon the son.

Would that have worked, would that have helped? Some formula as inane as the voodoo of Confession, as glib as the dying words of some tortured soul finding Hollywood redemption?

He felt himself moving across the darkness. No tunnels of light; no light at all. Sedative dreams, not near-death hallucinations. Death was hours or days away; by then he’d surely be comatose again. One small mercy.

He waited. No revelations, no insights, no lightning bolts of blinding faith. Just blackness and uncertainty and fear.

+ + +

Thomas sat motionless in front of the terminal long after the recording had finished.

The clone had been right: the ritual had been pointless, misguided. He was and always would be the murderer; nothing could make him see himself as the innocent software child of the dead Thomas Riemann, unfairly burdened with the killer’s guilt. Not unless he redefined himself completely: edited his memories, rewrote his personality. Sculpted his mind into someone new.

In other words: died.

That was the choice. He had to live with what he was in its entirety, or create another person who’d inherit only part of what he’d been.

He laughed angrily and shook his head. “I’m not passing through the eye of any needle. I killed Anna. I killed Anna. That’s who I am.” He reached for the scar which defined him, and stroked it as if it were a talisman.

He sat for a while longer, reliving the night in Hamburg one more time, weeping with shame at what he’d done.

Then he unlocked the drinks cabinet and proceeded to make himself confident and optimistic. The ritual had been pointless—but if nothing else, it had rid him of the delusion that it might have been otherwise.

Some time later, he thought about the clone. Drifting into narcosis. Suffering a crudely modeled extrapolation of the disease which had killed the original. And then, at the moment of simulated death, taking on a new body, young and healthy—with a face plucked from a photograph from Christmas, 1985.

Resurrection—for an instant. No more than a formality. The script had frozen the young murderer, without even waking him.

And then?

Thomas was too far gone to agonize about it. He’d done what he’d done for the sake of the ritual. He’d delivered the clone into Durham’s hands, to grant it—like the flesh-and-blood it believed itself to be—the remote chance of another life, in a world beyond death, unknowable.

And if the whole thing had been a mistake, there was no way, now, to undo it.

Загрузка...