Just in time to make 2018 a bit less unproductive, I managed to wake in the middle of the night with another few inventions. I’m finishing the year on only a third as many as 2016 and 2017, but better than some years. And I quite like these new ones.
Gel computing is a very old idea of mine, and I’m surprised no company has started doing it yet. Air gel is different. My original used a suspension of processing particles in gel, and the idea was that the gel would hold the particles in fixed locations with good free line of sight to neighbor devices for inter-device optical comms, while acting also as a coolant.
Air gel uses the same idea of suspending particles, but does so by using ultrasound, standing waves holding the particles aloft. They would form a semi-gel I suppose, much softer. The intention is that they will be more easily movable than in a gel, and maybe rotate. I imagine using rotating magnetic fields to rotate them, and use that mechanism to implement different configurations of inter-device nets. That would be the first pillar of running multiple neural nets in the same space at the same time, using spin-based TDM (time division multiplexing), or synchronized space multiplexing if you prefer. If a device uses on board processing that is fast compared to the signal transmission time to other devices (the speed of light may be fast but can still be severely limiting for processing and comms), then having the ability to deal with processing associated with several other networks while awaiting a response allows a processing network to be multiplied up several times. A neural net could become a hyper-neural net.
Given that this is intended for mid-century AI, I’m also making the assumption that true TDM can also be used on each net, my second pillar. Signals would carry a stream of slots holding bits for each processing instance. Since this allows a Turing machine to implement many different processes in parallel, I decided to call it Turing multiplexing. Again, it helps alleviate the potential gulf between processing and communication times. Combining Turing and spin multiplexing would allow a single neural net to be multiplied up potentially thousands or millions of times – hyper-neurons seems as good a term as any.
The third pillar of this system is that the processing particles (each could contain a large number of neurons or other IT objects) could be energized and clocked using very high speed alternating EM fields – radio, microwaves, light, even x-rays. I don’t have any suggestions for processing mechanisms that might operate at such frequencies, though Pauli switches might work at lower speeds, using Pauli exclusion principle to link electron spin states to make switches. I believe early versions of spin cubits use a similar principle. I’m agnostic whether conventional Turing machine or quantum processing would be used, or any combination. In any case, it isn’t my problem, I suspect that future AIs will figure out the physics and invent the appropriate IT.
Processing devices operating at high speed could use a lot of energy and generate a lot of heat, and encouraging the system to lase by design would be a good way to cool it as well as powering it.
A processor using such mechanisms need not be bulky. I always assumed a yogurt pot size for my gel computer before and an air gel processor could be the same, about 100ml. That is enough to suspend a trillion particles with good line of sight for optical interconnections, and each connection could utilise up to millions of alternative wavelengths. Each wavelength could support many TDM channels and spinning the particles multiplies that up again. A UV laser clock/power source driving processors at 10^16Hz would certainly need to use high density multiplexing to make use of such a volume, with transmission distances up to 10cm (but most sub-mm) otherwise being a strongly limiting performance factor, but 10 million-fold WDM/TDM is attainable.
A trillion of these hyper-neurons using that multiplexing would act very effectively as 10 million trillion neurons, each operating at 10^16Hz processing speed. That’s quite a lot of zeros, 35 of them, and yet each hyperneuron could have connections to thousands of others in each of many physical configurations. It would be an obvious platform for supporting a large population of electronically immortal people and AIs who each want a billion replicas, and if it only occupies 100ml of space, the environmental footprint isn’t an issue.
It’s hard to know how to talk to a computer that operates like a brain, but is 10^22 times faster, but I’d suggest ‘Yes Boss’.