Wsi helps bring neural computers a step closer Neural networks have been intensely researched during the past few years and they are set to provide a new direction for computing in the 1990s. At a recent technology exhibition in London, engineers from Hitachi's central research laboratories showed that by using wafer scale integration (wsi) for the implementation they have produced a system they believe is of unprecedented size and speed. The system is based on the wsi of digital neural network circuits with up to 144 neurons on each wafer. The wafers are of 5in diameter with the neurons and their interconnections, or synapses, being created using gate array methodology and 0.8um technology. Eight wafers provide up to 1152 neurons. Neural networks are modelled on the form of the brain. Each neuron produces an output as a threshold function of the sum of its weighted inputs. That output is then transmitted to the next 'layer' of neurons, where it is in turn weighted and summed with other neural outputs. The pattern of connection between the input layer and the output layer and the weights assigned to each connection determine the result produced from a given input. The ability to alter weights in response to feedback allows a network to be trained to generate an appropriate answer to a particular problem. Neural networks have shown promise in solving problems of optimisation and recognition. But the inherent parallelism of neural networks is not taken advantage of when run as software simulations on conventional serial processing computers. This results in much slower operations. In hardware, researchers are pursuing analogue and digital, and optical or electronic networks. The most general network architecture is the Hopfield network, in which each neuron - neurons can be disconnected by setting the synaptic weight to zero. But where the synapse connections are permanent and discrete their number rises with the square of the number of neurons. Each neuron must be able to store a synaptic weight for connections from every neuron but is obviously unwieldy. The solution to this chosen by Hitachi is a time multiplexed digital bus. Hitachi's engineers describe the system as a neurocomputer, and envision it running as an accelerator or coprocessor for a conventional computer or workstation host. The company hopes to bring such a system to market within two years. The engineers responsible for the neurocomputer feel that their combination of digital neural networks and wsi has, for the first time, produced a network of useful complexity and speed. They say that moves to cell based layout and more aggressive design rules could get them up to 1000 neurons per wafer quickly. There is also scope to improve the efficiency of gate utilisation from the currently required 2000 per neuron to 1000 per neuron. The system is contained in a frame measuring 12 x 8.3 x 9in. Operating from a 5v supply, each wafer dissipates 5W and the system as a whole approximately 50W with forced air cooling used to remove heat. The neurocomputer is digital in operation with each neuron producing 1 9bit result. Up to 64 synapse weights of 8 bits are stored in local sram at each neuron. Each wafer contains 60 gate arrays of 100,000 gate complexity. Of these, 48 are used to implement three neurons per gate array along with the associated bus connections and memory. The remaining 12 are used for buffering bus interconnections. To provide a Hopfield network architecture, each neural gate array contains its own global bus which links to two of its nearest neighbours. This bus joins a main bus block at the median line of the wafer, and from there it is taken off-wafer to a backplane which connects all the wafers. The bus is a 9 bit multiplexed bus allowing any neuron to be connected with any other neuron in the system. The bus protocol includes a broadcast facility allowing a neuron to be linked to all others simultaneously. This time multiplexed bus means that only one synapse weighting circuit is required per neuron. Summation is performed by adding the weighted connections on successive cycles. This is much slower than a fully and discreetly connected network but still allows neurons to operate in parallel, providing speed benefits over the software simulation. As for learning performance, Hitachi claims 2.3 billion connection updates per second, and a prototype system has been used to perform signature verification in 2s. Neural networks are well suited to wsi as they are inherently tolerant of individually defective neurons. Such a neuron can be effectively excised from the network by setting all its coefficients to zero. Neural networks are notoriously difficult to control or program but running one from a conventional host computer helps. But don't expect a neurocomputer to cost less than $10,000 for a minimal configuration. If such a configuration has a 1000 artificial neurons, what value does that put on the human brain with about 10 billion neurons?