Memcomputing: a brain-inspired computing paradigm to store and process information on the same physical platform
Massimiliano Di Ventra
Department of Physics, University of California, San Diego, La Jolla, USA

Thu., Oct. 2, 2014, 1 p.m.


I will discuss a novel computing paradigm we named memcomputing inspired by the operation of our own brain. Memcomputing - computing using memory circuit elements or memelements - satisfies important physical requirements: (i) it is intrinsically massively parallel, (ii) its information storing and computing units are physically the same, and (iii) it does not rely on active elements as the main tools of operation. I will then introduce the notion of universal memcomputing machines (UMMs) as a class of general-purpose computing machines based on systems with memory. We have shown that the memory properties of UMMs endow them with universal computing power- they are Turing-complete, intrinsic parallelism, functional polymorphism, and information overhead, namely their collective states can support exponential data compression directly in memory. We have proved that UMMs can solve NP-complete problems in polynomial time, and as an example I will provide the polynomial-time solution of the subset-sum problem when implemented in hardware. Even though we have not proved NP=P, the practical implementation of these UMMs would represent a paradigm shift from present von Neumann architectures bringing us closer to brain-like neural computation. In fact, I will discuss a practical CMOS-compatible realization of this computing paradigm that uses memcapacitors and we have named Dynamic Computing Random Access Memory (DCRAM).



Share
Memcomputing: a brain-inspired computing paradigm to store and process information on the same physical platform
Massimiliano Di Ventra
Department of Physics, University of California, San Diego, La Jolla, USA

Thu., Oct. 2, 2014, 1 p.m.


I will discuss a novel computing paradigm we named memcomputing inspired by the operation of our own brain. Memcomputing - computing using memory circuit elements or memelements - satisfies important physical requirements: (i) it is intrinsically massively parallel, (ii) its information storing and computing units are physically the same, and (iii) it does not rely on active elements as the main tools of operation. I will then introduce the notion of universal memcomputing machines (UMMs) as a class of general-purpose computing machines based on systems with memory. We have shown that the memory properties of UMMs endow them with universal computing power- they are Turing-complete, intrinsic parallelism, functional polymorphism, and information overhead, namely their collective states can support exponential data compression directly in memory. We have proved that UMMs can solve NP-complete problems in polynomial time, and as an example I will provide the polynomial-time solution of the subset-sum problem when implemented in hardware. Even though we have not proved NP=P, the practical implementation of these UMMs would represent a paradigm shift from present von Neumann architectures bringing us closer to brain-like neural computation. In fact, I will discuss a practical CMOS-compatible realization of this computing paradigm that uses memcapacitors and we have named Dynamic Computing Random Access Memory (DCRAM).



Share