This article introduces a variation of differential neural computer (DNC) architecture with a convertible short-term and long-term memory, named CSLM-DNC.
External memory-based neural networks, such as differentiable neural computers (DNCs), have recently gained importance and popularity to solve complex sequential learning tasks that pose challenges to conventional neural networks; however, a trained DNC usually has a low-memory utilization efficiency. Unlike the memory architecture of the original DNC, the new scheme of short-term and long-term memories offers different importance of memory locations for read and write, and they can be converted over time. This is mainly motivated by the human brain, where short-term memory stores large amounts of noisy and unimportant information and decays rapidly, while long-term memory stores important information and lasts for a long time. The conversion of these two types of memory is allowed and is able to be learned according to their reading and writing frequency. The authors quantitatively and qualitatively evaluate the proposed CSLM-DNC architecture on the tasks of question answering, copy and repeat copy, showing that it can significantly improve memory efficiency and learning performance. (Publisher abstract provided)
Downloads
Related Datasets
Similar Publications
- An Attractive Target: Do Perceptions of Physical Attractiveness Shape Victimization Risks in Women's Prisons?
- The Panacea That Is Love: The Influence of Romantic Relationships on Justice-Involved Young Adults' Self-Reported Health
- Enhancing Fault Ride-Through Capacity of DFIG-Based WPs by Adaptive Backstepping Command Using Parametric Estimation in Non-Linear Forward Power Controller Design