Continual Learning of Dense Signals

Placeholder Show Content

Abstract/Contents

Abstract
Coordinate networks are neural networks that parameterize low-dimensional signals like images and signed distance functions. They fail to sequentially learn on a stream of signals as they catastrophically forget the earlier part of the stream. This paper proposes a method to mitigate forgetting based on a hash encoding coordinate network. Specifically, the hash table is augmented with a cache data structure along with additional regularizations. Experiments show that this method eliminates forgetting when the network learns dense signals for different data collection patterns, approaching the empirical upper bound with little margin.

Description

Type of resource text
Publication date July 20, 2023

Creators/Contributors

Author Dong, Zhengyang
Thesis advisor Wetzstein, Gordon

Subjects

Subject neural network
Subject coordinate network
Subject signal representation
Subject continual learning
Subject catastrophic forgetting
Subject computational imaging
Genre Text
Genre Article
Genre Thesis

Bibliographic information

Access conditions

Use and reproduction
User agrees that, where applicable, content will not be used to identify or to otherwise infringe the privacy or confidentiality rights of individuals. Content distributed via the Stanford Digital Repository may be subject to additional license and use restrictions applied by the depositor.
License
This work is licensed under a Creative Commons Attribution Non Commercial 4.0 International license (CC BY-NC).

Preferred citation

Preferred citation
Dong, Z. (2023). Continual Learning of Dense Signals. Stanford Digital Repository. Available at https://purl.stanford.edu/rw194xv5100. https://doi.org/10.25740/rw194xv5100.

Collection

Undergraduate Theses, School of Engineering

View other items in this collection in SearchWorks

Contact information

Also listed in

Loading usage metrics...