Projects

assorted schoolwork, personal projects, independent research


Exposé: Neural Network Learning Dynamics Inhabit Three Phases, from Linear-Dynamical to Binary Matrix Graph, 2022.
Summary: A new analysis of learning dynamics of sigmoidally-activated neural networks identifies a linearity of young networks training and a graph structure of experienced networks. In addition to providing a new perspective on understanding networks and starting point for new questions and inquiries about deep learning, this analysis finds three quantifiable phases of learning, identifies ways to shortcut learning in young network, and provides insight into the deep learning success of the RELU nonlinearity.


Précis: A Cellular Automaton Model of Covid-19 Epidemiology, 2020, a side project that identifies a striking similarity between cellular automata and epidemic infectivity of diseases like Covid-19. Surprisingly, I haven't found any body of literature on this connection. Since cellular automata have a number of well-catalogued interesting properties, there might be insight to be gained about epidemics learned by applying and studying cellular automata models.
Go figure - cellular automata might be useful for something afterall! :)
(todo: move this to blog page)


Précis: Connectionism, Neural Networks, and Computational Neuroscience, ongoing research project, 2020. Motivated by my other work at the intersection of deep learning and cognitive science, I sketch a novel problem and theoretically-driven innovation towards biologically-plausible neural network technology. Part of a series of brief writeups outlining my current research directions. Here I state an associated mathematical problem. More assorted thoughts.


Expressiveness as Decomposable Default-VC, 2017, course project for CS 257 / Phil 356C: Logic and Artificial Intelligence. This project explored the application of philosophical probabilistic logic to statistical learning theory and model selection.


Tensors in the Petri Dish: Evolutionary Approaches to Designing and Training Neural Networks, 2017, course project with J. Beal for the course AP 293: Theoretical Neuroscience, taught by Surya Ganguli. In this survey, we review the current and recent work on evolutionary topics in deep learning, the recent advancement of neuroevolution, and perspectives on the rise of meta-learning.


Classification of Accents of English Speakers by Native Language, 2014, with A. Chow and S. Li, course project for CS 229: Machine Learning, taught by Andrew Ng. We analyzed the ability for machine learning classifers to determine whether an individual was a native speaker of English.


Offline Signature Verification with Convolutional Neural Networks, 2017, course project with G. Alvarez and B. Sheffer for the course CS 231N: Convolutional Neural Networks for Visual Recognition, taught be Andrej Karpathy and Fei-Fei Li. We build a model that uses deep learning to identify when a signature has been forged. This project has gathered industry attention for applications.


Discovery of Hidden Variables Using Probabilistic Methods, 2016, course project for Psych 204: Computation and Cognition: the Probabilistic Approach, with J. Cho. We discuss problems resulting from Bell number-sized hypothesis classes, and using Bayesian agent modeling and data analysis techniques, we provide an empirical method for generating, sampling, testing, and choosing hypothesis models.


Neural network attractors and dynamical systems, 2021. A project made of a few parts as of now. I'm asking questions about attractors of recurrent Hopfield/Boltzmann neural networks. Attractors serve as state systems and have been used to model associative memory and short-term memory of the cortex. So I'm curious about the shape and distribution of attractors to further understand these networks' computational complexity and expressiveness. To this end, first I employ standard theory of dynamical systems, stability theory, eigenspectra, and random matrix models of biologically-plausible networks. [ next steps are making distributed analogues, relate them to logical circuits & automata & LSTMs, and quantify the expressiveness per resources. ]

This is a hypothetical grant proposal that outlines, directly and succintly, the problem I'm working on, why I care about it, and how I'm going about it. Definitely draft material, but should be halfway legible to people who aren't me. I think I'm going to start labeling this project ESORMA, EigenSpectra Of Random Matrices / Attractors. This doesn't actually make perfect sense and is redundant, but it is helpful for getting the point across roughly and is a nice bite.

WIP Précis, 2021. This is a very rough document draft of new project I'm working for my own reference. Feel free to take a look, but it's very unedited thoughtdump/wordbarf!

Here is a github repo of computational analyses of (1) spectra of various random matrices and (2) finding the zeros of an equation related to spin-glasses, to attempt to gain an intuition around the classic result that attractor networks with N neurons can hold up to 0.138N patterns.


Multiple currencies to implement policy, including simultaneous policies towards universal basic income or divestment, via dynamic programming and principal components, to replace tax policy

This website!

Alpha-One (todo) - a theoretically-grounded projected in 2017-8 attempted to scale up experience replay to allow for experience 'time travel', but it was stopped by a Tensorflow missing feature that explicitly prevented storing of Heroku Go states.

Rho thoughts

A copy of my recent behavioral psychology experiment about spatial cognition. (I would not recommend trying to complete it; it takes about fifteen minutes of relatively rote work. A minute or two should give the idea.)