Can we find subnetworks that compute high-level intermediate variables in transformers?
We present a python package that simplifies subnetwork analysis.
Do neural networks self-organize into modular components when solving compositional tasks?
Can vision models learn same-different relations that generalize to different datasets?
Can we influence the solutions that neural networks learn by transferring subnetworks from trained models to randomly intialized models?
Which tree-structured neural network imparts the best inductive bias for syntactic agreement?
How might researchers make fair comparisons between human and machine perception systems?
We know BERT encodes racial and gender biases. Does it also encode intersectional biases?
Can we use representational similarity analysis to uncover linguistic structure in word embeddings?