John Hewitt @ NeurIPS Profile picture
CS PhD @stanford with @stanfordnlp. Frmr. @penn, intern @deepmind, @googleai, ++. Understanding and improving neural learning from language. Co-teach CS 224n.
Apr 5, 2019 5 tweets 4 min read
Does my unsupervised neural network learn syntax? In new #NAACL2019 paper with @chrmanning, our "structural probe" can show that your word representations embed entire parse trees.

paper: nlp.stanford.edu/pubs/hewitt201…
blog: nlp.stanford.edu/~johnhew/struc…
code: github.com/john-hewitt/st…
1/4 @chrmanning Key idea: Vector spaces have distance metrics (L2); trees do too (# edges between words). Vector spaces have norms (L2); rooted trees do too (# edges between word and ROOT.) Our probe finds a vector distance/norm on word representations that matches all tree distances/norms 2/4