On Neurons Invariant to Sentence Structural Changes in Neural Machine Translation
Patel, Gal, Choshen, Leshem, Abend, Omri
–arXiv.org Artificial Intelligence
We present a methodology that explores how sentence structure is reflected in neural representations of machine translation systems. We demonstrate our model-agnostic approach with the Transformer English-German translation model. We analyze neuron-level correlation of activations between paraphrases while discussing the methodology challenges and the need for confound analysis to isolate the effects of shallow cues. We find that similarity between activation patterns can be mostly accounted for by similarity in word choice and sentence length. Following that, we manipulate neuron activations to control the syntactic form of the output. We show this intervention to be somewhat successful, indicating that deep models capture sentence-structure distinctions, despite finding no such indication at the neuron level. To conduct our experiments, we develop a semi-automatic method to generate meaning-preserving minimal pair paraphrases (active-passive voice and adverbial clause-noun phrase) and compile a corpus of such pairs.
arXiv.org Artificial Intelligence
Nov-2-2022
- Country:
- Asia
- China
- Middle East > Israel
- Jerusalem District > Jerusalem (0.04)
- Europe
- Belgium > Brussels-Capital Region
- Brussels (0.04)
- Ireland > Leinster
- County Dublin > Dublin (0.04)
- Italy > Tuscany
- Florence (0.04)
- Pisa Province > Pisa (0.04)
- Belgium > Brussels-Capital Region
- North America > United States
- Georgia > Fulton County
- Atlanta (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- New York > New York County
- New York City (0.04)
- Georgia > Fulton County
- Oceania > Australia
- Asia
- Genre:
- Research Report (0.82)
- Technology: