Translation into morphologically-rich languages challenges neural machine translation (NMT) models with extremely sparse vocabularies
where atomic treatment of surface forms is unrealistic. This problem is typically addressed by either pre-processing words
into subword units or performing translation directly at the level of characters. The former is based on word segmentation
algorithms optimized using corpus-level statistics with no regard to the translation task. The latter learns directly from
translation data but requires rather deep architectures. In this paper, we propose to translate words by modeling word formation
through a hierarchical latent variable model which mimics the process of morphological inflection. Our model generates words
one character at a time by composing two latent representations: a continuous one, aimed at capturing the lexical semantics,
and a set of (approximately) discrete features, aimed at capturing the morphosyntactic function, which are shared among different
surface forms. Our model achieves better accuracy in translation into three morphologically-rich languages than conventional
open-vocabulary NMT methods, while also demonstrating a better generalization capacity under low to mid-resource settings.