Jörg Tiedemann gave a talk at the Learning Machines Seminar 2022 in Sweden on “Translations as semantic mirrors – Representation learning with multilingual data”. The talk is available online.
Our new project on High-Performance Language Technologies (HPLT) has started and we will scale data sets, language models and neural MT to a new level. In relation to that, the language technology group in Helsinki has also been selected for one of the first Finnish extreme scale projects on the supercomputer LUMI.
Our project there will be called LumiNMT and the goal of the project is to train neural machine translation models on a large scale using state-of-the-art transformer models and novel modular multilingual setups. Our project will focus on increasing language coverage and efficient use of massively parallel data sets. Our research group wants to use LUMI’s extensive parallel computing capabilities to reduce training time and scale up a model size.