ALIbaba’s Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab
AliceMind: ALIbaba’s Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba’s MinD (Machine IntelligeNce of Damo) Lab. The family of AliceMind: Language understanding model: StructBERT (ICLR 2020) Generative language model: PALM (EMNLP 2020) Cross-lingual language model: VECO (ACL 2021) Cross-modal language model: StructVBERT (CVPR 2020 VQA Challenge Runner-up) Structural language model: StructuralLM (ACL 2021) Chinese language understanding model with multi-granularity inputs: LatticeBERT (NAACL 2021) Pre-training […]
Read more