ALIbaba’s Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab

AliceMind: ALIbaba’s Collection of Encoder-decoders from MinD (Machine IntelligeNce of Damo) Lab
This repository provides pre-trained encoder-decoder models and its related optimization techniques developed by Alibaba’s MinD (Machine IntelligeNce of Damo) Lab.
The family of AliceMind:
- Language understanding model: StructBERT (
ICLR 2020
) - Generative language model: PALM (
EMNLP 2020
) - Cross-lingual language model: VECO (
ACL 2021
) - Cross-modal language model: StructVBERT (
CVPR 2020 VQA Challenge Runner-up
) - Structural language model: StructuralLM (
ACL 2021
) - Chinese language understanding model with multi-granularity inputs: LatticeBERT (
NAACL 2021
) - Pre-training table model: SDCUP (
Under Review
)
News
- March, 2021: AliceMind released!
- May, 2021: VECO and StructuralLM were accepted by ACL 2021.
- September, 2021: The