DiT Annotated Paper
DIT: SELF-SUPERVISED PRE-TRAINING FOR DOCUMENT IMAGE TRANSFORMER
DIT: SELF-SUPERVISED PRE-TRAINING FOR DOCUMENT IMAGE TRANSFORMER
WebFormer: The Web-page Transformer for Structure Information Extraction
LayoutLMv2: Multi-Modal Pre-Training For Visually-Rich Document Understanding
Fastformer: Additive Attention Can Be All You Need
LayoutLM: Pre-training of Text and Layout for Document Image Understanding
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Few-Shot Named Entity Recognition: A Comprehensive Study
EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks
EfficientNetV2: Smaller Models and Faster Training
Fine-Tune BERT for Text Classification with TensorFlow
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
MLP-MIXER: An all MLP Architecture for Vision
PICK: Processing Key Information Extraction from Documents using Improved Graph Learning-Convolutional Networks
Attention Is All You Need