xzbin
首页
文章
标签
分类
关于
0%
论文阅读
分类
2023
04-06
GLM: General Language Model Pretraining
03-30
GLM-130B 模型环境搭建流程
03-02
Image as a Foreign Language: BEIT Pretraining for All Vision and Vision-Language Tasks
03-01
BEIT: BERT Pre-Training of Image Transformers
03-01
ZERO and R2D2: A Large-scale Chinese Cross-modal Benchmark and a Vision-Language Framework
02-28
VLMO: Unified Vision-Language Pre-Training with Mixture-of-Modality-Experts
02-24
Pixel-BERT: Aligning Image Pixels with Text by Deep Multi-Modal Transformers
02-24
UNITER: learning universal image-text representations
02-24
an image is worth 16x16 words: transformers for image recognition at scale
02-07
ViLT Vision-and-Language Transformer Without Convolution or Region Supervision
1
2