We investigate the transformer’s capability to simulate the training process of deep models via in-context learning (ICL), i.e., in-context deep learning. Our key contribution is providing a positive example of using...
详细信息
We investigate the approximation and estimation rates of conditional diffusion transformers (DiTs) with classifier-free guidance. We present a comprehensive analysis for "in-context" conditional DiTs under f...
We investigate the statistical and computational limits of prompt tuning for transformer-based foundationmodels. Our key contributions are prompt tuning on single-head transformers with only a single self-attention l...
详细信息
暂无评论