본문 바로가기
반응형

분류 전체보기599

[2022-12-10] 오늘의 자연어처리 Robust Speech Recognition via Large-Scale Weak Supervision We study the capabilities of speech processing systems trained simply to predict large amounts of transcripts of audio on the internet. When scaled to 680,000 hours of multilingual and multitask supervision, the resulting models generalize well to standard benchmarks and are often competitive with prior fully supervised results but in a .. 2022. 12. 10.
[2022-12-09] 오늘의 자연어처리 Counterfactual reasoning: Do language models need world knowledge for causal understanding? Current pre-trained language models have enabled remarkable improvements in downstream tasks, but it remains difficult to distinguish effects of statistical correlation from more systematic logical reasoning grounded on understanding of the real world. In this paper we tease these factors apart by leverag.. 2022. 12. 9.
[2022-12-08] 오늘의 자연어처리 Semantic-Conditional Diffusion Networks for Image Captioning Recent advances on text-to-image generation have witnessed the rise of diffusion models which act as powerful generative models. Nevertheless, it is not trivial to exploit such latent variable models to capture the dependency among discrete words and meanwhile pursue complex visual-language alignment in image captioning. In this paper,.. 2022. 12. 8.
[2022-12-07] 오늘의 자연어처리 Retrieval as Attention: End-to-end Learning of Retrieval and Reading within a Single Transformer Systems for knowledge-intensive tasks such as open-domain question answering (QA) usually consist of two stages: efficient retrieval of relevant documents from a large corpus and detailed reading of the selected documents to generate answers. Retrievers and readers are usually modeled separately, whi.. 2022. 12. 7.
반응형