본문 바로가기
반응형

분류 전체보기599

[2023-04-21] 오늘의 자연어처리 Is ChatGPT Good at Search? Investigating Large Language Models as Re-Ranking Agent Large Language Models (LLMs) have demonstrated a remarkable ability to generalize zero-shot to various language-related tasks. This paper focuses on the study of exploring generative LLMs such as ChatGPT and GPT-4 for relevance ranking in Information Retrieval (IR). Surprisingly, our experiments reveal that proper.. 2023. 4. 21.
[2023-04-20] 오늘의 자연어처리 Romanization-based Large-scale Adaptation of Multilingual Language Models Large multilingual pretrained language models (mPLMs) have become the de facto state of the art for cross-lingual transfer in NLP. However, their large-scale deployment to many languages, besides pretraining data scarcity, is also hindered by the increase in vocabulary size and limitations in their parameter budget. In ord.. 2023. 4. 20.
[2023-04-19] 오늘의 자연어처리 LongForm: Optimizing Instruction Tuning for Long Text Generation with Corpus Extraction Instruction tuning enables language models to generalize more effectively and better follow user intent. However, obtaining instruction data can be costly and challenging. Prior works employ methods such as expensive human annotation, crowd-sourced datasets with alignment issues, or generating noisy examples .. 2023. 4. 19.
[2023-04-18] 오늘의 자연어처리 Research without Re-search: Maximal Update Parametrization Yields Accurate Loss Prediction across Scales As language models scale up, it becomes increasingly expensive to verify research ideas because conclusions on small models do not trivially transfer to large ones. A possible solution is to establish a generic system that directly predicts some metrics for large models solely based on the re.. 2023. 4. 18.
반응형