본문 바로가기
  • 책상 밖 세상을 경험할 수 있는 Playground를 제공하고, 수동적 학습에서 창조의 삶으로의 전환을 위한 새로운 라이프 스타일을 제시합니다.

전체 글78

[2024-2] 백승우 - Retrieval-Augmented Generation for Large Language Models: A Survey Retrieval-Augmented Generation for Large Language Models: A SurveyLarge Language Models (LLMs) showcase impressive capabilities but encounter challenges like hallucination, outdated knowledge, and non-transparent, untraceable reasoning processes. Retrieval-Augmented Generation (RAG) has emerged as a promising solution byarxiv.org0. AbstractLLM(Large Language Model)은 뛰어난 성과를 보이지만, hallucination, .. 2024. 11. 11.
[2024-2] 백승우 - VoxelMorph: A Learning Framework for Deformable Medical Image Registration VoxelMorph: A Learning Framework for Deformable Medical Image RegistrationWe present VoxelMorph, a fast learning-based framework for deformable, pairwise medical image registration. Traditional registration methods optimize an objective function for each pair of images, which can be time-consuming for large datasets or rich defoarxiv.org0. AbstractVoxelMorph은 변형 가능한 pair별 의료 image registration을 .. 2024. 11. 7.
[2024-2] 백승우 - (RAG) Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks Retrieval-Augmented Generation for Knowledge-Intensive NLP TasksLarge pre-trained language models have been shown to store factual knowledge in their parameters, and achieve state-of-the-art results when fine-tuned on downstream NLP tasks. However, their ability to access and precisely manipulate knowledge is still limarxiv.org0. AbstractPretrained LLM은 사실의 지식을 매개변수에 저장하고, downstream NLP 작업에서 미.. 2024. 11. 3.
[2024-2] 김경훈 - VoxelMorph : A Learning Framework for Deformable Medical Image Registration Link : https://arxiv.org/abs/1809.05231 VoxelMorph: A Learning Framework for Deformable Medical Image RegistrationWe present VoxelMorph, a fast learning-based framework for deformable, pairwise medical image registration. Traditional registration methods optimize an objective function for each pair of images, which can be time-consuming for large datasets or rich defoarxiv.org      0. Abstract 기.. 2024. 9. 10.