본문 바로가기
  • 책상 밖 세상을 경험할 수 있는 Playground를 제공하고, 수동적 학습에서 창조의 삶으로의 전환을 위한 새로운 라이프 스타일을 제시합니다.

전체 글391

[2025-2] 김정운 - CornerNet, CenterNet CornerNet: Detecting Objects as Paired Keypoints : https://arxiv.org/abs/1808.01244 CornerNet: Detecting Objects as Paired KeypointsWe propose CornerNet, a new approach to object detection where we detect an object bounding box as a pair of keypoints, the top-left corner and the bottom-right corner, using a single convolution neural network. By detecting objects as paired keypoints, wearxiv.org .. 2026. 1. 3.
[2025-2] 최민서 - SimPO: Simple Preference Optimization with a Reference-Free Reward [논문링크] https://arxiv.org/abs/2405.14734 SimPO: Simple Preference Optimization with a Reference-Free RewardDirect Preference Optimization (DPO) is a widely used offline preference optimization algorithm that reparameterizes reward functions in reinforcement learning from human feedback (RLHF) to enhance simplicity and training stability. In this work, we proposarxiv.org DPO에 대해 잘 모른다면 논문을 이해하는데 힘.. 2025. 12. 31.
[2025-2] 백승우 - MAS-Bench: A Unified Benchmark for Shortcut-Augmented Hybrid Mobile GUI Agents MAS-Bench: A Unified Benchmark for Shortcut-Augmented Hybrid Mobile GUI AgentsTo enhance the efficiency of GUI agents on various platforms like smartphones and computers, a hybrid paradigm that combines flexible GUI operations with efficient shortcuts (e.g., API, deep links) is emerging as a promising direction. However, a frameworkarxiv.org 2025. 12. 24.
[2025-2] 박제우 - Sharpness-Aware Minimization for Efficiently Improving Generalization https://arxiv.org/abs/2010.01412 Sharpness-Aware Minimization for Efficiently Improving GeneralizationIn today's heavily overparameterized models, the value of the training loss provides few guarantees on model generalization ability. Indeed, optimizing only the training loss value, as is commonly done, can easily lead to suboptimal model quality. Motivatearxiv.org Abstract현대 딥러닝 모델은 대부분 Overpar.. 2025. 12. 20.