这是indexloc提供的服务,不要输入任何密码
Skip to main content

Showing 1–50 of 250 results for author: Choi, E

Searching in archive cs. Search in all archives.
.
  1. arXiv:2507.11407  [pdf, ps, other

    cs.CL cs.AI

    EXAONE 4.0: Unified Large Language Models Integrating Non-reasoning and Reasoning Modes

    Authors: LG AI Research, :, Kyunghoon Bae, Eunbi Choi, Kibong Choi, Stanley Jungkyu Choi, Yemuk Choi, Kyubeen Han, Seokhee Hong, Junwon Hwang, Taewan Hwang, Joonwon Jang, Hyojin Jeon, Kijeong Jeon, Gerrard Jeongwon Jo, Hyunjik Jo, Jiyeon Jung, Euisoon Kim, Hyosang Kim, Jihoon Kim, Joonkee Kim, Seonghwan Kim, Soyeon Kim, Sunkyoung Kim, Yireun Kim , et al. (17 additional authors not shown)

    Abstract: This technical report introduces EXAONE 4.0, which integrates a Non-reasoning mode and a Reasoning mode to achieve both the excellent usability of EXAONE 3.5 and the advanced reasoning abilities of EXAONE Deep. To pave the way for the agentic AI era, EXAONE 4.0 incorporates essential features such as agentic tool use, and its multilingual capabilities are extended to support Spanish in addition to… ▽ More

    Submitted 15 July, 2025; originally announced July 2025.

    Comments: Technical Report, 30 Pages

  2. arXiv:2507.06996  [pdf, ps, other

    cs.LG cs.AI

    Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing

    Authors: Eunbyeol Cho, Jiyoun Kim, Minjae Lee, Sungjin Park, Edward Choi

    Abstract: Electronic Health Records (EHR) are time-series relational databases that record patient interactions and medical events over time, serving as a critical resource for healthcare research and applications. However, privacy concerns and regulatory restrictions limit the sharing and utilization of such sensitive data, necessitating the generation of synthetic EHR datasets. Unlike previous EHR synthes… ▽ More

    Submitted 9 July, 2025; originally announced July 2025.

  3. arXiv:2507.05228  [pdf, ps, other

    cs.LG cs.CR

    Cascade: Token-Sharded Private LLM Inference

    Authors: Rahul Thomas, Louai Zahran, Erica Choi, Akilesh Potti, Micah Goldblum, Arka Pal

    Abstract: As LLMs continue to increase in parameter size, the computational resources required to run them are available to fewer parties. Therefore, third-party inference services -- where LLMs are hosted by third parties with significant computational resources -- are becoming increasingly popular. However, third party inference raises critical concerns about user data privacy. To mitigate these risks, pr… ▽ More

    Submitted 7 July, 2025; originally announced July 2025.

    Comments: To be published in ICML 2025 Main Proceedings as "Hidden No More: Attacking and Defending Private Third-Party LLM Inference", together with arXiv:2505.18332

  4. arXiv:2506.23518  [pdf, ps, other

    cs.CV

    WAVE: Warp-Based View Guidance for Consistent Novel View Synthesis Using a Single Image

    Authors: Jiwoo Park, Tae Eun Choi, Youngjun Jun, Seong Jae Hwang

    Abstract: Generating high-quality novel views of a scene from a single image requires maintaining structural coherence across different views, referred to as view consistency. While diffusion models have driven advancements in novel view synthesis, they still struggle to preserve spatial continuity across views. Diffusion models have been combined with 3D models to address the issue, but such approaches lac… ▽ More

    Submitted 30 June, 2025; originally announced June 2025.

  5. arXiv:2506.17363  [pdf, ps, other

    cs.CY cs.AI

    A Large-Scale Real-World Evaluation of LLM-Based Virtual Teaching Assistant

    Authors: Sunjun Kweon, Sooyohn Nam, Hyunseung Lim, Hwajung Hong, Edward Choi

    Abstract: Virtual Teaching Assistants (VTAs) powered by Large Language Models (LLMs) have the potential to enhance student learning by providing instant feedback and facilitating multi-turn interactions. However, empirical studies on their effectiveness and acceptance in real-world classrooms are limited, leaving their practical impact uncertain. In this study, we develop an LLM-based VTA and deploy it in a… ▽ More

    Submitted 20 June, 2025; originally announced June 2025.

    Comments: ACL 2025 Industry Track

  6. arXiv:2506.11815  [pdf, ps, other

    eess.SP cs.AI cs.LG eess.IV

    Diffusion-Based Electrocardiography Noise Quantification via Anomaly Detection

    Authors: Tae-Seong Han, Jae-Wook Heo, Hakseung Kim, Cheol-Hui Lee, Hyub Huh, Eue-Keun Choi, Hye Jin Kim, Dong-Joo Kim

    Abstract: Electrocardiography (ECG) signals are frequently degraded by noise, limiting their clinical reliability in both conventional and wearable settings. Existing methods for addressing ECG noise, relying on artifact classification or denoising, are constrained by annotation inconsistencies and poor generalizability. Here, we address these limitations by reframing ECG noise quantification as an anomaly… ▽ More

    Submitted 22 July, 2025; v1 submitted 13 June, 2025; originally announced June 2025.

    Comments: This manuscript contains 17 pages, 10 figures, and 3 tables

  7. arXiv:2506.09014  [pdf, ps, other

    cs.CL

    Learning to Reason Across Parallel Samples for LLM Reasoning

    Authors: Jianing Qi, Xi Ye, Hao Tang, Zhigang Zhu, Eunsol Choi

    Abstract: Scaling test-time compute brings substantial performance gains for large language models (LLMs). By sampling multiple answers and heuristically aggregate their answers (e.g., either through majority voting or using verifiers to rank the answers), one can achieve consistent performance gains in math domains. In this paper, we propose a new way to leverage such multiple sample set. We train a compac… ▽ More

    Submitted 10 June, 2025; originally announced June 2025.

  8. arXiv:2506.08920  [pdf, ps, other

    cs.CL cs.AI cs.LG

    PropMEND: Hypernetworks for Knowledge Propagation in LLMs

    Authors: Zeyu Leo Liu, Greg Durrett, Eunsol Choi

    Abstract: Knowledge editing techniques for large language models (LLMs) can inject knowledge that is later reproducible verbatim, but they fall short on propagating that knowledge: models cannot answer questions that require reasoning with the injected knowledge. We present a hypernetwork-based approach for knowledge propagation, named PropMEND, where we meta-learn how to modify gradients of a language mode… ▽ More

    Submitted 10 June, 2025; originally announced June 2025.

    Comments: Under review

  9. arXiv:2506.08357  [pdf, ps, other

    cs.SD cs.AI eess.AS

    MD-ViSCo: A Unified Model for Multi-Directional Vital Sign Waveform Conversion

    Authors: Franck Meyer, Kyunghoon Hur, Edward Choi

    Abstract: Despite the remarkable progress of deep-learning methods generating a target vital sign waveform from a source vital sign waveform, most existing models are designed exclusively for a specific source-to-target pair. This requires distinct model architectures, optimization procedures, and pre-processing pipelines, resulting in multiple models that hinder usability in clinical settings. To address t… ▽ More

    Submitted 9 June, 2025; originally announced June 2025.

    Comments: Main paper (16 pages, 5 figures). Paper submitted for review. Code available at https://github.com/fr-meyer/MD-ViSCo

  10. arXiv:2506.01673  [pdf, ps, other

    cs.IR cs.AI cs.CL

    GRAM: Generative Recommendation via Semantic-aware Multi-granular Late Fusion

    Authors: Sunkyung Lee, Minjin Choi, Eunseong Choi, Hye-young Kim, Jongwuk Lee

    Abstract: Generative recommendation is an emerging paradigm that leverages the extensive knowledge of large language models by formulating recommendations into a text-to-text generation task. However, existing studies face two key limitations in (i) incorporating implicit item relationships and (ii) utilizing rich yet lengthy item information. To address these challenges, we propose a Generative Recommender… ▽ More

    Submitted 2 June, 2025; originally announced June 2025.

    Comments: ACL 2025 (Main Conference)

  11. arXiv:2505.21190  [pdf, ps, other

    cs.CL cs.AI

    Lunguage: A Benchmark for Structured and Sequential Chest X-ray Interpretation

    Authors: Jong Hak Moon, Geon Choi, Paloma Rabaey, Min Gwan Kim, Hyuk Gi Hong, Jung-Oh Lee, Hangyul Yoon, Eun Woo Doe, Jiyoun Kim, Harshita Sharma, Daniel C. Castro, Javier Alvarez-Valle, Edward Choi

    Abstract: Radiology reports convey detailed clinical observations and capture diagnostic reasoning that evolves over time. However, existing evaluation methods are limited to single-report settings and rely on coarse metrics that fail to capture fine-grained clinical semantics and temporal dependencies. We introduce LUNGUAGE,a benchmark dataset for structured radiology report generation that supports both s… ▽ More

    Submitted 27 May, 2025; originally announced May 2025.

  12. arXiv:2505.20875  [pdf, ps, other

    cs.CL cs.AI

    Trans-EnV: A Framework for Evaluating the Linguistic Robustness of LLMs Against English Varieties

    Authors: Jiyoung Lee, Seungho Kim, Jieun Han, Jun-Min Lee, Kitaek Kim, Alice Oh, Edward Choi

    Abstract: Large Language Models (LLMs) are predominantly evaluated on Standard American English (SAE), often overlooking the diversity of global English varieties. This narrow focus may raise fairness concerns as degraded performance on non-standard varieties can lead to unequal benefits for users worldwide. Therefore, it is critical to extensively evaluate the linguistic robustness of LLMs on multiple non-… ▽ More

    Submitted 4 June, 2025; v1 submitted 27 May, 2025; originally announced May 2025.

    Comments: 27 pages, 6 figures, 16 tables

  13. arXiv:2505.19429  [pdf, ps, other

    cs.CL

    Rhapsody: A Dataset for Highlight Detection in Podcasts

    Authors: Younghan Park, Anuj Diwan, David Harwath, Eunsol Choi

    Abstract: Podcasts have become daily companions for half a billion users. Given the enormous amount of podcast content available, highlights provide a valuable signal that helps viewers get the gist of an episode and decide if they want to invest in listening to it in its entirety. However, identifying highlights automatically is challenging due to the unstructured and long-form nature of the content. We in… ▽ More

    Submitted 25 May, 2025; originally announced May 2025.

  14. arXiv:2505.18332  [pdf, ps, other

    cs.CR cs.LG

    An Attack to Break Permutation-Based Private Third-Party Inference Schemes for LLMs

    Authors: Rahul Thomas, Louai Zahran, Erica Choi, Akilesh Potti, Micah Goldblum, Arka Pal

    Abstract: Recent advances in Large Language Models (LLMs) have led to the widespread adoption of third-party inference services, raising critical privacy concerns. Existing methods of performing private third-party inference, such as Secure Multiparty Computation (SMPC), often rely on cryptographic methods. However, these methods are thousands of times slower than standard unencrypted inference, and fail to… ▽ More

    Submitted 23 May, 2025; originally announced May 2025.

    Comments: To be published in ICML 2025 Main Proceedings as "Hidden No More: Attacking and Defending Private Third-Party LLM Inference"

  15. arXiv:2505.18231  [pdf, ps, other

    cs.LG cs.AI

    NSNQuant: A Double Normalization Approach for Calibration-Free Low-Bit Vector Quantization of KV Cache

    Authors: Donghyun Son, Euntae Choi, Sungjoo Yoo

    Abstract: Large Language Model (LLM) inference is typically memory-intensive, especially when processing large batch sizes and long sequences, due to the large size of key-value (KV) cache. Vector Quantization (VQ) is recently adopted to alleviate this issue, but we find that the existing approach is susceptible to distribution shift due to its reliance on calibration datasets. To address this limitation, w… ▽ More

    Submitted 23 May, 2025; originally announced May 2025.

  16. arXiv:2505.18087  [pdf, ps, other

    cs.CV cs.AI

    CXReasonBench: A Benchmark for Evaluating Structured Diagnostic Reasoning in Chest X-rays

    Authors: Hyungyung Lee, Geon Choi, Jung-Oh Lee, Hangyul Yoon, Hyuk Gi Hong, Edward Choi

    Abstract: Recent progress in Large Vision-Language Models (LVLMs) has enabled promising applications in medical tasks, such as report generation and visual question answering. However, existing benchmarks focus mainly on the final diagnostic answer, offering limited insight into whether models engage in clinically meaningful reasoning. To address this, we present CheXStruct and CXReasonBench, a structured p… ▽ More

    Submitted 23 May, 2025; originally announced May 2025.

  17. arXiv:2505.17818  [pdf, other

    cs.AI cs.CL

    PatientSim: A Persona-Driven Simulator for Realistic Doctor-Patient Interactions

    Authors: Daeun Kyung, Hyunseung Chung, Seongsu Bae, Jiho Kim, Jae Ho Sohn, Taerim Kim, Soo Kyung Kim, Edward Choi

    Abstract: Doctor-patient consultations require multi-turn, context-aware communication tailored to diverse patient personas. Training or evaluating doctor LLMs in such settings requires realistic patient interaction systems. However, existing simulators often fail to reflect the full range of personas seen in clinical practice. To address this, we introduce PatientSim, a patient simulator that generates rea… ▽ More

    Submitted 23 May, 2025; originally announced May 2025.

    Comments: 9 pages for main text, 4 pages for references, 27 pages for supplementary materials

  18. arXiv:2505.14489  [pdf, ps, other

    cs.AI cs.CL

    Reasoning Models Better Express Their Confidence

    Authors: Dongkeun Yoon, Seungone Kim, Sohee Yang, Sunkyoung Kim, Soyeon Kim, Yongil Kim, Eunbi Choi, Yireun Kim, Minjoon Seo

    Abstract: Despite their strengths, large language models (LLMs) often fail to communicate their confidence accurately, making it difficult to assess when they might be wrong and limiting their reliability. In this work, we demonstrate that reasoning models-LLMs that engage in extended chain-of-thought (CoT) reasoning-exhibit superior performance not only in problem-solving but also in accurately expressing… ▽ More

    Submitted 20 May, 2025; originally announced May 2025.

    Comments: Work in progress

  19. arXiv:2505.03810  [pdf, other

    cs.LG cs.AI cs.CL

    Grouped Sequency-arranged Rotation: Optimizing Rotation Transformation for Quantization for Free

    Authors: Euntae Choi, Sumin Song, Woosang Lim, Sungjoo Yoo

    Abstract: Large Language Models (LLMs) face deployment challenges due to high computational costs, and while Post-Training Quantization (PTQ) offers a solution, existing rotation-based methods struggle at very low bit-widths like 2-bit. We introduce a novel, training-free approach to construct an improved rotation matrix, addressing the limitations of current methods. The key contributions include leveragin… ▽ More

    Submitted 2 May, 2025; originally announced May 2025.

    Comments: 7 pages

  20. arXiv:2505.02722  [pdf, other

    cs.AI cs.LG

    Enhancing LLMs' Clinical Reasoning with Real-World Data from a Nationwide Sepsis Registry

    Authors: Junu Kim, Chaeeun Shim, Sungjin Park, Su Yeon Lee, Gee Young Suh, Chae-Man Lim, Seong Jin Choi, Song Mi Moon, Kyoung-Ho Song, Eu Suk Kim, Hong Bin Kim, Sejoong Kim, Chami Im, Dong-Wan Kang, Yong Soo Kim, Hee-Joon Bae, Sung Yoon Lim, Han-Gil Jeong, Edward Choi

    Abstract: Although large language models (LLMs) have demonstrated impressive reasoning capabilities across general domains, their effectiveness in real-world clinical practice remains limited. This is likely due to their insufficient exposure to real-world clinical data during training, as such data is typically not included due to privacy concerns. To address this, we propose enhancing the clinical reasoni… ▽ More

    Submitted 5 May, 2025; originally announced May 2025.

  21. arXiv:2505.00468  [pdf, other

    cs.CE

    Evaluation of Thermal Control Based on Spatial Thermal Comfort with Reconstructed Environmental Data

    Authors: Youngkyu Kim, Byounghyun Yoo, Ji Young Yun, Hyeokmin Lee, Sehyeon Park, Jin Woo Moon, Eun Ji Choi

    Abstract: Achieving thermal comfort while maintaining energy efficiency is a critical objective in building system control. Conventional thermal comfort models, such as the Predicted Mean Vote (PMV), rely on both environmental and personal variables. However, the use of fixed-location sensors limits the ability to capture spatial variability, which reduces the accuracy of occupant-specific comfort estimatio… ▽ More

    Submitted 4 May, 2025; v1 submitted 1 May, 2025; originally announced May 2025.

  22. arXiv:2504.09387  [pdf, other

    cs.CL

    On Language Models' Sensitivity to Suspicious Coincidences

    Authors: Sriram Padmanabhan, Kanishka Misra, Kyle Mahowald, Eunsol Choi

    Abstract: Humans are sensitive to suspicious coincidences when generalizing inductively over data, as they make assumptions as to how the data was sampled. This results in smaller, more specific hypotheses being favored over more general ones. For instance, when provided the set {Austin, Dallas, Houston}, one is more likely to think that this is sampled from "Texas Cities" over "US Cities" even though both… ▽ More

    Submitted 12 April, 2025; originally announced April 2025.

  23. arXiv:2504.01840  [pdf, other

    cs.CL

    LRAGE: Legal Retrieval Augmented Generation Evaluation Tool

    Authors: Minhu Park, Hongseok Oh, Eunkyung Choi, Wonseok Hwang

    Abstract: Recently, building retrieval-augmented generation (RAG) systems to enhance the capability of large language models (LLMs) has become a common practice. Especially in the legal domain, previous judicial decisions play a significant role under the doctrine of stare decisis which emphasizes the importance of making decisions based on (retrieved) prior documents. However, the overall performance of RA… ▽ More

    Submitted 24 April, 2025; v1 submitted 2 April, 2025; originally announced April 2025.

    Comments: 12 pages

  24. arXiv:2504.00698  [pdf

    cs.CL cs.AI cs.LG

    Command A: An Enterprise-Ready Large Language Model

    Authors: Team Cohere, :, Aakanksha, Arash Ahmadian, Marwan Ahmed, Jay Alammar, Milad Alizadeh, Yazeed Alnumay, Sophia Althammer, Arkady Arkhangorodsky, Viraat Aryabumi, Dennis Aumiller, Raphaël Avalos, Zahara Aviv, Sammie Bae, Saurabh Baji, Alexandre Barbet, Max Bartolo, Björn Bebensee, Neeral Beladia, Walter Beller-Morales, Alexandre Bérard, Andrew Berneshawi, Anna Bialas, Phil Blunsom , et al. (205 additional authors not shown)

    Abstract: In this report we describe the development of Command A, a powerful large language model purpose-built to excel at real-world enterprise use cases. Command A is an agent-optimised and multilingual-capable model, with support for 23 languages of global business, and a novel hybrid architecture balancing efficiency with top of the range performance. It offers best-in-class Retrieval Augmented Genera… ▽ More

    Submitted 14 April, 2025; v1 submitted 1 April, 2025; originally announced April 2025.

    Comments: 55 pages

  25. arXiv:2503.23228  [pdf, other

    eess.SY cs.RO

    Energy-Aware Lane Planning for Connected Electric Vehicles in Urban Traffic: Design and Vehicle-in-the-Loop Validation

    Authors: Hansung Kim, Eric Yongkeun Choi, Eunhyek Joa, Hotae Lee, Linda Lim, Scott Moura, Francesco Borrelli

    Abstract: Urban driving with connected and automated vehicles (CAVs) offers potential for energy savings, yet most eco-driving strategies focus solely on longitudinal speed control within a single lane. This neglects the significant impact of lateral decisions, such as lane changes, on overall energy efficiency, especially in environments with traffic signals and heterogeneous traffic flow. To address this… ▽ More

    Submitted 29 March, 2025; originally announced March 2025.

    Comments: Submitted to an Invited Session at 2025 IEEE Conference on Decision and Control

  26. arXiv:2503.12524  [pdf, other

    cs.CL cs.AI

    EXAONE Deep: Reasoning Enhanced Language Models

    Authors: LG AI Research, Kyunghoon Bae, Eunbi Choi, Kibong Choi, Stanley Jungkyu Choi, Yemuk Choi, Seokhee Hong, Junwon Hwang, Hyojin Jeon, Kijeong Jeon, Gerrard Jeongwon Jo, Hyunjik Jo, Jiyeon Jung, Hyosang Kim, Joonkee Kim, Seonghwan Kim, Soyeon Kim, Sunkyoung Kim, Yireun Kim, Yongil Kim, Youchul Kim, Edward Hwayoung Lee, Haeju Lee, Honglak Lee, Jinsik Lee , et al. (7 additional authors not shown)

    Abstract: We present EXAONE Deep series, which exhibits superior capabilities in various reasoning tasks, including math and coding benchmarks. We train our models mainly on the reasoning-specialized dataset that incorporates long streams of thought processes. Evaluation results show that our smaller models, EXAONE Deep 2.4B and 7.8B, outperform other models of comparable size, while the largest model, EXAO… ▽ More

    Submitted 19 March, 2025; v1 submitted 16 March, 2025; originally announced March 2025.

    Comments: arXiv admin note: substantial text overlap with arXiv:2412.04862, arXiv:2408.03541

  27. arXiv:2503.04713  [pdf, other

    eess.AS cs.AI cs.CL cs.LG cs.SD

    Scaling Rich Style-Prompted Text-to-Speech Datasets

    Authors: Anuj Diwan, Zhisheng Zheng, David Harwath, Eunsol Choi

    Abstract: We introduce Paralinguistic Speech Captions (ParaSpeechCaps), a large-scale dataset that annotates speech utterances with rich style captions. While rich abstract tags (e.g. guttural, nasal, pained) have been explored in small-scale human-annotated datasets, existing large-scale datasets only cover basic tags (e.g. low-pitched, slow, loud). We combine off-the-shelf text and speech embedders, class… ▽ More

    Submitted 6 March, 2025; originally announced March 2025.

  28. arXiv:2503.03444  [pdf, other

    cs.CL cs.AI

    Taxation Perspectives from Large Language Models: A Case Study on Additional Tax Penalties

    Authors: Eunkyung Choi, Young Jin Suh, Hun Park, Wonseok Hwang

    Abstract: How capable are large language models (LLMs) in the domain of taxation? Although numerous studies have explored the legal domain in general, research dedicated to taxation remain scarce. Moreover, the datasets used in these studies are either simplified, failing to reflect the real-world complexities, or unavailable as open source. To address this gap, we introduce PLAT, a new benchmark designed t… ▽ More

    Submitted 5 March, 2025; originally announced March 2025.

    Comments: 5 pages

  29. arXiv:2503.03064  [pdf, other

    cs.CL

    Improving LLM-as-a-Judge Inference with the Judgment Distribution

    Authors: Victor Wang, Michael J. Q. Zhang, Eunsol Choi

    Abstract: Using language models to scalably approximate human preferences on text quality (LLM-as-a-judge) has become a standard practice applicable to many tasks. A judgment is often extracted from the judge's textual output alone, typically with greedy decoding. However, LLM judges naturally provide distributions over judgment tokens, inviting a breadth of inference methods for extracting fine-grained pre… ▽ More

    Submitted 4 March, 2025; originally announced March 2025.

  30. arXiv:2503.02328  [pdf, other

    cs.CL cs.CY cs.HC cs.SI

    Limited Effectiveness of LLM-based Data Augmentation for COVID-19 Misinformation Stance Detection

    Authors: Eun Cheol Choi, Ashwin Balasubramanian, Jinhu Qi, Emilio Ferrara

    Abstract: Misinformation surrounding emerging outbreaks poses a serious societal threat, making robust countermeasures essential. One promising approach is stance detection (SD), which identifies whether social media posts support or oppose misleading claims. In this work, we finetune classifiers on COVID-19 misinformation SD datasets consisting of claims and corresponding tweets. Specifically, we test cont… ▽ More

    Submitted 4 March, 2025; originally announced March 2025.

  31. arXiv:2502.15779  [pdf, other

    cs.LG cs.AI cs.CL

    Rotate, Clip, and Partition: Towards W2A4KV4 Quantization by Integrating Rotation and Learnable Non-uniform Quantizer

    Authors: Euntae Choi, Sumin Song, Woosang Lim, Sungjoo Yoo

    Abstract: We propose Rotate, Clip, and Partition (RCP), a quantization-aware training (QAT) approach that first realizes extreme compression of LLMs with W2A4KV4(2-bit weight, 4-bit activation, and 4-bit KV cache) configuration. RCP integrates recent rotation techniques with a novel non-uniform weight quantizer design, by quantitatively analyzing the impact of random rotation on 2-bit weight quantization. O… ▽ More

    Submitted 17 February, 2025; originally announced February 2025.

  32. arXiv:2502.14259  [pdf, ps, other

    cs.LG

    LabTOP: A Unified Model for Lab Test Outcome Prediction on Electronic Health Records

    Authors: Sujeong Im, Jungwoo Oh, Edward Choi

    Abstract: Lab tests are fundamental for diagnosing diseases and monitoring patient conditions. However, frequent testing can be burdensome for patients, and test results may not always be immediately available. To address these challenges, we propose LabTOP, a unified model that predicts lab test outcomes by leveraging a language modeling approach on EHR data. Unlike conventional methods that estimate only… ▽ More

    Submitted 5 July, 2025; v1 submitted 19 February, 2025; originally announced February 2025.

    Comments: 11 pages for main text, 13 pages for appendix

  33. arXiv:2502.12767  [pdf, other

    cs.CL cs.AI

    R2-KG: General-Purpose Dual-Agent Framework for Reliable Reasoning on Knowledge Graphs

    Authors: Sumin Jo, Junseong Choi, Jiho Kim, Edward Choi

    Abstract: Recent studies have combined Large Language Models (LLMs) with Knowledge Graphs (KGs) to enhance reasoning, improving inference accuracy without additional training while mitigating hallucination. However, existing frameworks still suffer two practical drawbacks: they must be re-tuned whenever the KG or reasoning task changes, and they depend on a single, high-capacity LLM for reliable (i.e., trus… ▽ More

    Submitted 20 May, 2025; v1 submitted 18 February, 2025; originally announced February 2025.

  34. Transactional Dynamics in Hyperledger Fabric: A Stochastic Modeling and Performance Evaluation of Permissioned Blockchains

    Authors: Carlos Melo, Glauber Gonçalves, Francisco Airton Silva, Iure Fé, Ericksulino Moura, André Soares, Eunmi Choi, Dugki Min, Jae-Woo Lee, Tuan Anh Nguyen

    Abstract: Blockchain, often integrated with distributed systems and security enhancements, has significant potential in various industries. However, environmental concerns and the efficiency of consortia-controlled permissioned networks remain critical issues. We use a Stochastic Petri Net model to analyze transaction flows in Hyperledger Fabric networks, achieving a 95% confidence interval for response tim… ▽ More

    Submitted 13 February, 2025; originally announced February 2025.

  35. Optimal Resource Utilization in Hyperledger Fabric: A Comprehensive SPN-Based Performance Evaluation Paradigm

    Authors: Carlos Melo, Glauber Gonçalves, Francisco A. Silva, Leonel Feitosa, Iure Fé, André Soares, Eunmi Choi, Tuan Anh Nguyen, Dugki Min

    Abstract: Hyperledger Fabric stands as a leading framework for permissioned blockchain systems, ensuring data security and auditability for enterprise applications. As applications on this platform grow, understanding its complex configuration concerning various blockchain parameters becomes vital. These configurations significantly affect the system's performance and cost. In this research, we introduce a… ▽ More

    Submitted 12 February, 2025; originally announced February 2025.

  36. arXiv:2502.04757  [pdf, other

    cs.CV cs.CL

    ELITE: Enhanced Language-Image Toxicity Evaluation for Safety

    Authors: Wonjun Lee, Doehyeon Lee, Eugene Choi, Sangyoon Yu, Ashkan Yousefpour, Haon Park, Bumsub Ham, Suhyun Kim

    Abstract: Current Vision Language Models (VLMs) remain vulnerable to malicious prompts that induce harmful outputs. Existing safety benchmarks for VLMs primarily rely on automated evaluation methods, but these methods struggle to detect implicit harmful content or produce inaccurate evaluations. Therefore, we found that existing benchmarks have low levels of harmfulness, ambiguous data, and limited diversit… ▽ More

    Submitted 9 February, 2025; v1 submitted 7 February, 2025; originally announced February 2025.

  37. arXiv:2502.01122  [pdf, other

    cs.LG

    Learning Efficient Positional Encodings with Graph Neural Networks

    Authors: Charilaos I. Kanatsoulis, Evelyn Choi, Stephanie Jegelka, Jure Leskovec, Alejandro Ribeiro

    Abstract: Positional encodings (PEs) are essential for effective graph representation learning because they provide position awareness in inherently position-agnostic transformer architectures and increase the expressive capacity of Graph Neural Networks (GNNs). However, designing powerful and efficient PEs for graphs poses significant challenges due to the absence of canonical node ordering and the scale o… ▽ More

    Submitted 3 February, 2025; originally announced February 2025.

  38. arXiv:2501.17715  [pdf, other

    cs.CL

    RICoTA: Red-teaming of In-the-wild Conversation with Test Attempts

    Authors: Eujeong Choi, Younghun Jeong, Soomin Kim, Won Ik Cho

    Abstract: User interactions with conversational agents (CAs) evolve in the era of heavily guardrailed large language models (LLMs). As users push beyond programmed boundaries to explore and build relationships with these systems, there is a growing concern regarding the potential for unauthorized access or manipulation, commonly referred to as "jailbreaking." Moreover, with CAs that possess highly human-lik… ▽ More

    Submitted 29 January, 2025; originally announced January 2025.

    Comments: PACLIC 38

  39. arXiv:2501.17270  [pdf, other

    cs.CL cs.DB

    Comprehensive Evaluation for a Large Scale Knowledge Graph Question Answering Service

    Authors: Saloni Potdar, Daniel Lee, Omar Attia, Varun Embar, De Meng, Ramesh Balaji, Chloe Seivwright, Eric Choi, Mina H. Farid, Yiwen Sun, Yunyao Li

    Abstract: Question answering systems for knowledge graph (KGQA), answer factoid questions based on the data in the knowledge graph. KGQA systems are complex because the system has to understand the relations and entities in the knowledge-seeking natural language queries and map them to structured queries against the KG to answer them. In this paper, we introduce Chronos, a comprehensive evaluation framework… ▽ More

    Submitted 28 January, 2025; originally announced January 2025.

  40. arXiv:2501.12422  [pdf, other

    cs.LG cs.AI cs.CV

    CroMe: Multimodal Fake News Detection using Cross-Modal Tri-Transformer and Metric Learning

    Authors: Eunjee Choi, Junhyun Ahn, XinYu Piao, Jong-Kook Kim

    Abstract: Multimodal Fake News Detection has received increasing attention recently. Existing methods rely on independently encoded unimodal data and overlook the advantages of capturing intra-modality relationships and integrating inter-modal similarities using advanced techniques. To address these issues, Cross-Modal Tri-Transformer and Metric Learning for Multimodal Fake News Detection (CroMe) is propose… ▽ More

    Submitted 21 January, 2025; originally announced January 2025.

  41. arXiv:2412.19391  [pdf, other

    cs.CV cs.AI cs.LG

    An In-Depth Analysis of Adversarial Discriminative Domain Adaptation for Digit Classification

    Authors: Eugene Choi, Julian Rodriguez, Edmund Young

    Abstract: Domain adaptation is an active area of research driven by the growing demand for robust machine learning models that perform well on real-world data. Adversarial learning for deep neural networks (DNNs) has emerged as a promising approach to improving generalization ability, particularly for image classification. In this paper, we implement a specific adversarial learning technique known as Advers… ▽ More

    Submitted 6 January, 2025; v1 submitted 26 December, 2024; originally announced December 2024.

    Comments: Replacement: Updated methodology section to include grayscale preprocessing of SVHN data

  42. arXiv:2412.15797  [pdf, other

    cs.CL

    Ensembling Large Language Models with Process Reward-Guided Tree Search for Better Complex Reasoning

    Authors: Sungjin Park, Xiao Liu, Yeyun Gong, Edward Choi

    Abstract: Despite recent advances in large language models, open-source models often struggle to consistently perform well on complex reasoning tasks. Existing ensemble methods, whether applied at the token or output levels, fail to address these challenges. In response, we present Language model Ensemble with Monte Carlo Tree Search (LE-MCTS), a novel framework for process-level ensembling of language mode… ▽ More

    Submitted 20 December, 2024; originally announced December 2024.

  43. arXiv:2412.04862  [pdf, other

    cs.CL

    EXAONE 3.5: Series of Large Language Models for Real-world Use Cases

    Authors: LG AI Research, Soyoung An, Kyunghoon Bae, Eunbi Choi, Kibong Choi, Stanley Jungkyu Choi, Seokhee Hong, Junwon Hwang, Hyojin Jeon, Gerrard Jeongwon Jo, Hyunjik Jo, Jiyeon Jung, Yountae Jung, Hyosang Kim, Joonkee Kim, Seonghwan Kim, Soyeon Kim, Sunkyoung Kim, Yireun Kim, Yongil Kim, Youchul Kim, Edward Hwayoung Lee, Haeju Lee, Honglak Lee, Jinsik Lee , et al. (8 additional authors not shown)

    Abstract: This technical report introduces the EXAONE 3.5 instruction-tuned language models, developed and released by LG AI Research. The EXAONE 3.5 language models are offered in three configurations: 32B, 7.8B, and 2.4B. These models feature several standout capabilities: 1) exceptional instruction following capabilities in real-world scenarios, achieving the highest scores across seven benchmarks, 2) ou… ▽ More

    Submitted 9 December, 2024; v1 submitted 6 December, 2024; originally announced December 2024.

    Comments: arXiv admin note: text overlap with arXiv:2408.03541

  44. arXiv:2412.02043  [pdf

    cs.IR cs.AI

    Future of Information Retrieval Research in the Age of Generative AI

    Authors: James Allan, Eunsol Choi, Daniel P. Lopresti, Hamed Zamani

    Abstract: In the fast-evolving field of information retrieval (IR), the integration of generative AI technologies such as large language models (LLMs) is transforming how users search for and interact with information. Recognizing this paradigm shift at the intersection of IR and generative AI (IR-GenAI), a visioning workshop supported by the Computing Community Consortium (CCC) was held in July 2024 to dis… ▽ More

    Submitted 2 December, 2024; originally announced December 2024.

  45. arXiv:2411.15927  [pdf, other

    cs.CL cs.AI

    Generative Prompt Internalization

    Authors: Haebin Shin, Lei Ji, Yeyun Gong, Sungdong Kim, Eunbi Choi, Minjoon Seo

    Abstract: Prompts used in recent large language model based applications are often fixed and lengthy, leading to significant computational overhead. To address this challenge, we propose Generative Prompt Internalization (GenPI), a lightweight method that employs a joint training approach. GenPI not only replicates the behavior of models with prompt inputs but also generates the content of the prompt along… ▽ More

    Submitted 24 March, 2025; v1 submitted 24 November, 2024; originally announced November 2024.

    Comments: NAACL 2025 (Main Conference)

  46. arXiv:2411.14042  [pdf, other

    cs.CL cs.AI

    Forecasting Future International Events: A Reliable Dataset for Text-Based Event Modeling

    Authors: Daehoon Gwak, Junwoo Park, Minho Park, Chaehun Park, Hyunchan Lee, Edward Choi, Jaegul Choo

    Abstract: Predicting future international events from textual information, such as news articles, has tremendous potential for applications in global policy, strategic decision-making, and geopolitics. However, existing datasets available for this task are often limited in quality, hindering the progress of related research. In this paper, we introduce WORLDREP (WORLD Relationship and Event Prediction), a n… ▽ More

    Submitted 21 November, 2024; originally announced November 2024.

    Comments: EMNLP 2024 Findings

  47. arXiv:2411.05787  [pdf, other

    cs.CL

    RefreshKV: Updating Small KV Cache During Long-form Generation

    Authors: Fangyuan Xu, Tanya Goyal, Eunsol Choi

    Abstract: Generating long sequences of tokens given a long-context input is a very compute-intensive inference scenario for large language models (LLMs). One prominent inference speed-up approach is to construct a smaller key-value (KV) cache, relieving LLMs from computing attention over a long sequence of tokens. While such methods work well to generate short sequences, their performance degrades rapidly f… ▽ More

    Submitted 3 March, 2025; v1 submitted 8 November, 2024; originally announced November 2024.

  48. arXiv:2411.02551  [pdf, other

    cs.SD cs.AI cs.MM eess.AS

    PIAST: A Multimodal Piano Dataset with Audio, Symbolic and Text

    Authors: Hayeon Bang, Eunjin Choi, Megan Finch, Seungheon Doh, Seolhee Lee, Gyeong-Hoon Lee, Juhan Nam

    Abstract: While piano music has become a significant area of study in Music Information Retrieval (MIR), there is a notable lack of datasets for piano solo music with text labels. To address this gap, we present PIAST (PIano dataset with Audio, Symbolic, and Text), a piano music dataset. Utilizing a piano-specific taxonomy of semantic tags, we collected 9,673 tracks from YouTube and added human annotations… ▽ More

    Submitted 7 November, 2024; v1 submitted 4 November, 2024; originally announced November 2024.

    Comments: Accepted for publication at the 3rd Workshop on NLP for Music and Audio (NLP4MusA 2024)

  49. arXiv:2411.01813  [pdf, other

    cs.RO cs.AI

    So You Think You Can Scale Up Autonomous Robot Data Collection?

    Authors: Suvir Mirchandani, Suneel Belkhale, Joey Hejna, Evelyn Choi, Md Sazzad Islam, Dorsa Sadigh

    Abstract: A long-standing goal in robot learning is to develop methods for robots to acquire new skills autonomously. While reinforcement learning (RL) comes with the promise of enabling autonomous data collection, it remains challenging to scale in the real-world partly due to the significant effort required for environment design and instrumentation, including the need for designing reset functions or acc… ▽ More

    Submitted 4 November, 2024; originally announced November 2024.

    Comments: 21 pages, 25 figures. Conference on Robot Learning (CoRL) 2024

  50. arXiv:2410.23820  [pdf, other

    cs.LG cs.AI cs.CV

    Disentangling Disentangled Representations: Towards Improved Latent Units via Diffusion Models

    Authors: Youngjun Jun, Jiwoo Park, Kyobin Choo, Tae Eun Choi, Seong Jae Hwang

    Abstract: Disentangled representation learning (DRL) aims to break down observed data into core intrinsic factors for a profound understanding of the data. In real-world scenarios, manually defining and labeling these factors are non-trivial, making unsupervised methods attractive. Recently, there have been limited explorations of utilizing diffusion models (DMs), which are already mainstream in generative… ▽ More

    Submitted 31 October, 2024; originally announced October 2024.