All Practice Areas

AI & Technology Law

AI·기술법

Jurisdiction: All US KR EU Intl
LOW Academic International

The Illusion of Stochasticity in LLMs

arXiv:2604.06543v1 Announce Type: new Abstract: In this work, we demonstrate that reliable stochastic sampling is a fundamental yet unfulfilled requirement for Large Language Models (LLMs) operating as agents. Agentic systems are frequently required to sample from distributions, often inferred from...

1 min 1 week, 5 days ago
ai llm
LOW News International

Atlassian launches visual AI tools and third-party agents in Confluence

Confluence users can now create visual assets within the software in addition to new third-party agents working with Lovable, Replit, and Gamma.

1 min 1 week, 5 days ago
ai artificial intelligence
LOW Academic United States

Beyond Facts: Benchmarking Distributional Reading Comprehension in Large Language Models

arXiv:2604.06201v1 Announce Type: new Abstract: While most reading comprehension benchmarks for LLMs focus on factual information that can be answered by localizing specific textual evidence, many real-world tasks require understanding distributional information, such as population-level trends and preferences expressed across...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

AE-ViT: Stable Long-Horizon Parametric Partial Differential Equations Modeling

arXiv:2604.06475v1 Announce Type: new Abstract: Deep Learning Reduced Order Models (ROMs) are becoming increasingly popular as surrogate models for parametric partial differential equations (PDEs) due to their ability to handle high-dimensional data, approximate highly nonlinear mappings, and utilize GPUs. Existing...

1 min 1 week, 5 days ago
ai deep learning
LOW Academic International

SensorPersona: An LLM-Empowered System for Continual Persona Extraction from Longitudinal Mobile Sensor Streams

arXiv:2604.06204v1 Announce Type: new Abstract: Personalization is essential for Large Language Model (LLM)-based agents to adapt to users' preferences and improve response quality and task performance. However, most existing approaches infer personas from chat histories, which capture only self-disclosed information...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

The Stepwise Informativeness Assumption: Why are Entropy Dynamics and Reasoning Correlated in LLMs?

arXiv:2604.06192v1 Announce Type: new Abstract: Recent work uses entropy-based signals at multiple representation levels to study reasoning in large language models, but the field remains largely empirical. A central unresolved puzzle is why internal entropy dynamics, defined under the predictive...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Distributional Open-Ended Evaluation of LLM Cultural Value Alignment Based on Value Codebook

arXiv:2604.06210v1 Announce Type: new Abstract: As LLMs are globally deployed, aligning their cultural value orientations is critical for safety and user engagement. However, existing benchmarks face the Construct-Composition-Context ($C^3$) challenge: relying on discriminative, multiple-choice formats that probe value knowledge rather...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

State-of-the-Art Arabic Language Modeling with Sparse MoE Fine-Tuning and Chain-of-Thought Distillation

arXiv:2604.06421v1 Announce Type: new Abstract: This paper introduces Arabic-DeepSeek-R1, an application-driven open-source Arabic LLM that leverages a sparse MoE backbone to address the digital equity gap for under-represented languages, and establishes a new SOTA across the entire Open Arabic LLM...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

DiffuMask: Diffusion Language Model for Token-level Prompt Pruning

arXiv:2604.06627v1 Announce Type: new Abstract: In-Context Learning and Chain-of-Thought prompting improve reasoning in large language models (LLMs). These typically come at the cost of longer, more expensive prompts that may contain redundant information. Prompt compression based on pruning offers a...

1 min 1 week, 5 days ago
ai llm
LOW Academic United States

Application-Driven Pedagogical Knowledge Optimization of Open-Source LLMs via Reinforcement Learning and Supervised Fine-Tuning

arXiv:2604.06385v1 Announce Type: new Abstract: We present an innovative multi-stage optimization strategy combining reinforcement learning (RL) and supervised fine-tuning (SFT) to enhance the pedagogical knowledge of large language models (LLMs), as illustrated by EduQwen 32B-RL1, EduQwen 32B-SFT, and an optional...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

SubFLOT: Submodel Extraction for Efficient and Personalized Federated Learning via Optimal Transport

arXiv:2604.06631v1 Announce Type: new Abstract: Federated Learning (FL) enables collaborative model training while preserving data privacy, but its practical deployment is hampered by system and statistical heterogeneity. While federated network pruning offers a path to mitigate these issues, existing methods...

1 min 1 week, 5 days ago
ai data privacy
LOW Academic United States

Asymptotic-Preserving Neural Networks for Viscoelastic Parameter Identification in Multiscale Blood Flow Modeling

arXiv:2604.06287v1 Announce Type: new Abstract: Mathematical models and numerical simulations offer a non-invasive way to explore cardiovascular phenomena, providing access to quantities that cannot be measured directly. In this study, we start with a one-dimensional multiscale blood flow model that...

1 min 1 week, 5 days ago
ai neural network
LOW Academic International

Busemann energy-based attention for emotion analysis in Poincar\'e discs

arXiv:2604.06752v1 Announce Type: new Abstract: We present EmBolic - a novel fully hyperbolic deep learning architecture for fine-grained emotion analysis from textual messages. The underlying idea is that hyperbolic geometry efficiently captures hierarchies between both words and emotions. In our...

1 min 1 week, 5 days ago
ai deep learning
LOW Academic International

Extracting Breast Cancer Phenotypes from Clinical Notes: Comparing LLMs with Classical Ontology Methods

arXiv:2604.06208v1 Announce Type: new Abstract: A significant amount of data held in Oncology Electronic Medical Records (EMRs) is contained in unstructured provider notes -- including but not limited to the chemotherapy (or cancer treatment) outcome, different biomarkers, the tumor's location,...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Consistency-Guided Decoding with Proof-Driven Disambiguation for Three-Way Logical Question Answering

arXiv:2604.06196v1 Announce Type: new Abstract: Three-way logical question answering (QA) assigns $True/False/Unknown$ to a hypothesis $H$ given a premise set $S$. While modern large language models (LLMs) can be accurate on isolated examples, we identify two recurring failure modes in...

1 min 1 week, 5 days ago
ai llm
LOW Academic European Union

Time-Series Classification with Multivariate Statistical Dependence Features

arXiv:2604.06537v1 Announce Type: new Abstract: In this paper, we propose a novel framework for non-stationary time-series analysis that replaces conventional correlation-based statistics with direct estimation of statistical dependence in the normalized joint density of input and target signals, the cross...

1 min 1 week, 5 days ago
algorithm neural network
LOW Academic International

Illocutionary Explanation Planning for Source-Faithful Explanations in Retrieval-Augmented Language Models

arXiv:2604.06211v1 Announce Type: new Abstract: Natural language explanations produced by large language models (LLMs) are often persuasive, but not necessarily scrutable: users cannot easily verify whether the claims in an explanation are supported by evidence. In XAI, this motivates a...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

ART: Attention Replacement Technique to Improve Factuality in LLMs

arXiv:2604.06393v1 Announce Type: new Abstract: Hallucination in large language models (LLMs) continues to be a significant issue, particularly in tasks like question answering, where models often generate plausible yet incorrect or irrelevant information. Although various methods have been proposed to...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

TalkLoRA: Communication-Aware Mixture of Low-Rank Adaptation for Large Language Models

arXiv:2604.06291v1 Announce Type: new Abstract: Low-Rank Adaptation (LoRA) enables parameter-efficient fine-tuning of Large Language Models (LLMs), and recent Mixture-of-Experts (MoE) extensions further enhance flexibility by dynamically combining multiple LoRA experts. However, existing MoE-augmented LoRA methods assume that experts operate independently,...

1 min 1 week, 5 days ago
ai llm
LOW Academic European Union

Toward a universal foundation model for graph-structured data

arXiv:2604.06391v1 Announce Type: new Abstract: Graphs are a central representation in biomedical research, capturing molecular interaction networks, gene regulatory circuits, cell--cell communication maps, and knowledge graphs. Despite their importance, currently there is not a broadly reusable foundation model available for...

1 min 1 week, 5 days ago
ai neural network
LOW Academic International

The Depth Ceiling: On the Limits of Large Language Models in Discovering Latent Planning

arXiv:2604.06427v1 Announce Type: new Abstract: The viability of chain-of-thought (CoT) monitoring hinges on models being unable to reason effectively in their latent representations. Yet little is known about the limits of such latent reasoning in LLMs. We test these limits...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

LLM-Augmented Knowledge Base Construction For Root Cause Analysis

arXiv:2604.06171v1 Announce Type: new Abstract: Communications networks now form the backbone of our digital world, with fast and reliable connectivity. However, even with appropriate redundancy and failover mechanisms, it is difficult to guarantee "five 9s" (99.999 %) reliability, requiring rapid...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Scientific Knowledge-driven Decoding Constraints Improving the Reliability of LLMs

arXiv:2604.06603v1 Announce Type: new Abstract: Large language models (LLMs) have shown strong knowledge reserves and task-solving capabilities, but still face the challenge of severe hallucination, hindering their practical application. Though scientific theories and rules can efficiently direct the behaviors of...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Learning to Interrupt in Language-based Multi-agent Communication

arXiv:2604.06452v1 Announce Type: new Abstract: Multi-agent systems using large language models (LLMs) have demonstrated impressive capabilities across various domains. However, current agent communication suffers from verbose output that overload context and increase computational costs. Although existing approaches focus on compressing...

1 min 1 week, 5 days ago
ai llm
LOW News International

OpenAI releases a new safety blueprint to address the rise in child sexual exploitation

OpenAI's new Child Safety Blueprint aims to tackle the alarming rise in child sexual exploitation linked to advancements in AI.

1 min 1 week, 5 days ago
ai chatgpt
LOW Academic European Union

Blending Human and LLM Expertise to Detect Hallucinations and Omissions in Mental Health Chatbot Responses

arXiv:2604.06216v1 Announce Type: new Abstract: As LLM-powered chatbots are increasingly deployed in mental health services, detecting hallucinations and omissions has become critical for user safety. However, state-of-the-art LLM-as-a-judge methods often fail in high-risk healthcare contexts, where subtle errors can have...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

SHAPE: Stage-aware Hierarchical Advantage via Potential Estimation for LLM Reasoning

arXiv:2604.06636v1 Announce Type: new Abstract: Process supervision has emerged as a promising approach for enhancing LLM reasoning, yet existing methods fail to distinguish meaningful progress from mere verbosity, leading to limited reasoning capabilities and unresolved token inefficiency. To address this,...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Cross-Lingual Transfer and Parameter-Efficient Adaptation in the Turkic Language Family: A Theoretical Framework for Low-Resource Language Models

arXiv:2604.06202v1 Announce Type: new Abstract: Large language models (LLMs) have transformed natural language processing, yet their capabilities remain uneven across languages. Most multilingual models are trained primarily on high-resource languages, leaving many languages with large speaker populations underrepresented in both...

1 min 1 week, 5 days ago
ai llm
LOW Academic European Union

A Comparative Study of Demonstration Selection for Practical Large Language Models-based Next POI Prediction

arXiv:2604.06207v1 Announce Type: new Abstract: This paper investigates demonstration selection strategies for predicting a user's next point-of-interest (POI) using large language models (LLMs), aiming to accurately forecast a user's subsequent location based on historical check-in data. While in-context learning (ICL)...

1 min 1 week, 5 days ago
ai llm
LOW Academic International

Attention Flows: Tracing LLM Conceptual Engagement via Story Summaries

arXiv:2604.06416v1 Announce Type: new Abstract: Although LLM context lengths have grown, there is evidence that their ability to integrate information across long-form texts has not kept pace. We evaluate one such understanding task: generating summaries of novels. When human authors...

1 min 1 week, 5 days ago
ai llm
Previous Page 35 of 200 Next

Impact Distribution

Critical 0
High 57
Medium 938
Low 4987