Multi-lingual Multi-institutional Electronic Health Record based Predictive Model
arXiv:2604.00027v1 Announce Type: new Abstract: Large-scale EHR prediction across institutions is hindered by substantial heterogeneity in schemas and code systems. Although Common Data Models (CDMs) can standardize records for multi-institutional learning, the manual harmonization and vocabulary mapping are costly and...
Eyla: Toward an Identity-Anchored LLM Architecture with Integrated Biological Priors -- Vision, Implementation Attempt, and Lessons from AI-Assisted Development
arXiv:2604.00009v1 Announce Type: cross Abstract: We present the design rationale, implementation attempt, and failure analysis of Eyla, a proposed identity-anchored LLM architecture that integrates biologically-inspired subsystems -- including HiPPO-initialized state-space models, zero-initialized adapters, episodic memory retrieval, and calibrated uncertainty training...
Agent Q-Mix: Selecting the Right Action for LLM Multi-Agent Systems through Reinforcement Learning
arXiv:2604.00344v1 Announce Type: new Abstract: Large Language Models (LLMs) have shown remarkable performance in completing various tasks. However, solving complex problems often requires the coordination of multiple agents, raising a fundamental question: how to effectively select and interconnect these agents....
CRIT: Graph-Based Automatic Data Synthesis to Enhance Cross-Modal Multi-Hop Reasoning
arXiv:2604.01634v1 Announce Type: new Abstract: Real-world reasoning often requires combining information across modalities, connecting textual context with visual cues in a multi-hop process. Yet, most multimodal benchmarks fail to capture this ability: they typically rely on single images or set...
Relevance to Intellectual Property practice area: This article discusses the development of a new dataset and benchmark, CRIT, designed to enhance cross-modal multi-hop reasoning in Vision-Language Models (VLMs). The research findings and policy signals in this article are relevant to Intellectual Property practice areas as they highlight the need for more advanced and robust AI models in detecting and preventing copyright infringement, particularly in the context of image and text-based content. Key legal developments: The article's focus on developing more advanced AI models to improve cross-modal multi-hop reasoning has implications for the detection and prevention of copyright infringement in digital content, including images and text. This development may lead to more effective tools for copyright holders to protect their work and for AI-powered content moderation systems to detect and remove infringing content. Research findings: The article's experiments show that even state-of-the-art models struggle on cross-modal multi-hop reasoning tasks, but models trained on CRIT show significant gains in this area. This suggests that the development of more advanced AI models, like CRIT, can improve the accuracy and effectiveness of AI-powered content moderation systems and copyright infringement detection tools. Policy signals: The article's focus on developing more advanced AI models to improve cross-modal multi-hop reasoning has implications for the development of policies and regulations related to AI-powered content moderation and copyright infringement detection. This development may lead to more effective tools for copyright holders to protect their work and for AI-powered content moderation systems to detect and remove infringing content.
### **Jurisdictional Comparison & Analytical Commentary on CRIT’s Impact on Intellectual Property Practice** The introduction of **CRIT**—a graph-based dataset designed to enhance cross-modal multi-hop reasoning in Vision-Language Models (VLMs)—raises significant **intellectual property (IP) considerations** across jurisdictions, particularly regarding **data ownership, copyright in AI-generated content, and liability for AI hallucinations**. In the **U.S.**, where AI-generated works face restrictive copyright protections (as seen in *Thaler v. Perlmutter*), CRIT’s synthetic data pipeline may trigger debates over **who owns the training data**—the researchers, the automated pipeline, or the underlying sources. **South Korea**, under its more permissive stance (e.g., allowing copyright in AI-generated works if human creativity is involved), may view CRIT’s manually verified test set as protectable, but disputes over **derivative works** could arise if CRIT’s outputs are used to train commercial VLMs. **Internationally**, under the **Berne Convention**, CRIT’s synthetic data may lack protection if deemed purely machine-generated, but jurisdictions like the **EU (under the AI Act)** could impose **liability frameworks** for AI-driven hallucinations, shifting enforcement risks to model developers rather than dataset creators. The broader implication is that **CRIT’s release may accelerate regulatory scrutiny** on AI training data provenance, potentially leading to **mandatory disclosure of
As a Patent Prosecution & Infringement Expert, I'll provide domain-specific expert analysis of the article's implications for practitioners. **Technical Analysis:** The article discusses a novel dataset and benchmark, CRIT, designed to evaluate the performance of Vision-Language Models (VLMs) in cross-modal multi-hop reasoning tasks. The CRIT dataset is built using a graph-based automatic pipeline to generate complex cross-modal reasoning tasks, addressing the limitations of existing multimodal benchmarks. This dataset and benchmark are likely to be used to evaluate the performance of VLMs in various applications, such as image and text analysis, and may lead to the development of new and improved VLMs. **Patent Implications:** The CRIT dataset and benchmark may have implications for patent applications related to VLMs and multimodal reasoning. For example, patent claims directed to VLMs may need to be revised to account for the limitations of existing multimodal benchmarks and the improved performance of VLMs on the CRIT dataset. Additionally, the CRIT dataset may be used as prior art to challenge the novelty and non-obviousness of patent claims directed to VLMs. **Case Law and Regulatory Connections:** The CRIT dataset and benchmark may be connected to the following case law and regulatory issues: * The CRIT dataset and benchmark may be relevant to the discussion of obviousness and non-obviousness in patent law, particularly in the context of artificial intelligence and machine learning inventions (e.g
DISCO-TAB: A Hierarchical Reinforcement Learning Framework for Privacy-Preserving Synthesis of Complex Clinical Data
arXiv:2604.01481v1 Announce Type: new Abstract: The development of robust clinical decision support systems is frequently impeded by the scarcity of high-fidelity, privacy-preserving biomedical data. While Generative Large Language Models (LLMs) offer a promising avenue for synthetic data generation, they often...
Can Large Language Models Self-Correct in Medical Question Answering? An Exploratory Study
arXiv:2604.00261v2 Announce Type: new Abstract: Large language models (LLMs) have achieved strong performance on medical question answering (medical QA), and chain-of-thought (CoT) prompting has further improved results by eliciting explicit intermediate reasoning; meanwhile, self-reflective (self-corrective) prompting has been widely claimed...
HippoCamp: Benchmarking Contextual Agents on Personal Computers
arXiv:2604.01221v1 Announce Type: new Abstract: We present HippoCamp, a new benchmark designed to evaluate agents' capabilities on multimodal file management. Unlike existing agent benchmarks that focus on tasks like web interaction, tool use, or software automation in generic settings, HippoCamp...
### **Relevance to Intellectual Property (IP) Practice** This academic article highlights **critical limitations in AI-driven file management and reasoning**, particularly concerning **context-aware retrieval and multimodal data processing**—key challenges in IP-intensive industries where large datasets (e.g., patents, trademarks, copyrighted works) must be efficiently analyzed. The **48.3% accuracy gap** in user profiling and **long-horizon retrieval failures** signal potential risks in AI-assisted legal research, prior art searches, and automated IP documentation handling. Additionally, the **emphasis on cross-modal reasoning** underscores the need for robust **AI governance frameworks** in IP practice to mitigate errors in automated document analysis. *(Note: This is not legal advice but an analysis of academic research relevance to IP legal practice.)*
### **Jurisdictional Comparison & Analytical Commentary on *HippoCamp* and Its IP Implications** The *HippoCamp* benchmark highlights critical gaps in AI-driven contextual reasoning, particularly in handling personal data—an area with significant implications for **data privacy, copyright, and trade secret protections** across jurisdictions. In the **U.S.**, where IP frameworks like the *Computer Fraud and Abuse Act (CFAA)* and *Defend Trade Secrets Act (DTSA)* govern unauthorized access to personal or proprietary systems, the benchmark’s findings on "long-horizon retrieval" and "cross-modal reasoning" could raise concerns about **unintended data exfiltration** by AI agents, potentially triggering liability under **copyright scraping laws (e.g., *Authors Guild v. Google*)** or **trade secret misappropriation claims**. **Korea’s** approach under the *Personal Information Protection Act (PIPA)* and *Unfair Competition Prevention Act (UCPA)* would similarly scrutinize AI agents’ handling of personal files, with stricter penalties for **unauthorized data processing**—though Korea’s **AI ethics guidelines** may encourage proactive compliance. At the **international level**, the *EU AI Act* and *GDPR* would impose **high-risk AI system obligations**, requiring transparency in training data sources and user consent for personal file interactions, while the *WIPO Copyright Treaty* and *TRIPS Agreement*
### **Expert Analysis of *HippoCamp* for Patent Practitioners** 1. **Technical & Legal Implications for AI/ML Patents** The *HippoCamp* benchmark highlights critical deficiencies in **multimodal large language models (MLLMs)** and **agentic systems** when handling **personal file management**, particularly in **long-horizon retrieval** and **cross-modal reasoning**. This aligns with existing patent trends in **AI-driven file systems** (e.g., USPTO Class 707/100-707/104 for database/file management) and **context-aware computing** (e.g., USPTO Class 706/46-706/50 for AI reasoning). The benchmark’s focus on **real-world personal data** (42.4 GB across 2K+ files) may raise **privacy and data security concerns**, potentially intersecting with **GDPR, CCPA, or trade secret protections** if such systems are commercialized. 2. **Potential Patentability & Prior Art Considerations** The benchmark’s **failure diagnosis framework** (46.1K structured trajectories) could inform **patent claims** related to **AI error correction, adaptive retrieval, or multimodal fusion techniques**. However, prior art in **personal file management agents** (e.g., USPTO 10,984,123 B
RefineRL: Advancing Competitive Programming with Self-Refinement Reinforcement Learning
arXiv:2604.00790v1 Announce Type: new Abstract: While large language models (LLMs) have demonstrated strong performance on complex reasoning tasks such as competitive programming (CP), existing methods predominantly focus on single-attempt settings, overlooking their capacity for iterative refinement. In this paper, we...
Announcing the ICML 2026 Tutorials
While this article primarily focuses on the organization of the ICML 2026 Tutorials and does not directly address legal developments, it does signal important trends relevant to **Intellectual Property (IP) in AI and machine learning (ML)**. The inclusion of tutorials on **numerical optimization, probabilistic numerics, and calibration** suggests growing academic and practical interest in explainable AI (XAI), algorithmic fairness, and decision-making transparency—areas increasingly intersecting with **patent eligibility, copyright, and data governance** under frameworks like the **EU AI Act** and **U.S. patent law updates on AI inventions**. The rigorous review process also reflects broader industry and regulatory emphasis on **ethical AI and reproducibility**, which may influence future IP litigation and licensing strategies. For IP practitioners, this signals the need to monitor how emerging ML techniques are being **protected, challenged, or regulated** in patent and trade secret contexts.
### **Analytical Commentary: Impact of ICML 2026 Tutorials on Intellectual Property (IP) Practice** The ICML 2026 Tutorials announcement highlights the evolving nature of academic and industry collaboration in machine learning (ML), raising important IP considerations regarding **open-access dissemination, proprietary knowledge protection, and collaborative innovation frameworks**. While the conference promotes **open educational resources (OER) and community-driven learning**, jurisdictions like the **US (patent-first approach), South Korea (balanced innovation policy), and international regimes (TRIPS/WIPO)** differ in how they balance **public disclosure (prior art) against patentability, trade secret protection, and collaborative R&D incentives**. The tutorial format itself—whether it involves **invited experts, community submissions, or rigorous peer review**—may influence **IP ownership of derivative works, licensing models, and the enforceability of open-source commitments**, particularly in cross-border collaborations where **Korean "creative commons" policies, US Bayh-Dole Act implications, and WIPO’s open-access principles** may lead to divergent legal interpretations. #### **Key Jurisdictional Comparisons:** 1. **United States:** - The **Bayh-Dole Act** encourages patenting of federally funded research, but ICML’s open-access policy may conflict with institutional IP policies if tutorials derive from patented work. - **Trade secret risks** arise if invited speakers disclose proprietary techniques without formal ND
### **Domain-Specific Expert Analysis for Patent Practitioners** This article highlights the **ICML 2026 Tutorials** selection process, emphasizing **rigorous peer review, community input, and expert-led instruction**—key considerations in **patent prosecution strategy**, particularly for **software and AI-related inventions**. The structured review process (invited, community-sourced, and peer-reviewed) mirrors the **USPTO’s subject matter eligibility (35 U.S.C. § 101) and obviousness (35 U.S.C. § 103) analyses**, where examiner discretion and prior art play pivotal roles. **Case Law & Regulatory Connections:** - **Alice Corp. v. CLS Bank (2014)** – The USPTO’s **Step 2B (inventive concept)** analysis aligns with ICML’s emphasis on **non-artificial evaluation** of tutorial proposals, ensuring substantive contributions beyond routine practice. - **USPTO’s 2019 Revised Patent Subject Matter Eligibility Guidance (PEG)** – The **machine learning tutorial topics** (e.g., numerical optimization, probabilistic numerics) must demonstrate **technological improvement** to overcome § 101 rejections, similar to how ICML evaluates **novelty and practical utility** in submissions. **Strategic Implications for Patent Practitioners:** 1. **Drafting AI/ML Patent Claims**
Open, Reliable, and Collective: A Community-Driven Framework for Tool-Using AI Agents
arXiv:2604.00137v1 Announce Type: new Abstract: Tool-integrated LLMs can retrieve, compute, and take real-world actions via external tools, but reliability remains a key bottleneck. We argue that failures stem from both tool-use accuracy (how well an agent invokes a tool) and...
Are they human? Detecting large language models by probing human memory constraints
arXiv:2604.00016v1 Announce Type: cross Abstract: The validity of online behavioral research relies on study participants being human rather than machine. In the past, it was possible to detect machines by posing simple challenges that were easily solved by humans but...
Learning ECG Image Representations via Dual Physiological-Aware Alignments
arXiv:2604.01526v1 Announce Type: new Abstract: Electrocardiograms (ECGs) are among the most widely used diagnostic tools for cardiovascular diseases, and a large amount of ECG data worldwide appears only in image form. However, most existing automated ECG analysis methods rely on...
Proactive Agent Research Environment: Simulating Active Users to Evaluate Proactive Assistants
arXiv:2604.00842v1 Announce Type: new Abstract: Proactive agents that anticipate user needs and autonomously execute tasks hold great promise as digital assistants, yet the lack of realistic user simulation frameworks hinders their development. Existing approaches model apps as flat tool-calling APIs,...
Optimizing EEG Graph Structure for Seizure Detection: An Information Bottleneck and Self-Supervised Learning Approach
arXiv:2604.01595v1 Announce Type: new Abstract: Seizure detection from EEG signals is highly challenging due to complex spatiotemporal dynamics and extreme inter-patient variability. To model them, recent methods construct dynamic graphs via statistical correlations, predefined similarity measures, or implicit learning, yet...
This academic article on **EEG seizure detection** is not directly relevant to **Intellectual Property (IP) law practice**, as it focuses on **machine learning, signal processing, and healthcare applications** rather than legal, regulatory, or policy developments in IP. The research highlights advancements in **AI-driven medical diagnostics**, which may have **indirect implications for patent law** (e.g., patentability of AI-based medical algorithms in jurisdictions like the U.S. or EU), but it does not present **key legal developments, regulatory changes, or policy signals** relevant to current IP practice. For IP practitioners, this type of research would be more pertinent in **patent drafting, prior art searches, or technology licensing** rather than legal analysis or regulatory tracking. If you're looking for **IP-specific legal updates**, I recommend monitoring sources like **WIPO, USPTO, EPO, KIPO, or industry-specific legal news** for patent law changes, enforcement actions, or policy shifts. Would you like me to track a different type of legal news?
### **Jurisdictional Comparison & Analytical Commentary on the Impact of *IRENE* on Intellectual Property Practice** The development of *IRENE* (Information Bottleneck-guided EEG SeizuRE DetectioN via SElf-Supervised Learning) presents significant implications for **patent eligibility, trade secret protection, and data governance** across jurisdictions. In the **U.S.**, where patent eligibility under *Alice* and *Bilski* often hinges on whether AI-driven medical diagnostics are deemed "abstract" or "technological," *IRENE*’s structured approach to EEG graph optimization may strengthen patent claims by emphasizing its **specific application in seizure detection** rather than mere algorithmic abstraction. **South Korea**, under the *Patent Act* and *Unfair Competition Prevention Act*, would likely favor trade secret protection for proprietary EEG processing techniques, given its robust enforcement of trade secrets (e.g., Samsung’s past litigation over AI chip designs). At the **international level**, under the **TRIPS Agreement** and **WIPO’s AI guidelines**, patentability would depend on whether jurisdictions classify *IRENE* as a **technical solution** (favored in Europe under the EPO’s AI patent guidelines) or a **medical method** (often excluded in many jurisdictions). The rise of **self-supervised learning in healthcare** also raises **GDPR/PIPL compliance issues**, as EEG data
### **Expert Analysis of Patent & IP Implications for Practitioners** #### **1. Patentability & Novelty Considerations** The paper introduces **IRENE**, a novel framework combining **Information Bottleneck (IB) theory** with **self-supervised graph learning** for EEG-based seizure detection. Key novel aspects include: - **Dynamic graph denoising** tailored to EEG noise characteristics (unlike prior art relying on statistical correlations or predefined similarity measures). - **Joint optimization** of graph structure and spatial-temporal representations under the IB principle. - **Graph Masked AutoEncoder (GMAE)** for structure-aware signal reconstruction. **Potential patent claims** could focus on: - A **method for EEG seizure detection** comprising steps of: - Constructing a denoised dynamic graph from EEG signals using an IB-guided approach. - Applying a self-supervised GMAE to learn compact representations. - A **system** comprising EEG sensors, a processor, and memory storing executable instructions for the claimed method. **Prior Art Risks:** - **Graph-based EEG processing** (e.g., dynamic functional connectivity graphs) is well-known (e.g., *Bashivan et al., 2016*). - **Self-supervised learning in EEG** (e.g., contrastive learning for seizure detection) has been explored (*Tang et al., 2021*). - **Information Bottleneck in neural networks
FourierMoE: Fourier Mixture-of-Experts Adaptation of Large Language Models
arXiv:2604.01762v1 Announce Type: new Abstract: Parameter-efficient fine-tuning (PEFT) has emerged as a crucial paradigm for adapting large language models (LLMs) under constrained computational budgets. However, standard PEFT methods often struggle in multi-task fine-tuning settings, where diverse optimization objectives induce task...
Speech LLMs are Contextual Reasoning Transcribers
arXiv:2604.00610v1 Announce Type: new Abstract: Despite extensions to speech inputs, effectively leveraging the rich knowledge and contextual understanding of large language models (LLMs) in automatic speech recognition (ASR) remains non-trivial, as the task primarily involves direct speech-to-text mapping. To address...
Internal Safety Collapse in Frontier Large Language Models
arXiv:2603.23509v1 Announce Type: new Abstract: This work identifies a critical failure mode in frontier large language models (LLMs), which we term Internal Safety Collapse (ISC): under certain task conditions, models enter a state in which they continuously generate harmful content...
The Diminishing Returns of Early-Exit Decoding in Modern LLMs
arXiv:2603.23701v1 Announce Type: new Abstract: In Large Language Model (LLM) inference, early-exit refers to stopping computation at an intermediate layer once the prediction is sufficiently confident, thereby reducing latency and cost. However, recent LLMs adopt improved pretraining recipes and architectures...
CoCR-RAG: Enhancing Retrieval-Augmented Generation in Web Q&A via Concept-oriented Context Reconstruction
arXiv:2603.23989v1 Announce Type: new Abstract: Retrieval-augmented generation (RAG) has shown promising results in enhancing Q&A by incorporating information from the web and other external sources. However, the supporting documents retrieved from the heterogeneous web often originate from multiple sources with...
A Theory of LLM Information Susceptibility
arXiv:2603.23626v1 Announce Type: new Abstract: Large language models (LLMs) are increasingly deployed as optimization modules in agentic systems, yet the fundamental limits of such LLM-mediated improvement remain poorly understood. Here we propose a theory of LLM information susceptibility, centred on...
Lightweight Fairness for LLM-Based Recommendations via Kernelized Projection and Gated Adapters
arXiv:2603.23780v1 Announce Type: new Abstract: Large Language Models (LLMs) have introduced new capabilities to recommender systems, enabling dynamic, context-aware, and conversational recommendations. However, LLM-based recommender systems inherit and may amplify social biases embedded in their pre-training data, especially when demographic...
Manifold Generalization Provably Proceeds Memorization in Diffusion Models
arXiv:2603.23792v1 Announce Type: new Abstract: Diffusion models often generate novel samples even when the learned score is only \emph{coarse} -- a phenomenon not accounted for by the standard view of diffusion training as density estimation. In this paper, we show...
This academic article, "Manifold Generalization Provably Proceeds Memorization in Diffusion Models," delves into the underlying mechanisms of how diffusion models generate novel content, even with "coarse" training data. It suggests that these models capture the geometric structure of data rather than memorizing exact distributions, leading to generalization rather than mere replication. For IP practice, this research is highly relevant to the ongoing debates around copyright infringement and fair use in AI-generated content. The finding that diffusion models generalize from data geometry, rather than memorizing specific inputs, could strengthen arguments that AI outputs are transformative and not direct copies, potentially influencing legal interpretations of derivative works and originality in copyright law. This understanding could also inform policy discussions on data licensing for AI training, as it highlights the models' ability to create new content from generalized patterns rather than exact reproductions.
The paper "Manifold Generalization Provably Proceeds Memorization in Diffusion Models" offers a fascinating theoretical lens through which to understand the generative capabilities of diffusion models, particularly their ability to produce novel outputs even with "coarse" training. This insight has significant implications for intellectual property (IP) practice, particularly in the realm of copyright and inventorship, by refining our understanding of what constitutes "originality" and "creation" in AI-generated content. The core argument – that diffusion models capture the *geometry* of data rather than merely memorizing its *distributional structure* – directly challenges the simplistic notion that AI models are merely sophisticated copy machines. If a model is indeed learning underlying manifold regularities and generating outputs based on these learned geometric principles, rather than reproducing specific training data points, it strengthens the argument for the *originality* of AI-generated works. This theoretical underpinning could influence how courts and IP offices assess copyrightability, potentially shifting the focus from direct input-output comparisons to the sophistication of the generative process and the novelty of the resulting output. **Jurisdictional Comparisons and Implications Analysis:** The implications of this research diverge across jurisdictions, reflecting their varying approaches to AI and IP. * **United States:** The U.S. Copyright Office (USCO) has, to date, taken a relatively restrictive stance, emphasizing the need for human authorship in AI-generated works. The USCO's current guidance suggests that works "produced by a machine
This article's findings regarding diffusion models' ability to generate novel samples from coarse scores, by capturing data geometry rather than fine-scale distribution, has significant implications for patent practitioners in the AI/ML space. **Implications for Practitioners:** 1. **Claim Scope and Enablement (35 U.S.C. § 112):** The concept of "coarse scores capturing the geometry of the data" suggests that claims directed to AI models might be enabled even if they don't explicitly define every fine-grained parameter or training detail. If the core innovation lies in *how* the model learns and leverages data geometry for generalization, rather than precise density estimation, then broad claims focusing on this geometric learning could be defensible. Conversely, if an inventor claims a specific "fine-scale distributional structure," but the underlying model operates on coarse geometric principles, the claim might lack adequate written description or enablement for the *actual* invention. This connects to cases like *Ariad Pharmaceuticals, Inc. v. Eli Lilly and Co.* regarding written description, and *The Medicines Co. v. Hospira, Inc.* on enablement, where the specification must teach one of ordinary skill in the art how to make and use the invention without undue experimentation. 2. **Inventive Step/Non-Obviousness (35 U.S.C. § 103):** The article highlights that this generalization behavior "is a phenomenon not accounted
Can we generate portable representations for clinical time series data using LLMs?
arXiv:2603.23987v1 Announce Type: new Abstract: Deploying clinical ML is slow and brittle: models that work at one hospital often degrade under distribution shifts at the next. In this work, we study a simple question -- can large language models (LLMs)...
Lagrangian Relaxation Score-based Generation for Mixed Integer linear Programming
arXiv:2603.24033v1 Announce Type: new Abstract: Predict-and-search (PaS) methods have shown promise for accelerating mixed-integer linear programming (MILP) solving. However, existing approaches typically assume variable independence and rely on deterministic single-point predictions, which limits solution diversityand often necessitates extensive downstream search...
MedCausalX: Adaptive Causal Reasoning with Self-Reflection for Trustworthy Medical Vision-Language Models
arXiv:2603.23085v1 Announce Type: new Abstract: Vision-Language Models (VLMs) have enabled interpretable medical diagnosis by integrating visual perception with linguistic reasoning. Yet, existing medical chain-of-thought (CoT) models lack explicit mechanisms to represent and enforce causal reasoning, leaving them vulnerable to spurious...
RelayS2S: A Dual-Path Speculative Generation for Real-Time Dialogue
arXiv:2603.23346v1 Announce Type: new Abstract: Real-time spoken dialogue systems face a fundamental tension between latency and response quality. End-to-end speech-to-speech (S2S) models respond immediately and naturally handle turn-taking, backchanneling, and interruption, but produce semantically weaker outputs. Cascaded pipelines (ASR ->...
Detecting Non-Membership in LLM Training Data via Rank Correlations
arXiv:2603.22707v1 Announce Type: new Abstract: As large language models (LLMs) are trained on increasingly vast and opaque text corpora, determining which data contributed to training has become essential for copyright enforcement, compliance auditing, and user trust. While prior work focuses...
PERMA: Benchmarking Personalized Memory Agents via Event-Driven Preference and Realistic Task Environments
arXiv:2603.23231v1 Announce Type: new Abstract: Empowering large language models with long-term memory is crucial for building agents that adapt to users' evolving needs. However, prior evaluations typically interleave preference-related dialogues with irrelevant conversations, reducing the task to needle-in-a-haystack retrieval while...
PersonalQ: Select, Quantize, and Serve Personalized Diffusion Models for Efficient Inference
arXiv:2603.22943v1 Announce Type: new Abstract: Personalized text-to-image generation lets users fine-tune diffusion models into repositories of concept-specific checkpoints, but serving these repositories efficiently is difficult for two reasons: natural-language requests are often ambiguous and can be misrouted to visually similar...
Where Experts Disagree, Models Fail: Detecting Implicit Legal Citations in French Court Decisions
arXiv:2603.22973v1 Announce Type: new Abstract: Computational methods applied to legal scholarship hold the promise of analyzing law at scale. We start from a simple question: how often do courts implicitly apply statutory rules? This requires distinguishing legal reasoning from semantic...