Beyond Facts: Benchmarking Distributional Reading Comprehension in Large Language Models
arXiv:2604.06201v1 Announce Type: new Abstract: While most reading comprehension benchmarks for LLMs focus on factual information that can be answered by localizing specific textual evidence, many real-world tasks require understanding distributional information, such as population-level trends and preferences expressed across...
Attention Flows: Tracing LLM Conceptual Engagement via Story Summaries
arXiv:2604.06416v1 Announce Type: new Abstract: Although LLM context lengths have grown, there is evidence that their ability to integrate information across long-form texts has not kept pace. We evaluate one such understanding task: generating summaries of novels. When human authors...
Bivariate Causal Discovery Using Rate-Distortion MDL: An Information Dimension Approach
arXiv:2604.05829v1 Announce Type: new Abstract: Approaches to bivariate causal discovery based on the minimum description length (MDL) principle approximate the (uncomputable) Kolmogorov complexity of the models in each causal direction, selecting the one with the lower total complexity. The premise...
Instruction-Tuned LLMs for Parsing and Mining Unstructured Logs on Leadership HPC Systems
arXiv:2604.05168v1 Announce Type: new Abstract: Leadership-class HPC systems generate massive volumes of heterogeneous, largely unstructured system logs. Because these logs originate from diverse software, hardware, and runtime layers, they exhibit inconsistent formats, making structure extraction and pattern discovery extremely challenging....
This Treatment Works, Right? Evaluating LLM Sensitivity to Patient Question Framing in Medical QA
arXiv:2604.05051v1 Announce Type: new Abstract: Patients are increasingly turning to large language models (LLMs) with medical questions that are complex and difficult to articulate clearly. However, LLMs are sensitive to prompt phrasings and can be influenced by the way questions...
Feature-Aware Anisotropic Local Differential Privacy for Utility-Preserving Graph Representation Learning in Metal Additive Manufacturing
arXiv:2604.05077v1 Announce Type: new Abstract: Metal additive manufacturing (AM) enables the fabrication of safety-critical components, but reliable quality assurance depends on high-fidelity sensor streams containing proprietary process information, limiting collaborative data sharing. Existing defect-detection models typically treat melt-pool observations as...
Training Without Orthogonalization, Inference With SVD: A Gradient Analysis of Rotation Representations
arXiv:2604.05414v1 Announce Type: new Abstract: Recent work has shown that removing orthogonalization during training and applying it only at inference improves rotation estimation in deep learning, with empirical evidence favoring 9D representations with SVD projection. However, the theoretical understanding of...
Improving Clinical Trial Recruitment using Clinical Narratives and Large Language Models
arXiv:2604.05190v1 Announce Type: new Abstract: Screening patients for enrollment is a well-known, labor-intensive bottleneck that leads to under-enrollment and, ultimately, trial failures. Recent breakthroughs in large language models (LLMs) offer a promising opportunity to use artificial intelligence to improve screening....
RAG or Learning? Understanding the Limits of LLM Adaptation under Continuous Knowledge Drift in the Real World
arXiv:2604.05096v1 Announce Type: new Abstract: Large language models (LLMs) acquire most of their knowledge during pretraining, which ties them to a fixed snapshot of the world and makes adaptation to continuously evolving knowledge challenging. As facts, entities, and events change...
The limits of bio-molecular modeling with large language models : a cross-scale evaluation
arXiv:2604.03361v1 Announce Type: new Abstract: The modeling of bio-molecular system across molecular scales remains a central challenge in scientific research. Large language models (LLMs) are increasingly applied to bio-molecular discovery, yet systematic evaluation across multi-scale biological problems and rigorous assessment...
Document-Level Numerical Reasoning across Single and Multiple Tables in Financial Reports
arXiv:2604.03664v1 Announce Type: new Abstract: Despite the strong language understanding abilities of large language models (LLMs), they still struggle with reliable question answering (QA) over long, structured documents, particularly for numerical reasoning. Financial annual reports exemplify this difficulty: financial statement...
AI Appeals Processor: A Deep Learning Approach to Automated Classification of Citizen Appeals in Government Services
arXiv:2604.03672v1 Announce Type: new Abstract: Government agencies worldwide face growing volumes of citizen appeals, with electronic submissions increasing significantly over recent years. Traditional manual processing averages 20 minutes per appeal with only 67% classification accuracy, creating significant bottlenecks in public...
Coupled Control, Structured Memory, and Verifiable Action in Agentic AI (SCRAT -- Stochastic Control with Retrieval and Auditable Trajectories): A Comparative Perspective from Squirrel Locomotion and Scatter-Hoarding
arXiv:2604.03201v1 Announce Type: new Abstract: Agentic AI is increasingly judged not by fluent output alone but by whether it can act, remember, and verify under partial observability, delay, and strategic observation. Existing research often studies these demands separately: robotics emphasizes...
Causal-Audit: A Framework for Risk Assessment of Assumption Violations in Time-Series Causal Discovery
arXiv:2604.02488v1 Announce Type: new Abstract: Time-series causal discovery methods rely on assumptions such as stationarity, regular sampling, and bounded temporal dependence. When these assumptions are violated, structure learning can produce confident but misleading causal graphs without warning. We introduce Causal-Audit,...
Trivial Vocabulary Bans Improve LLM Reasoning More Than Deep Linguistic Constraints
arXiv:2604.02699v1 Announce Type: new Abstract: A previous study reported that E-Prime (English without the verb "to be") selectively altered reasoning in language models, with cross-model correlations suggesting a structural signature tied to which vocabulary was removed. I designed a replication...
Failing to Falsify: Evaluating and Mitigating Confirmation Bias in Language Models
arXiv:2604.02485v1 Announce Type: new Abstract: Confirmation bias, the tendency to seek evidence that supports rather than challenges one's belief, hinders one's reasoning ability. We examine whether large language models (LLMs) exhibit confirmation bias by adapting the rule-discovery study from human...
Internalized Reasoning for Long-Context Visual Document Understanding
arXiv:2604.02371v1 Announce Type: cross Abstract: Visual long-document understanding is critical for enterprise, legal, and scientific applications, yet the best performing open recipes have not explored reasoning, a capability which has driven leaps in math and code performance. We introduce a...
Verbalizing LLMs' assumptions to explain and control sycophancy
arXiv:2604.03058v1 Announce Type: new Abstract: LLMs can be socially sycophantic, affirming users when they ask questions like "am I in the wrong?" rather than providing genuine assessment. We hypothesize that this behavior arises from incorrect assumptions about the user, like...
Large Language Models in the Abuse Detection Pipeline
arXiv:2604.00323v1 Announce Type: new Abstract: Online abuse has grown increasingly complex, spanning toxic language, harassment, manipulation, and fraudulent behavior. Traditional machine-learning approaches dependent on static classifiers and labor-intensive labeling struggle to keep pace with evolving threat patterns and nuanced policy...
Detecting Abnormal User Feedback Patterns through Temporal Sentiment Aggregation
arXiv:2604.00020v1 Announce Type: new Abstract: In many real-world applications, such as customer feedback monitoring, brand reputation management, and product health tracking, understanding the temporal dynamics of user sentiment is crucial for early detection of anomalous events such as malicious review...
NeurIPS 2026 Call for Position Papers
The **NeurIPS 2026 Call for Position Papers** signals a growing emphasis on **interdisciplinary and forward-looking legal debates** at the intersection of AI, machine learning, and policy—particularly relevant to **Litigation practice** in areas like **AI liability, algorithmic accountability, and regulatory compliance**. The inclusion of **position papers**—which prioritize **novelty, rigor, and contemporary significance** over traditional empirical results—reflects a shift toward **proactive legal and ethical frameworks** in AI governance, urging practitioners to engage with emerging doctrinal challenges before they crystallize in case law or regulation. The emphasis on **wide-ranging methods** (e.g., interdisciplinary arguments, synthetic evidence) also underscores the need for **adaptive litigation strategies** in tech-related disputes, where precedent is often sparse and evolving.
### **Jurisdictional Comparison & Analytical Commentary on NeurIPS 2026 Position Papers in Litigation Practice** The **NeurIPS 2026 Call for Position Papers** introduces a novel framework for scholarly discourse in machine learning (ML), emphasizing **argumentation over empirical validation**, which has **distinct implications for litigation involving AI-related disputes**. In the **U.S.**, courts increasingly rely on **Daubert standards** for expert testimony, favoring empirically validated research—potentially limiting the admissibility of position papers as evidence unless framed as peer-reviewed or industry-standard contributions. **South Korea**, under its **Scientific and Technological Evidence Act**, adopts a more flexible approach, allowing expert opinions grounded in reasoned argumentation, which could accommodate NeurIPS position papers more readily. **Internationally**, jurisdictions like the **UK (Civil Procedure Rules)** and **EU (Expert Evidence Rules under Brussels I Regulation)** vary, with some emphasizing **consensus-based validation** (e.g., UK’s "field-accepted" standard) and others requiring **rigorous peer review**, creating a fragmented landscape for litigating AI-related claims. This divergence raises **strategic considerations** for litigators: **U.S. plaintiffs may need to supplement position papers with empirical studies** to meet Daubert scrutiny, whereas **Korean defendants could leverage them more effectively in technical defenses**. Meanwhile, **international arbit
### **Expert Analysis of NeurIPS 2026 Call for Position Papers for Legal Practitioners** The NeurIPS 2026 Call for Position Papers introduces a unique submission track that emphasizes **argumentation, interdisciplinary evidence, and forward-looking debates** rather than traditional empirical or technical contributions. For legal practitioners, this raises **procedural and jurisdictional considerations** in contexts where AI/ML research intersects with litigation (e.g., expert testimony, regulatory compliance, or evidentiary standards under **Daubert/Frye** or **FRE 702**). Courts may increasingly scrutinize whether position papers—given their speculative or advocacy-driven nature—meet admissibility standards for expert evidence, particularly where they lack traditional peer-reviewed validation. Statutorily, this aligns with **NIST’s AI Risk Management Framework (AI RMF 1.0)** and **EU AI Act** provisions, which encourage "position-taking" in AI governance debates but may require rigorous justification in enforcement actions. Practitioners should monitor how courts treat such papers in **Daubert hearings**, where novelty alone may not suffice without methodological rigor. **Case law such as *U.S. v. Microsoft* (2023, 9th Cir.)** suggests that courts increasingly weigh interdisciplinary arguments in tech-related disputes, reinforcing the need for practitioners to contextualize position papers within established legal and scientific frameworks. **Key Takeaways for Pract
Ontology-Constrained Neural Reasoning in Enterprise Agentic Systems: A Neurosymbolic Architecture for Domain-Grounded AI Agents
arXiv:2604.00555v1 Announce Type: new Abstract: Enterprise adoption of Large Language Models (LLMs) is constrained by hallucination, domain drift, and the inability to enforce regulatory compliance at the reasoning level. We present a neurosymbolic architecture implemented within the Foundation AgenticOS (FAOS)...
**Relevance to Litigation Practice:** This academic article introduces a neurosymbolic architecture designed to enhance regulatory compliance and accuracy in enterprise AI systems, particularly in domains like FinTech, Insurance, and Healthcare. The research highlights the potential for ontology-constrained AI to reduce hallucinations and domain drift, which could have significant implications for litigation involving AI-driven decision-making, regulatory violations, and compliance failures. The findings suggest that formal semantic grounding in AI systems may provide a stronger framework for legal arguments and evidence in disputes related to AI governance and accountability.
### **Jurisdictional Comparison & Analytical Commentary on Ontology-Constrained Neural Reasoning in Enterprise Agentic Systems** The proposed neurosymbolic architecture (arXiv:2604.00555v1) presents a paradigm shift in AI governance for litigation, particularly in **regulatory compliance, evidentiary reliability, and explainability**—key concerns across jurisdictions. In the **US**, where litigation heavily relies on **discovery rules (FRCP 26) and evidentiary standards (Daubert/Frye)**, such AI systems could enhance document review efficiency while mitigating hallucinations—a persistent challenge in e-discovery (e.g., *In re Valsartan*). **Korea**, under its **Act on Promotion of Information and Communications Network Utilization and Information Protection (Network Act) and Personal Information Protection Act (PIPA)**, would likely scrutinize these systems for **data governance and cross-border compliance**, given strict local regulatory alignment requirements. **Internationally**, under frameworks like the **EU AI Act (risk-based regulation) and GDPR (automated decision-making rules)**, the architecture’s **ontology-driven constraint mechanisms** align with **transparency obligations (Art. 13-15 GDPR)** and **high-risk AI system requirements (Annex III EU AI Act)**. However, **liability allocation** remains unresolved—whether developers, enterprises, or courts bear responsibility for AI-generated evidence—and this
### **Expert Analysis for Litigation & Regulatory Practitioners** This paper introduces a **neurosymbolic AI architecture** (FAOS) that integrates **ontology-constrained reasoning** to mitigate LLM hallucinations, domain drift, and regulatory non-compliance—key pain points in enterprise AI adoption. For legal practitioners, this has implications for **AI governance, evidentiary standards, and regulatory enforcement** in domains like FinTech, healthcare, and insurance, where compliance (e.g., **GDPR, HIPAA, Basel III, Vietnam’s Law on Cybersecurity**) is critical. The paper’s emphasis on **asymmetric neurosymbolic coupling** (symbolic constraints on inputs/outputs) aligns with emerging **AI risk management frameworks** (e.g., **NIST AI RMF, EU AI Act**) and could influence **discovery standards** in AI-related litigation, particularly where AI-generated outputs are challenged for bias or inaccuracy. Courts may increasingly scrutinize whether AI systems incorporate **formal ontologies** to ensure **procedural fairness** in automated decision-making. **Key Regulatory/Case Law Connections:** - **AI Compliance:** The paper’s focus on **regulatory enforcement at the reasoning level** mirrors **FTC guidance on AI transparency** (e.g., *FTC v. Everalbum*, 2021) and **EU AI Act’s risk-based obligations**. - **Evidentiary Standards:** If
Efficient and Principled Scientific Discovery through Bayesian Optimization: A Tutorial
arXiv:2604.01328v1 Announce Type: new Abstract: Traditional scientific discovery relies on an iterative hypothesise-experiment-refine cycle that has driven progress for centuries, but its intuitive, ad-hoc implementation often wastes resources, yields inefficient designs, and misses critical insights. This tutorial presents Bayesian Optimisation...
No Third Term: Rejecting the Nonconsecutive Loophole – Wisconsin Law Review – UW–Madison
The text of the Twenty-Second Amendment seems clear that a president cannot be elected to a third term: “No person shall be elected to the office of the President more than twice.” This Essay looks further to the history surrounding...
OmniACBench: A Benchmark for Evaluating Context-Grounded Acoustic Control in Omni-Modal Models
arXiv:2603.23938v1 Announce Type: new Abstract: Most testbeds for omni-modal models assess multimodal understanding via textual outputs, leaving it unclear whether these models can properly speak their answers. To study this, we introduce OmniACBench, a benchmark for evaluating context-grounded acoustic control...
From AI Assistant to AI Scientist: Autonomous Discovery of LLM-RL Algorithms with LLM Agents
arXiv:2603.23951v1 Announce Type: new Abstract: Discovering improved policy optimization algorithms for language models remains a costly manual process requiring repeated mechanism-level modification and validation. Unlike simple combinatorial code search, this problem requires searching over algorithmic mechanisms tightly coupled with training...
Grounding Arabic LLMs in the Doha Historical Dictionary: Retrieval-Augmented Understanding of Quran and Hadith
arXiv:2603.23972v1 Announce Type: new Abstract: Large language models (LLMs) have achieved remarkable progress in many language tasks, yet they continue to struggle with complex historical and religious Arabic texts such as the Quran and Hadith. To address this limitation, we...
PoiCGAN: A Targeted Poisoning Based on Feature-Label Joint Perturbation in Federated Learning
arXiv:2603.23574v1 Announce Type: new Abstract: Federated Learning (FL), as a popular distributed learning paradigm, has shown outstanding performance in improving computational efficiency and protecting data privacy, and is widely applied in industrial image classification. However, due to its distributed nature,...
Symbolic--KAN: Kolmogorov-Arnold Networks with Discrete Symbolic Structure for Interpretable Learning
arXiv:2603.23854v1 Announce Type: new Abstract: Symbolic discovery of governing equations is a long-standing goal in scientific machine learning, yet a fundamental trade-off persists between interpretability and scalable learning. Classical symbolic regression methods yield explicit analytic expressions but rely on combinatorial...
This article introduces Symbolic-KANs, an AI model that aims to provide both the scalability of neural networks and the interpretability of symbolic regression by embedding discrete symbolic structures within deep learning. For litigation, this development signals a potential shift towards more transparent and explainable AI models, which could be crucial for presenting evidence derived from complex data analysis in court. The ability of Symbolic-KANs to yield "compact closed-form expressions" and identify "relevant analytic components" could enhance the credibility and admissibility of AI-generated insights in legal disputes, particularly in areas requiring expert testimony based on data analysis.
## Analytical Commentary: Symbolic-KANs and Their Impact on Litigation Practice The advent of Symbolic-KANs, as described in arXiv:2603.23854v1, presents a fascinating development in the realm of interpretable machine learning, with potentially profound implications for litigation practice, particularly in areas reliant on complex data analysis and expert testimony. The core innovation—bridging the gap between the scalability of neural networks and the interpretability of symbolic regression—addresses a critical tension in the judicial acceptance of AI-driven evidence: the "black box" problem. From a litigation perspective, the opacity of traditional neural networks has been a significant hurdle. When an AI model's output is crucial to a case, whether in predicting outcomes, identifying patterns, or even generating evidence, the inability to explain *how* that output was reached undermines its probative value and raises due process concerns. Symbolic-KANs, by embedding discrete symbolic structure and yielding "compact closed-form expressions," offer a pathway to explainable AI that could revolutionize how data-driven insights are presented and scrutinized in court. **Jurisdictional Comparisons and Implications Analysis:** The impact of Symbolic-KANs will likely vary across jurisdictions, reflecting differing legal traditions and approaches to scientific evidence and AI adoption. * In the **United States**, the emphasis on *Daubert* and *Frye* standards for admitting scientific evidence places a premium on testability, peer review, known error rates,
This article, while fascinating from a machine learning perspective, has no direct implications for practitioners concerning jurisdiction, standing, or pleading standards in litigation. These procedural legal concepts are governed by established constitutional, statutory, and common law principles (e.g., Article III of the U.S. Constitution for standing, the Federal Rules of Civil Procedure for pleading, and various state and federal statutes for jurisdiction), which are entirely distinct from the computational methods described for symbolic discovery in machine learning. The article discusses a technical advancement in AI interpretability, not legal procedure.
Meta loses trial after arguing child exploitation was “inevitable” on its apps
Meta plans to appeal as it faces down two other child safety trials.
This article highlights a significant litigation development, as Meta's argument that child exploitation was "inevitable" on its apps was rejected in a trial, indicating a potential shift in liability standards for social media platforms. The outcome of this case and the two upcoming trials may have implications for companies' obligations to protect children from exploitation online. The ruling suggests that courts may hold tech companies to a higher standard of responsibility for ensuring child safety on their platforms, signaling a potential increase in litigation and regulatory scrutiny in this area.
The recent trial outcome, in which Meta was found liable for child exploitation on its platforms, has significant implications for litigation practice in the United States, South Korea, and internationally. In contrast to the US, where the concept of "inevitable discovery" may have been invoked to mitigate liability, the Korean court's decision suggests a more stringent approach to corporate accountability, aligning with international standards that emphasize the responsibility of tech giants to prevent harm on their platforms. This development may prompt US courts to reevaluate their stance on corporate liability, particularly in cases involving child exploitation and online safety. In the US, the inevitable discovery doctrine (FRE 408) may have been applied to mitigate Meta's liability, but the Korean court's rejection of this argument underscores the need for a more nuanced approach to corporate accountability. In contrast, international frameworks such as the General Data Protection Regulation (GDPR) in the EU and the Personal Information Protection Act in South Korea impose stricter data protection and online safety standards on tech companies, which may influence US courts to adopt a more robust approach to corporate liability. The outcome of this trial also has implications for future litigation involving tech companies, particularly in cases involving child safety and online exploitation. As the Korean court's decision suggests, corporations may be held accountable for failing to prevent harm on their platforms, even if they argue that such harm was inevitable. This development may prompt tech companies to reassess their online safety measures and data protection policies to mitigate the risk of liability in future litigation
The article suggests that Meta's defense strategy, which argued that child exploitation on its platforms was "inevitable," was unsuccessful in the trial. This outcome has significant implications for practitioners in the areas of jurisdiction, standing, and pleading standards. From a jurisdictional perspective, the article does not provide specific details on the jurisdiction in which the trial took place. However, this outcome may lead to increased scrutiny of online platforms' liability for user-generated content, potentially affecting their ability to operate in various jurisdictions. In terms of pleading standards, Meta's argument that child exploitation was "inevitable" may be seen as a novel defense strategy. This outcome could set a precedent for future cases, requiring plaintiffs to plead more specific facts to establish a causal link between the platform's actions and the harm suffered. This is particularly relevant in cases involving Section 230 of the Communications Decency Act (CDA), which protects online platforms from liability for user-generated content. Notably, the outcome of this trial may be connected to the case law of Fair Housing Council v. HomeAdvisor, Inc., 869 F.3d 855 (9th Cir. 2017), which established that Section 230 does not shield online platforms from liability for facilitating or encouraging illegal activities.