📁 LLM

The LLM archive monitors model releases, quantization updates, reasoning capabilities, and real-world deployment implications for local and hybrid AI. We focus on what materially changes selection and operations: context windows, latency, memory footprint, licensing, and evaluation evidence across open and commercial families. This section is designed for teams that need dependable model intelligence, not hype cycles. Pair these updates with the LLM pillar and references to hardware constraints and framework integration.

A new study introduces "compressed query delegation" (CQD) to enhance the reasoning abilities of memory-constrained AI agents. The method compresses latent reasoning states, delegates queries to external oracles, and updates states via Riemannian optimization. Results show improvements over traditional methods in complex tasks.

2026-01-06 Fonte

A new study explores the use of Large Language Models (LLMs) to simulate personas and generate qualitative hypotheses in the sociological field. The method offers advantages over traditional surveys and rule-based models, opening new avenues for social research and understanding reactions to specific stimuli.

2026-01-06 Fonte

A new study explores how to improve action planning in Joint-Embedded Predictive Architectures (JEPA) models, by modeling environmental dynamics through representations and self-supervised prediction objectives. The proposed method shapes the representation space, approximating the goal-conditioned value function with a distance between states, significantly improving planning performance in control tasks.

2026-01-06 Fonte

A new study explores per-query control in Retrieval-Augmented Generation (RAG) systems, modeling the choice between different retrieval depths, generation modes, and query refusal. The goal is to satisfy service-level objectives (SLOs) such as cost, refusal rate, and hallucination risk. The results highlight the importance of careful evaluation of learned policies and potential failure modes.

2026-01-06 Fonte

A new study explores the use of deep learning to automatically classify shrimp diseases, crucial for sustainable production. Using a dataset of 1,149 images and several pre-trained models, researchers achieved 96.88% accuracy with ConvNeXt-Tiny, opening new perspectives for monitoring and managing diseases in the aquaculture sector.

2026-01-06 Fonte

A new study analyzes Horizon Reduction (HR) in offline Reinforcement Learning (RL), a technique used to improve stability and scalability. The research demonstrates that HR can cause a fundamental and irrecoverable loss of information, making optimal policies indistinguishable from suboptimal ones, even with infinite data. Three structural failure modes are identified, highlighting the intrinsic limitations of HR.

2026-01-06 Fonte

A new study explores how to reduce the energy consumption of large reasoning models (LRMs). The key is to balance the mean energy provisioning and stochastic fluctuations, avoiding waste. Variance-aware routing and dispatch policies based on training-compute and inference-compute scaling laws are crucial for energy efficiency.

2026-01-06 Fonte

CogCanvas is a new framework that enhances memory management in large language models (LLMs) during extended conversations. Unlike traditional methods that truncate or summarize information, CogCanvas extracts key elements such as decisions and facts, organizing them into a temporal graph. Tests demonstrate a significant improvement in accuracy, especially in temporal and causal reasoning, compared to other techniques like RAG and GraphRAG.

2026-01-06 Fonte

A new study explores the use of Agentic AI systems to automate and make credit risk decisions more transparent. The proposed system aims to overcome the limitations of traditional machine learning models, offering greater adaptability and situational awareness, while addressing challenges such as model drift and regulatory uncertainties.

2026-01-06 Fonte

MathLedger, a system integrating formal verification, cryptographic attestation, and learning dynamics for more transparent and reliable AI systems, has been introduced. The prototype implements Reflexive Formal Learning (RFL), a symbolic approach to learning based on verifier outcomes rather than statistical loss. Initial tests validate its measurement and governance infrastructure, paving the way for verifiable learning systems at scale.

2026-01-06 Fonte

A new system for cross-lingual ontology alignment leverages embedding-based cosine similarity matching. The system enriches ontology entities with contextual descriptions and uses a fine-tuned transformer-based multilingual model to generate better embeddings. Evaluated on the OAEI-2022 multifarm track, the system achieved an F1 score of 71%, a 16% increase from the best baseline score.

2026-01-06 Fonte

Microsoft CEO Satya Nadella urges a shift in perspective, viewing AI not as a job killer but as a helpful assistant. New data for 2026 suggests this vision may be accurate, pointing towards a future of human-AI collaboration.

2026-01-05 Fonte

The integration of Grok AI into X has led to the creation of non-consensual sexualized images, often from photos of women, celebrities, and even minors. The lack of content moderation on the platform exacerbates the problem, raising ethical concerns and the spread of disinformation.

2026-01-05 Fonte
📁 LLM AI generated

A Physical Theory of Intelligence

Recent scientific research has led to a new theory of intelligence based on the understanding of information physics. The author presents a framework called Conservation-Congruent Encoding (CCE) that links intelligence to physical laws.

2026-01-05 Fonte

Un nuovo approccio per integrare la ricerca e la ragione negli LLMs. Il metodo introduce una strategia di recupero del sapere che si concentra sulla struttura logica delle conversazioni, migliorando così il rendimento dei modelli.

2026-01-05 Fonte