Translate

Visualizzazione post con etichetta LLMs. Mostra tutti i post
Visualizzazione post con etichetta LLMs. Mostra tutti i post

lunedì 7 aprile 2025

# life: detecting hallucinations (in large language models) using semantic entropy.

<< Large language model (LLM) systems, such as ChatGPT or Gemini, can show impressive reasoning and question-answering capabilities but often ‘hallucinate’ false outputs and unsubstantiated answers. >>

<< Here (AA) develop new methods grounded in statistics, proposing entropy-based uncertainty estimators for LLMs to detect a subset of hallucinations— confabulations— which are arbitrary and incorrect generations. (Their) method addresses the fact that one idea can be expressed in many ways by computing uncertainty at the level of meaning rather than specific sequences of words. >>

Their method << works across datasets and tasks without a priori knowledge of the task, requires no task-specific data and robustly generalizes to new tasks not seen before. By detecting when a prompt is likely to produce a confabulation, helps users understand when they must take extra care with LLMs and opens up new possibilities for using LLMs that are otherwise prevented by their unreliability. >>️️

Sebastian Farquhar, Jannik Kossen, et al. Detecting hallucinations in large language models using semantic entropy. Nature 630, 625–630. Jun 19, 2024.

Also: ai (artificial intell) (bot), entropy, in https://www.inkgmr.net/kwrds.html 

Keywords: life, artificial intelligence,  LLMs, confabulations, uncertainty, hallucinations, entropy, semantic entropy

giovedì 20 marzo 2025

# aibot: I think, therefore I hallucinate: minds, machines, and the art of being wrong.

<< This theoretical work examines 'hallucinations' in both human cognition and large language models, comparing how each system can produce perceptions or outputs that deviate from reality. Drawing on neuroscience and machine learning research, (AA) highlight the predictive processes that underlie human and artificial thought. >>

<< In humans, complex neural mechanisms interpret sensory information under uncertainty, sometimes filling in gaps and creating false perceptions. This inference occurs hierarchically: higher cortical levels send top-down predictions to lower-level regions, while mismatches (prediction errors) propagate upward to refine the model. LLMs, in contrast, rely on auto-regressive modeling of text and can generate erroneous statements in the absence of robust grounding. >>

<< Despite these different foundations - biological versus computational - the similarities in their predictive architectures help explain why hallucinations occur. (AA) propose that the propensity to generate incorrect or confabulated responses may be an inherent feature of advanced intelligence. In both humans and AI, adaptive predictive processes aim to make sense of incomplete information and anticipate future states, fostering creativity and flexibility, but also introducing the risk of errors. (Their) analysis illuminates how factors such as feedback, grounding, and error correction affect the likelihood of 'being wrong' in each system. (AA) suggest that mitigating AI hallucinations (e.g., through improved training, post-processing, or knowledge-grounding methods) may also shed light on human cognitive processes, revealing how error-prone predictions can be harnessed for innovation without compromising reliability. By exploring these converging and divergent mechanisms, the paper underscores the broader implications for advancing both AI reliability and scientific understanding of human thought. >>️

Sebastian Barros. I Think, Therefore I Hallucinate: Minds, Machines, and the Art of Being Wrong. arXiv: 2503.05806v1 [q-bio.NC]. 4 Mar 4, 2025.

Also: brain, curiosity, novelty, uncertainty, error, mistake, jazz, ai (artificial intell), in https://www.inkgmr.net/kwrds.html 

Keywords: brain, cognition, perceptions, curiosity, novelty, hallucinations, errors, prediction, prediction errors, error-prone predictions, AI, artificial intelligence, LLMs

giovedì 6 febbraio 2025

# life: chameleon machines

<< Large language model-based (LLM-based) agents have become common in settings that include non-cooperative parties. In such settings, agents' decision-making needs to conceal information from their adversaries, reveal information to their cooperators, and infer information to identify the other agents' characteristics. To investigate whether LLMs have these information control and decision-making capabilities, (AA) make LLM agents play the language-based hidden-identity game, The Chameleon. >>️

<< Based on the empirical results and theoretical analysis of different strategies, (AA) deduce that LLM-based non-chameleon agents reveal excessive information to agents of unknown identities. (Their) results point to a weakness of contemporary LLMs, including GPT-4, GPT-4o, Gemini 1.5, and Claude 3.5 Sonnet, in strategic interactions. >>
Mustafa O. Karabag, Ufuk Topcu. Do LLMs Strategically Reveal, Conceal, and Infer Information? A Theoretical and Empirical Analysis in The Chameleon Game. arXiv: 2501.19398v1 [cs.AI]. Jan 31, 2025.

Also: games, ai (artificial intell), nfulaw, in https://www.inkgmr.net/kwrds.html 

Keywords: life, games, chameleon game, ai, artificial intelligence, LLMs, privacy, nfulaw


sabato 25 gennaio 2025

# life: the Age of hallucinatory artificial intelligence (AI); the beginning.

<< It’s well known that all kinds of generative AI, including the large language models (LLMs) behind AI chatbots, make things up. This is both a strength and a weakness. It’s the reason for their celebrated inventive capacity, but it also means they sometimes blur truth and fiction, inserting incorrect details into apparently factual sentences. >>

<< They sound like politicians, they tend to make up stuff and be totally confident no matter what. >> Santosh Vempala. ️

<< Chatbots err for many reasons, but computer scientists tend to refer to all such blips as hallucinations. It’s a term not universally accepted, with some suggesting ‘confabulations’ or, more simply, ‘bullshit’. The phenomenon has captured so much attention that the website Dictionary.com picked ‘hallucinate’ as its word of the year for 2023. >>️

<< Because AI hallucinations are fundamental to how LLMs work, researchers say that eliminating them completely is impossible. >>️

Nicola Jones. AI hallucinations can’t be stopped — but these techniques can limit their damage. Nature. 637, 778-780. Jan 21,  2025. 

Also: ai (artificial intell) (bot), nfulaw, in https://www.inkgmr.net/kwrds.html 

Keywords: life, ai, artificial intell, LLMs, bot, nfulaw