Ana de Prado's picture
2 17

Ana de Prado

AnaDP

AI & ML interests

None yet

Recent Activity

Organizations

Intelygenz's profile picture

AnaDP's activity

upvoted 2 articles 4 days ago
view article
Article

Introduction to 3D Gaussian Splatting

โ€ข 46
view article
Article

Hunyuan video LoRA training study (Single image/style training)

By neph1 โ€ข
โ€ข 5
reacted to vladbogo's post with โค๏ธ 12 months ago
view post
Post
xAI releases the weights for Grok-1. Apparently it's a 314B MoE with 25% of the weights active on a given token.

Blog: https://x.ai/blog/grok-os
Code: https://github.com/xai-org/grok
Model: xai-org/grok-1
Weights: magnet:?xt=urn:btih:5f96d43576e3d386c9ba65b883210a393b68210e&tr=https%3A%2F%2Facademictorrents.com%2Fannounce.php&tr=udp%3A%2F%2Ftracker.coppersurfer.tk%3A6969&tr=udp%3A%2F%2Ftracker.opentrackr.org%3A1337%2Fannounce
  • 2 replies
ยท
reacted to santiviquez's post with ๐Ÿ‘ about 1 year ago
view post
Post
Eigenvalues to the rescue? ๐Ÿ›Ÿ๐Ÿค”

I found out about this paper thanks to @gsarti 's post from last week; I got curious, so I want to post my take on it. ๐Ÿค—

The paper proposes a new metric called EigenScore to detect LLM hallucinations. ๐Ÿ“„

Their idea is that given an input question, they generate K different answers, take their internal embedding states, calculate a covariance matrix with them, and use it to calculate an EigenScore.

We can think of the EigenScore as the mean of the eigenvalues of the covariance matrix of the embedding space of the K-generated answers.

โ“But why eigenvalues?

Well, if the K generations have similar semantics, the sentence embeddings will be highly correlated, and most eigenvalues will be close to 0.

On the other hand, if the LLM hallucinates, the K generations will have diverse semantics, and the eigenvalues will be significantly different from 0.

The idea is pretty neat and shows better results when compared to other methods like sequence probabilities, length-normalized entropy, and other uncertainty quantification-based methods.

๐Ÿ’ญ What I'm personally missing from the paper is that they don't compare their results with other methods like LLM-Eval and SelfcheckGPT. They do mention that EigenScore is much cheaper to implement than SelfcheckGPT, but that's all on the topic.

Paper: INSIDE: LLMs' Internal States Retain the Power of Hallucination Detection (2402.03744)