[논문리뷰] Can LLMs Predict Their Own Failures? Self-Awareness via Internal CircuitsarXiv에 게시된 'Can LLMs Predict Their Own Failures? Self-Awareness via Internal Circuits' 논문에 대한 자세한 리뷰입니다.#Review#LLM Self-Awareness#Failure Prediction#Internal States#Attention Mechanisms#Neural Network Probes#Computational Efficiency#Zero-Shot Transfer2026년 1월 5일댓글 수 로딩 중
[논문리뷰] Large Language Models Do NOT Really Know What They Don't KnowarXiv에 게시된 'Large Language Models Do NOT Really Know What They Don't Know' 논문에 대한 자세한 리뷰입니다.#Review#LLMs#Hallucination Detection#Mechanistic Interpretability#Internal States#Knowledge Recall#Refusal Tuning#Factual Associations#Associated Hallucinations2025년 10월 17일댓글 수 로딩 중