[2603.25112] Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory
Abstract page for arXiv paper 2603.25112: Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory
Open weights models, datasets, and frameworks
Abstract page for arXiv paper 2603.25112: Do LLMs Know What They Know? Measuring Metacognitive Efficiency with Signal Detection Theory
Abstract page for arXiv paper 2603.24772: Evaluating Fine-Tuned LLM Model For Medical Transcription With Small Low-Resource Languages Val...
Abstract page for arXiv paper 2603.25325: How Pruning Reshapes Features: Sparse Autoencoder Analysis of Weight-Pruned Language Models
MiroFlow is an innovative open-source agent framework designed to enhance the performance and robustness of large language models in comp...
The article discusses a humorous attempt to create an auto-complete AI using Family Guy episodes as a database, highlighting the unexpect...
NXP has released a new Linux accelerator driver for their Neutron NPU, enhancing support for machine learning applications and improving ...
IronCurtain is an open-source AI assistant designed to enhance security and control over AI agents, preventing them from executing harmfu...
Mistral AI has partnered with Accenture to enhance enterprise AI adoption by leveraging Mistral's AI models, marking a significant collab...
This article discusses the limitations of current PyTorch schedulers and introduces a flexible suite for scheduling various optimizer hyp...
This article discusses the emulation of FP8 inference on Ampere GPUs, specifically the RTX 3050, using custom Triton kernels to optimize ...
The article discusses Mixture of Experts (MoEs) in Transformer models, highlighting their efficiency and scalability compared to traditio...
PerpetualBooster v1.9.0 introduces significant enhancements to its gradient boosting machine, including built-in causal ML, drift detecti...
SymTorch is a new library that automates the symbolic distillation of deep neural networks, converting them into interpretable mathematic...
The OGD4All framework enhances citizen interaction with geospatial Open Government Data using Large Language Models, achieving high accur...
The EO-1 model is introduced as a unified foundation for general robot control, enhancing multimodal reasoning through a large dataset an...
The article introduces xai-cola, an open-source Python library designed to sparsify counterfactual explanations, enhancing interpretabili...
This study explores the use of small language models for extracting clinical information from low-resource languages, focusing on a priva...
AngelSlim introduces a versatile toolkit for large model compression, integrating advanced algorithms for efficient deployment and improv...
This article discusses the reproduction of Google's Nested Learning/HOPE framework in PyTorch, addressing the lack of available code and ...
This article discusses a lightweight TensorRT implementation of FoundationPose, aimed at improving robotics research by eliminating the h...
A Reddit discussion among ML engineers on effective strategies for learning PyTorch, addressing common challenges like forgetting concept...
The article discusses how users of the OpenClaw AI tool are leveraging an open-source project called Scrapling to bypass anti-bot systems...
Sovereign Mohawk is a Go-based runtime for federated learning that addresses scaling and trust issues, achieving empirical validation for...
Get the latest news, tools, and insights delivered to your inbox.
Daily or weekly digest • Unsubscribe anytime