Large Language Models

GPT, Claude, Gemini, and other LLMs

Top This Week

Claude for Creative Work
Llms

Claude for Creative Work

Anthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.

AI Tools & Products · 5 min ·
Claude AI Goes Down for Thousands of Users Tuesday, Downdetector Shows
Llms

Claude AI Goes Down for Thousands of Users Tuesday, Downdetector Shows

Learn about the recent Claude AI outage affecting users. Read more on the status and issues reported with Claude services.

AI Tools & Products · 6 min ·
People Who Refuse To Use AI & ChatGPT Have 10 Very Rare Traits
Llms

People Who Refuse To Use AI & ChatGPT Have 10 Very Rare Traits

Whether it's because they understand the environmental impact or value their critical thinking skills, people who refuse to use AI and Ch...

AI Tools & Products · 8 min ·

All Content

[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis
Llms

[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

Abstract page for arXiv paper 2603.03527: Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

arXiv - Machine Learning · 4 min ·
[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis
Llms

[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis

Abstract page for arXiv paper 2603.03524: Test-Time Meta-Adaptation with Self-Synthesis

arXiv - AI · 3 min ·
[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery
Llms

[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

Abstract page for arXiv paper 2603.03517: MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

arXiv - AI · 4 min ·
[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations
Llms

[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

Abstract page for arXiv paper 2603.03332: Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

arXiv - AI · 4 min ·
[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts
Llms

[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts

Abstract page for arXiv paper 2603.03330: Certainty robustness: Evaluating LLM stability under self-challenging prompts

arXiv - AI · 3 min ·
[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness
Llms

[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness

Abstract page for arXiv paper 2603.03329: AutoHarness: improving LLM agents by automatically synthesizing a code harness

arXiv - AI · 4 min ·
[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees
Llms

[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

Abstract page for arXiv paper 2603.03328: StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

arXiv - AI · 3 min ·
[2603.03326] Controllable and explainable personality sliders for LLMs at inference time
Llms

[2603.03326] Controllable and explainable personality sliders for LLMs at inference time

Abstract page for arXiv paper 2603.03326: Controllable and explainable personality sliders for LLMs at inference time

arXiv - AI · 3 min ·
[2603.03325] IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference
Llms

[2603.03325] IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference

Abstract page for arXiv paper 2603.03325: IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference

arXiv - Machine Learning · 4 min ·
[2603.03324] Controlling Chat Style in Language Models via Single-Direction Editing
Llms

[2603.03324] Controlling Chat Style in Language Models via Single-Direction Editing

Abstract page for arXiv paper 2603.03324: Controlling Chat Style in Language Models via Single-Direction Editing

arXiv - AI · 3 min ·
[2603.03323] Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement
Llms

[2603.03323] Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

Abstract page for arXiv paper 2603.03323: Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

arXiv - AI · 3 min ·
[2603.03322] Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery
Llms

[2603.03322] Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery

Abstract page for arXiv paper 2603.03322: Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Di...

arXiv - AI · 4 min ·
[2603.03321] DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following
Llms

[2603.03321] DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

Abstract page for arXiv paper 2603.03321: DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

arXiv - AI · 3 min ·
[2603.03389] Towards Improved Sentence Representations using Token Graphs
Llms

[2603.03389] Towards Improved Sentence Representations using Token Graphs

Abstract page for arXiv paper 2603.03389: Towards Improved Sentence Representations using Token Graphs

arXiv - Machine Learning · 4 min ·
[2603.03320] From We to Me: Theory Informed Narrative Shift with Abductive Reasoning
Llms

[2603.03320] From We to Me: Theory Informed Narrative Shift with Abductive Reasoning

Abstract page for arXiv paper 2603.03320: From We to Me: Theory Informed Narrative Shift with Abductive Reasoning

arXiv - AI · 3 min ·
[2603.03319] Automated Concept Discovery for LLM-as-a-Judge Preference Analysis
Llms

[2603.03319] Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

Abstract page for arXiv paper 2603.03319: Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

arXiv - AI · 4 min ·
[2603.03378] AOI: Turning Failed Trajectories into Training Signals for Autonomous Cloud Diagnosis
Llms

[2603.03378] AOI: Turning Failed Trajectories into Training Signals for Autonomous Cloud Diagnosis

Abstract page for arXiv paper 2603.03378: AOI: Turning Failed Trajectories into Training Signals for Autonomous Cloud Diagnosis

arXiv - Machine Learning · 4 min ·
[2603.03318] Quantum-Inspired Self-Attention in a Large Language Model
Llms

[2603.03318] Quantum-Inspired Self-Attention in a Large Language Model

Abstract page for arXiv paper 2603.03318: Quantum-Inspired Self-Attention in a Large Language Model

arXiv - AI · 3 min ·
[2603.03314] Towards Self-Robust LLMs: Intrinsic Prompt Noise Resistance via CoIPO
Llms

[2603.03314] Towards Self-Robust LLMs: Intrinsic Prompt Noise Resistance via CoIPO

Abstract page for arXiv paper 2603.03314: Towards Self-Robust LLMs: Intrinsic Prompt Noise Resistance via CoIPO

arXiv - Machine Learning · 4 min ·
[2603.03313] How does fine-tuning improve sensorimotor representations in large language models?
Llms

[2603.03313] How does fine-tuning improve sensorimotor representations in large language models?

Abstract page for arXiv paper 2603.03313: How does fine-tuning improve sensorimotor representations in large language models?

arXiv - AI · 3 min ·
Previous Page 258 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime