Large Language Models

GPT, Claude, Gemini, and other LLMs

Top This Week

Upwork Launches Hiring App Inside ChatGPT
Llms

Upwork Launches Hiring App Inside ChatGPT

Upwork Launches Hiring App Inside ChatGPT - CDO Magazine

AI Tools & Products · 5 min ·
Open A.I. and Chat GPT are under criminal investigation
Llms

Open A.I. and Chat GPT are under criminal investigation

Open A.I. and Chat GPT are under criminal investigation after a deadly shooting last year at Florida State University.

AI Tools & Products · 1 min ·
Ulta Partners With Google Gemini To Power Agentic AI For Beauty Shoppers
Llms

Ulta Partners With Google Gemini To Power Agentic AI For Beauty Shoppers

Ulta Beauty is integrating Google's Gemini AI into its website and app, and extending its catalog across Google's platforms giving it a c...

AI Tools & Products · 5 min ·

All Content

[2603.03612] Why Are Linear RNNs More Parallelizable?
Llms

[2603.03612] Why Are Linear RNNs More Parallelizable?

Abstract page for arXiv paper 2603.03612: Why Are Linear RNNs More Parallelizable?

arXiv - Machine Learning · 4 min ·
[2603.03371] Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs
Llms

[2603.03371] Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs

Abstract page for arXiv paper 2603.03371: Sleeper Cell: Injecting Latent Malice Temporal Backdoors into Tool-Using LLMs

arXiv - AI · 4 min ·
[2603.03597] NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training
Llms

[2603.03597] NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training

Abstract page for arXiv paper 2603.03597: NuMuon: Nuclear-Norm-Constrained Muon for Compressible LLM Training

arXiv - Machine Learning · 3 min ·
[2603.03538] Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs
Llms

[2603.03538] Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs

Abstract page for arXiv paper 2603.03538: Online Learnability of Chain-of-Thought Verifiers: Soundness and Completeness Trade-offs

arXiv - Machine Learning · 4 min ·
[2603.03535] Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts
Llms

[2603.03535] Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts

Abstract page for arXiv paper 2603.03535: Trade-offs in Ensembling, Merging and Routing Among Parameter-Efficient Experts

arXiv - Machine Learning · 3 min ·
[2603.03352] Perfect score on IPhO 2025 theory by Gemini agent
Llms

[2603.03352] Perfect score on IPhO 2025 theory by Gemini agent

Abstract page for arXiv paper 2603.03352: Perfect score on IPhO 2025 theory by Gemini agent

arXiv - AI · 3 min ·
[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis
Llms

[2603.03527] Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

Abstract page for arXiv paper 2603.03527: Logit-Level Uncertainty Quantification in Vision-Language Models for Histopathology Image Analysis

arXiv - Machine Learning · 4 min ·
[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis
Llms

[2603.03524] Test-Time Meta-Adaptation with Self-Synthesis

Abstract page for arXiv paper 2603.03524: Test-Time Meta-Adaptation with Self-Synthesis

arXiv - AI · 3 min ·
[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery
Llms

[2603.03517] MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

Abstract page for arXiv paper 2603.03517: MMAI Gym for Science: Training Liquid Foundation Models for Drug Discovery

arXiv - AI · 4 min ·
[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations
Llms

[2603.03332] Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

Abstract page for arXiv paper 2603.03332: Fragile Thoughts: How Large Language Models Handle Chain-of-Thought Perturbations

arXiv - AI · 4 min ·
[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts
Llms

[2603.03330] Certainty robustness: Evaluating LLM stability under self-challenging prompts

Abstract page for arXiv paper 2603.03330: Certainty robustness: Evaluating LLM stability under self-challenging prompts

arXiv - AI · 3 min ·
[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness
Llms

[2603.03329] AutoHarness: improving LLM agents by automatically synthesizing a code harness

Abstract page for arXiv paper 2603.03329: AutoHarness: improving LLM agents by automatically synthesizing a code harness

arXiv - AI · 4 min ·
[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees
Llms

[2603.03328] StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

Abstract page for arXiv paper 2603.03328: StructLens: A Structural Lens for Language Models via Maximum Spanning Trees

arXiv - AI · 3 min ·
[2603.03326] Controllable and explainable personality sliders for LLMs at inference time
Llms

[2603.03326] Controllable and explainable personality sliders for LLMs at inference time

Abstract page for arXiv paper 2603.03326: Controllable and explainable personality sliders for LLMs at inference time

arXiv - AI · 3 min ·
[2603.03325] IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference
Llms

[2603.03325] IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference

Abstract page for arXiv paper 2603.03325: IntPro: A Proxy Agent for Context-Aware Intent Understanding via Retrieval-conditioned Inference

arXiv - Machine Learning · 4 min ·
[2603.03324] Controlling Chat Style in Language Models via Single-Direction Editing
Llms

[2603.03324] Controlling Chat Style in Language Models via Single-Direction Editing

Abstract page for arXiv paper 2603.03324: Controlling Chat Style in Language Models via Single-Direction Editing

arXiv - AI · 3 min ·
[2603.03323] Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement
Llms

[2603.03323] Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

Abstract page for arXiv paper 2603.03323: Discern Truth from Falsehood: Reducing Over-Refusal via Contrastive Refinement

arXiv - AI · 3 min ·
[2603.03322] Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery
Llms

[2603.03322] Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Discovery

Abstract page for arXiv paper 2603.03322: Can Large Language Models Derive New Knowledge? A Dynamic Benchmark for Biological Knowledge Di...

arXiv - AI · 4 min ·
[2603.03321] DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following
Llms

[2603.03321] DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

Abstract page for arXiv paper 2603.03321: DIALEVAL: Automated Type-Theoretic Evaluation of LLM Instruction Following

arXiv - AI · 3 min ·
[2603.03389] Towards Improved Sentence Representations using Token Graphs
Llms

[2603.03389] Towards Improved Sentence Representations using Token Graphs

Abstract page for arXiv paper 2603.03389: Towards Improved Sentence Representations using Token Graphs

arXiv - Machine Learning · 4 min ·
Previous Page 228 Next

Related Topics

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime