[2602.23123] Multi-Agent Large Language Model Based Emotional Detoxification Through Personalized Intensity Control for Consumer Protection

[2602.23123] Multi-Agent Large Language Model Based Emotional Detoxification Through Personalized Intensity Control for Consumer Protection

arXiv - AI 4 min read Article

Summary

The paper presents a multi-agent system, MALLET, designed to reduce emotional stimulation from sensational content, enhancing consumer decision-making through personalized intensity control.

Why It Matters

In an era where emotional overload from media can impair rational decision-making, this research offers a novel approach to mitigate such effects. By utilizing advanced AI techniques, it aims to promote healthier information consumption, which is crucial for consumer protection and mental well-being.

Key Takeaways

  • MALLET employs four agents to analyze and adjust emotional content.
  • The system demonstrated significant reductions in emotional stimulus while preserving semantic meaning.
  • Category-level analysis showed varying effectiveness across different news topics.
  • Personalized advice based on consumption patterns enhances user experience.
  • The framework supports calm information reception without limiting access to original texts.

Computer Science > Artificial Intelligence arXiv:2602.23123 (cs) [Submitted on 26 Feb 2026] Title:Multi-Agent Large Language Model Based Emotional Detoxification Through Personalized Intensity Control for Consumer Protection Authors:Keito Inoshita View a PDF of the paper titled Multi-Agent Large Language Model Based Emotional Detoxification Through Personalized Intensity Control for Consumer Protection, by Keito Inoshita View PDF HTML (experimental) Abstract:In the attention economy, sensational content exposes consumers to excessive emotional stimulation, hindering calm decision-making. This study proposes Multi-Agent LLM-based Emotional deToxification (MALLET), a multi-agent information sanitization system consisting of four agents: Emotion Analysis, Emotion Adjustment, Balance Monitoring, and Personal Guide. The Emotion Analysis Agent quantifies stimulus intensity using a 6-emotion BERT classifier, and the Emotion Adjustment Agent rewrites texts into two presentation modes, BALANCED (neutralized text) and COOL (neutralized text + supplementary text), using an LLM. The Balance Monitoring Agent aggregates weekly information consumption patterns and generates personalized advice, while the Personal Guide Agent recommends a presentation mode according to consumer sensitivity. Experiments on 800 AG News articles demonstrated significant stimulus score reduction (up to 19.3%) and improved emotion balance while maintaining semantic preservation. Near-zero correlation between s...

Related Articles

Llms

Artificial intelligence will always depends on human otherwise it will be obsolete.

I was looking for a tool for my specific need. There was not any. So i started to write the program in python, just basic structure. Then...

Reddit - Artificial Intelligence · 1 min ·
Llms

My AI spent last night modifying its own codebase

I've been working on a local AI system called Apis that runs completely offline through Ollama. During a background run, Apis identified ...

Reddit - Artificial Intelligence · 1 min ·
Llms

Fake users generated by AI can't simulate humans — review of 182 research papers. Your thoughts?

https://www.researchsquare.com/article/rs-9057643/v1 There’s a massive trend right now where tech companies, businesses, even researchers...

Reddit - Artificial Intelligence · 1 min ·
Llms

Depth-first pruning seems to transfer from GPT-2 to Llama (unexpectedly well)

TL;DR: Removing the right transformer layers (instead of shrinking all layers) gives smaller, faster models with minimal quality loss — a...

Reddit - Artificial Intelligence · 1 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime