đź’¬ Community

Discussion · 20 items

GitHub to Use User Data for AI Training by Default

submitted by /u/i-drake [link] [comments]

·

Introducing TRIBE v2: A Predictive Foundation Model Trained to Understand How the Human Brain Processes Complex Stimuli

"Understanding how the human brain processes the world around us is one of the greatest open challenges in neuroscience. Breakthroughs here could transform how we understand and treat neurological conditions affecting hundreds of millions of people — and improve AI systems by directly guiding their development from neuroscientific principles. Today, we're announcing TRIBE v2: our first AI model of human brain responses to sights, sounds, and language. Building on our Algonauts 2025 award-winn...

·

AI system learns to prevent warehouse robot traffic jams, boosting throughput 25%

"Inside a giant autonomous warehouse, hundreds of robots dart down aisles as they collect and distribute items to fulfill a steady stream of customer orders. In this busy environment, even small traffic jams or minor collisions can snowball into massive slowdowns. To avoid such an avalanche of inefficiencies, researchers from MIT and the tech firm Symbotic developed a new method that automatically keeps a fleet of robots moving smoothly. Their method learns which robots should go first at eac...

·

Could factories run faster and greener? How AI 'digital twins' reshape production

Researchers at Örebro University have developed a new production system that uses artificial intelligence (AI) to improve efficiency and sustainability across industries such as automotive manufacturing. The research is published in the journal IOP Conference Series: Materials Science and Engineering. "Our results show that production can become both faster and more sustainable at the same time," says Rajesh Patil, researcher in mechanical engineering. Together with Professor Magnus Löfstrand...

·

AI agent accelerates catalyst discovery for sustainable fuel development

A multi-institutional team based in China recently used AI to identify a key characteristic of compounds called catalysts that are used to initiate and speed up the chemical reactions that convert carbon dioxide into molecules that can be used to develop sustainable fuels. The team then used the AI—dubbed Catalysis AI Agent—to guide their catalyst designs, ultimately discovering the universal design principle for copper-based single-atom alloy (SAAs) catalysts. They published their results on...

·

How to see through the opaqueness of pricing of tokens?

I was reflecting on this after reading articles like these The rise of China’s hottest new commodity: AI tokens More! More! More! Tech Workers Max Out Their A.I. Use. (NYT Paywall) While conceptually a "unit," the pricing of Tokens is all over the place. Almost every 'AI service' provider provides a Freemium model where you sign up and get a few tokens and max it out with a couple of queries, prompting you to buy a plan that gives "x or y Tokens.' And the pricing is all over the place. How to...

·

[D] On conferences and page limitations

What is your opinion on long appendices in conference papers? I am observing that appendix lengths in conference papers (ICML, NeurIPS, etc.) are getting longer and longer, and in some fields they are now basically the standard and a central part of the paper. From my point of view, this is becoming a bit problematic. I have many times been asked to add more experiments which, in order to be included, require several extra pages beyond the main 8–10 pages. This effectively makes the appendix ...

·

Judge rejects Pentagon's attempt to 'cripple' Anthropic

submitted by /u/esporx [link] [comments]

·

[R] Which place should I commit to ACL SRW or ICML workshop or AACL?

Hello everyone, I got ARR review set on March 12 with submitted paper. OA 3, 2.5, 2.5 and 2. Meta review is 2.5 the harsh (2) guy criticised the most but he overused LLM so around 4 times he made mistakes (wrong facts) in his reviews. However, generally the 2.5 guys are also show agreements in incremental work/novelty. Actually this is the revised submission (after October cycle last year), the topic moved too fast and I think my work would soon become outdated. with metareview 2.5, I chose n...

·

Retraining vs Fine-tuning or Transfer Learning? [D]

Hi! I am currently working on a project that is basically an e-commerce clickstream data. We take in data, find the intent of the user(XGboost) and price sensitivity(Xgboost), segregate the user in different segments based on their purchasing intent or their research or price behaviour(Xgboost), recommend the benefit like discount or free shipping(Linucp or Thompson sampling), etc. My question is this - when the data comes in daily to train our models, is it better to retrain the models from ...

·

Community Building

I made 3 repos public and in a week I have a total of 16 stars and 5 forks. I realize that the platforms are extremely complex and definitely not for casual coders. But I think even they could find something useful. Sadly, I have no idea how to build a community. Any advice would be appreciated. submitted by /u/Sure_Excuse_8824 [link] [comments]

·

The Silver Alien Head

Here is a piece I’ve been working on that ties into my experiential grooves theory. submitted by /u/AbleYak9996 [link] [comments]

·

CodexLib — compressed knowledge packs any AI can ingest instantly (100+ packs, 50 domains, REST API)

I built CodexLib (https://codexlib.io) — a curated repository of 100+ deep knowledge bases in compressed, AI-optimized format. The idea: instead of pasting long documents into your context window, you use a pre-compressed knowledge pack with a Rosetta decoder header. The AI decompresses it on the fly, and you get the same depth at ~15% fewer tokens. Each pack covers a specific domain (quantum computing, cardiology, cybersecurity, etc.) with abbreviations like ML=Machine Learning, NN=Neural Ne...

·

Claude's system prompt + XML tags is the most underused power combo right now

Most people just type into ChatGPT like it's Google. Claude with a structured system prompt using XML tags behaves like a completely different tool. Example system prompt: <role>You are a senior equity analyst</role> <task>Analyse this earnings transcript and extract: 1) forward guidance tone 2) margin surprises 3) management deflections</task> <output>Return as structured JSON</output> Then paste the entire earnings call transcript. You get institutional-g...

·

Ridiculous. Anthropic is behaving exactly like OpenAI.

Claude was fantastic when I paid monthly, right up until I chose to commit to a yearly Pro subscription. Now, a mere thirty-four text prompts—mostly two or three sentences long—burn through 94% of my five-hour limit. To make matters worse, six of those prompts were wasted because I had to repeat what I had just stated. Claude kept pulling web calls for information already established one or two prompts earlier. This is machinery designed to eat your usage. This is the exact same bait-and-swit...

·

[R] Interested in recent research into recall vs recognition in LLMs

I've casually seen LLMs correctly verify exact quotations that they either couldn't or wouldn't quote directly for me. I'm aware that they're trained to avoid quoting potentially copywritten content, and the implications of that, but it made me wonder a few things: Can LLMs verify knowledge more (or less) accurately than they can recall knowledge? 1b. Can LLMs verify more (or less) knowledge accurately than they can recall accurately? What research exists into LLM accuracy in recalling facts ...

·

Pretrained ADAM v2 weights [D]

Hi everyone, I'm a master's student working on anatomy-aware unsupervised anomaly detection in chest X-rays. My thesis uses ADAM v2 (Autodidactic Dense Anatomical Model v2) from the paper "Representing Part-Whole Hierarchies in Foundation Models by Learning Localizability, Composability and Decomposability from Anatomy via Self Supervision" by Taher et al., CVPR 2024. I need the pretrained ConvNeXt-B weights from this model to use as a feature extractor for my downstream anomaly detection tas...

·

[D] Why evaluating only final outputs is misleading for local LLM agents

Been running local agents with Ollama + LangChain lately and noticed something kind of uncomfortable — you can get a completely correct final answer while the agent is doing absolute nonsense internally. I’m talking about stuff like calling the wrong tool first and then “recovering,” using tools it didn’t need at all, looping a few times before converging, or even getting dangerously close to calling something it shouldn’t. And if you’re only checking the final output, all of that just… passe...

·

[D] - 1M tokens/second serving Qwen 3.5 27B on B200 GPUs, benchmark results and findings

Wrote up the process of pushing Qwen 3.5 27B (dense, FP8) to 1.1M total tok/s on 96 B200 GPUs with vLLM v0.18.0. DP=8 nearly 4x'd throughput over TP=8. Model is too small for tensor parallelism to help on B200s. MTP-1 mattered more than anything else (GPU utilization was 0% without it). MTP-5 crashed with cudaErrorIllegalAddress. 97.1% scaling efficiency at 8 nodes, 96.5% at 12. TPOT flat at ~46ms regardless of node count. Inference Gateway (KV-cache-aware routing) added ~35% overhead vs Clus...

·

New Project - 3D + AI - Animation

running a pipeline between Blender - Unreal Engine - the chat + Kling IA - Im pretty happy with this work, should still work on more consistency, let me know what you think submitted by /u/Commercial-Army-5843 [link] [comments]

·

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime