女大危矣!

清華和騰訊碼工的AI新作CALM:直接顛覆了現有LLM

Robert Youssef

 

@rryssf_

 

Holy shit... this might be the next big paradigm shift in AI.

 

 

 

Tencent + Tsinghua just dropped a paper called Continuous Autoregressive Language Models (CALM) and it basically kills the “next-token” paradigm every LLM is built on.

 

 

 

Instead of predicting one token at a time, CALM predicts continuous vectors that represent multiple tokens at once.

 

 

 

Meaning: the model doesn’t think “word by word”… it thinks in ideas per step.

 

 

 

Here’s why that’s insane

 

 

 

→ 4× fewer prediction steps (each vector = ~4 tokens)

 

→ 44% less training compute

 

→ No discrete vocabulary pure continuous reasoning

 

→ New metric (BrierLM) replaces perplexity entirely

 

 

 

They even built a new energy-based transformer that learns without softmax no token sampling, no vocab ceiling.

 

 

 

It’s like going from speaking Morse code… to streaming full thoughts.

 

 

 

If this scales, every LLM today is obsolete.

 

 

所有跟帖: 

論文到落地可能是一步可能是正無窮 -食神OG- 給 食神OG 發送悄悄話 (0 bytes) () 11/04/2025 postreply 12:50:00

噓! 快去買點撲 -FBE63- 給 FBE63 發送悄悄話 (0 bytes) () 11/04/2025 postreply 12:50:56

這就是為什麽open AI現在急於變現。 -TalkToMi- 給 TalkToMi 發送悄悄話 (0 bytes) () 11/04/2025 postreply 12:53:48

本來嘛,算法的進步才是最重要的,以為piling on compute就能work的人估計算法方麵不靈 -過來人2- 給 過來人2 發送悄悄話 (0 bytes) () 11/04/2025 postreply 12:59:47

FOMO -大好時光- 給 大好時光 發送悄悄話 (0 bytes) () 11/04/2025 postreply 13:04:37

對OpenAI也許有影響,但是目前對NVDA還沒有,因為隻要AI scaling laws沒有失效,算力就永遠短缺~ -ocliving2005_4ever- 給 ocliving2005_4ever 發送悄悄話 (0 bytes) () 11/04/2025 postreply 13:09:34

你的無知無與倫比 -cnrhm2017- 給 cnrhm2017 發送悄悄話 cnrhm2017 的博客首頁 (0 bytes) () 11/04/2025 postreply 13:13:08

請您先登陸,再發跟帖!