Gate Booster 第 4 期:發帖瓜分 1,500 $USDT
🔹 發布 TradFi 黃金福袋原創內容,可得 15 $USDT,名額有限先到先得
🔹 本期支持 X、YouTube 發布原創內容
🔹 無需複雜操作,流程清晰透明
🔹 流程:申請成為 Booster → 領取任務 → 發布原創內容 → 回鏈登記 → 等待審核及發獎
📅 任務截止時間:03月20日16:00(UTC+8)
立即領取任務:https://www.gate.com/booster/10028?pid=allPort&ch=KTag1BmC
更多詳情:https://www.gate.com/announcements/article/50203
Since o1 launched, the biggest complaint is that it's "too verbose."
I just wanted to fix a simple bug, and it gave me three background explanations, two solution approaches plus error handling, and then wished me good luck on top of that.
I was only looking for a spelling mistake on line 12, but ended up having to review Python naming conventions all over again.
This blame falls squarely on RLHF. Annotators tend to give higher scores to longer responses, thinking more text looks more professional.
So the model desperately piles up "seemingly useful" filler, while the actual core information gets diluted.
Look at Claude next door—it's much more sensible about this, knowing what length matches what question.
The most painful part is the wallet: o1's output pricing is $60/1M tokens. For something that should take 100 tokens to explain, it deliberately pads it to 500, multiplying costs by five on the spot.
Now when asking questions you have to specifically add "code only," and even that doesn't always work.
The model's current state is: genius-level IQ, but EQ completely offline—it simply doesn't know when to shut up.