logoalt Hacker News

runeblazeyesterday at 2:09 AM1 replyview on HN

Is it though? There is a reason gpt has codex variants. RL on a specific task raises the performance on that task


Replies

jjmarryesterday at 2:30 AM

Post-training doesn't transfer over when a new base model arrives so anyone who adopted a task-specific LLM gets burned when a new generational advance comes out.

show 1 reply