Sure, running an LLM is cheaper, but the way we use LLMs now requires way more tokens than last year.
10x more tokens today cost less than than half of X tokens from ~mid 2024.
ok but the capabilities are also rising. what point are you trying to make?
10x more tokens today cost less than than half of X tokens from ~mid 2024.