That is unfortunate but they do present some theoretical insights about scaling context length and probably a more efficient way to do RL. Even knowledge about it can have an effect on next iterations from other labs.