this is smart, but I think NVIDIA's paper on fine tuning small language models presents a sightly more efficient approach