In a single-task setting (the situation described in the intro) throughput and latency are just the inverse of one another (in the mathematical sense of “inverse”: throughput = nb task per seconds = 1/time taken to process the task = 1/latency).
They only diverge when you consider multiple tasks.
That’s not the way “latency” is commonly used in my experience.
Latency numbers always include queuing time - so the measures are not related or derivable from each other.
A process might have a throughput of 1 million jobs per second but if the average size of the queue is 10 million then your job latency is going to be 10 seconds on average and not 1 microsecond.