1. Take a large collection of text, e.g. from https://opus.nlpl.eu/corpora-search/zh-CN&en
2. Split into sentences and tokenize sentences into words, e.g. using https://github.com/fxsjy/jieba
3. Count how often each word appears and sort sentences by descending frequency of the least common word.
4. Use binary search to find a location in the sorted collection of sentences where the difficulty feels about right.
Of course this gives you a collection of disjointed sentences, but you can always go to the original file and look at the surrounding context when you find an interesting or confusing one.