logoalt Hacker News

kgeisttoday at 1:17 AM0 repliesview on HN

Probably because there's a ton of code that deals with nested parentheses across languages in the training data, and models have learned how to work around tokenization limitations, when it comes to parentheses.