A bit of hype in the AI wording here. This could be called a chip with hardcoded logic obtained with machine learning
ML is part of AI, and has always been. AI is not equal to chatgpt and AI wasn't coined/conceived in November 2022.
Calling it "AI" is marketing sugar. It is closer to an inference-only state machine where gradient descent did the wiring instead of an engineer, and the annoying part is that once the detector setup or noise profile moves, retraining and redeploy stop being normal ML chores and turn into hardware respins, validation, and a lot of waiting. That distinction stops sounding pedantic the first time a bug fix means touching silicon instead of pushing to a repo.
Is a LLM logic in weights derived from machine learning?
AI is not a new thing, and machine learned logic definitely counts as AI.