I think you're conflating intelligence and consciousness. Pain and suffering requires consciousness but intelligence does not imply pain or suffering or happiness. LLMs are already "intelligent" to a certain degree in some aspects, though not generally intelligent like humans. But there is no reason to believe that you couldn't have a generally intelligent artificial agent that lacks consciousness and thus can feel no pain or suffering.