Simplest cure is probably to add a directory to robots.txt that you ask to ignore, and fill the directory with a few GB of random text in a few thousand files, along with some blobs of random data, either set as filetype of being an excel file, a pdf or something, all small random files that will break the input parser of the AI. Enough companies do that and AI scrapers will start to respect the limiters. Yes a bit of heavy traffic, and you probably will want to do a little bit of web server work that gives each file a current date and time, and also serves them randomly to filename requests, something like a canary does.