Search results for: 'training-data poisoning, AI model theft or adversarial samples to attack AI-powered systems"'