Machine Learning Models Vulnerable to New Black-Box Attack

Friday 31 January 2025


A team of researchers has discovered a new way for hackers to exploit machine learning models, specifically those that adapt to changing circumstances during testing. This technique, dubbed RIP (Reusing Incorrect Predictions), is a type of black-box attack that takes advantage of the model’s tendency to make confident predictions even when they’re wrong.


The problem arises when machine learning models are trained on one set of data and then tested on another set that has different characteristics. To adapt to these changes, the model makes predictions based on its previous training, but this can lead to incorrect conclusions if the new data is significantly different from what it was trained on.


RIP attacks work by manipulating the testing data in a way that causes the model to make confident but incorrect predictions. The attackers then use these predictions as input for the next round of adaptation, further solidifying their hold on the model’s behavior.


The researchers found that RIP attacks can be particularly effective against models that rely heavily on augmenting the training data with random transformations, such as rotation or flipping. This is because these models are more likely to overfit to the original training data and make incorrect predictions when faced with new and unfamiliar data.


To demonstrate the effectiveness of RIP attacks, the researchers conducted a series of experiments using different machine learning models and testing datasets. They found that even simple models like decision trees could be easily manipulated by RIP attacks, while more complex models like neural networks were also vulnerable to attack.


The researchers believe that their findings have important implications for the development of machine learning models that can adapt to changing circumstances during testing. They suggest that model developers should take steps to prevent or detect RIP attacks, such as using more robust training data or incorporating mechanisms to identify and correct incorrect predictions.


Overall, the discovery of RIP attacks highlights the importance of developing secure and robust machine learning models that can adapt effectively to changing circumstances during testing. By better understanding the limitations and vulnerabilities of these models, researchers and developers can take steps to improve their performance and prevent malicious exploitation.


Cite this article: “Machine Learning Models Vulnerable to New Black-Box Attack”, The Science Archive, 2025.


Machine Learning, Black-Box Attack, Rip Attacks, Machine Learning Models, Testing Data, Incorrect Predictions, Overfitting, Neural Networks, Decision Trees, Secure Ai.


Reference: Trung-Hieu Hoang, Duc Minh Vo, Minh N. Do, “R.I.P.: A Simple Black-box Attack on Continual Test-time Adaptation” (2024).


Leave a Reply