Researchers Develop Technique to Allow AI Models to “Forget” Specific Data Entries

About the Author
Ryan Daws is a senior editor at TechForge Media with over a decade of experience in crafting compelling narratives and making complex topics accessible. His articles and interviews with industry leaders have earned him recognition as a key influencer by organizations like Onalytica. Under his leadership, publications have been praised by analyst firms such as Forrester for their excellence and performance.
Connect with him on X (@gadget_ry), Bluesky (@gadgetry.bsky.social), and/or Mastodon (@gadgetry@techhub.social).
Researchers Develop Method to Enable AI Models to Selectively ‘Forget’ Specific Classes of Data
Progress in Artificial Intelligence (AI) has provided tools capable of revolutionizing various domains, from healthcare to autonomous driving. However, as technology advances, so do its complexities and ethical considerations.
The paradigm of large-scale pre-trained AI systems, such as OpenAI’s ChatGPT and CLIP (Contrastive Language–Image Pre-training), has reshaped expectations for machines. These highly generalist models, capable of handling a vast array of tasks with consistent precision, have seen widespread adoption for both professional and personal use.
However, such versatility comes at a hefty price. Training and running these models demands prodigious amounts of energy and time, raising sustainability concerns, as well as requiring cutting-edge hardware significantly more expensive than standard computers. Compounding these issues is that generalist tendencies may hinder the efficiency of AI models when applied to specific tasks.
"For instance, ‘in practical applications, the classification of all kinds of object classes is rarely required,’" explains Associate Professor Go Irie, who led the research. "For example, in an autonomous driving system, it would be sufficient to recognize limited classes of objects such as cars, pedestrians, and traffic signs. We would not need to recognize food, furniture, or animal species. Retaining classes that do not need to be recognized may decrease overall classification accuracy, as well as cause operational disadvantages such as the waste of computational resources and the risk of information leakage."
A potential solution lies in training models to ‘forget’ redundant or unnecessary information—streamlining their processes to focus solely on what is required. While some existing methods already cater to this need, they tend to assume a ‘white-box’ approach where users have access to a model’s internal architecture and parameters. Oftentimes, however, users get no such visibility.
"Black-box" AI systems, more common due to commercial and ethical restrictions, conceal their inner mechanisms, rendering it difficult to induce selective forgetting in these models.
Addressing the Challenges of Machine Unlearning
The researchers developed a method that enables black-box vision-language models to selectively forget specific classes of data. This approach has significant potential for real-world applications where task-specific precision is paramount.
"Simplifying models for specialized tasks could make them faster, more resource-efficient, and capable of running on less powerful devices—hastening the adoption of AI in areas previously deemed unfeasible," notes Associate Professor Irie.
Another key use lies in image generation, where forgetting entire categories of visual context could prevent models from inadvertently creating undesirable or harmful content, be it offensive material or misinformation.
Perhaps most importantly, this method addresses one of AI’s greatest ethical quandaries: privacy. AI models, particularly large-scale ones, are often trained on massive datasets that may inadvertently contain sensitive or outdated information. Requests to remove such data—especially in light of laws advocating for the ‘Right to be Forgotten’—pose significant challenges.
"Retraining entire models to exclude problematic data is costly and time-intensive, yet the risks of leaving it unaddressed can have far-reaching consequences," warns Associate Professor Irie. "Selective forgetting, or so-called machine unlearning, may provide an efficient solution to this problem."
These privacy-focused applications are especially relevant in high-stakes industries like healthcare and finance, where sensitive data is central to operations.
The Future of AI: Balancing Efficiency and Ethics
As the global race to advance AI accelerates, the Tokyo University of Science’s black-box forgetting approach charts an important path forward—not only by making the technology more adaptable and efficient but also by adding significant safeguards for users.
"While the potential for misuse remains, methods like selective forgetting demonstrate that researchers are proactively addressing both ethical and practical challenges," notes Associate Professor Irie.
This innovation has far-reaching implications for various industries and applications. By enabling AI models to selectively forget specific classes of data, we can create more efficient, adaptable, and responsible technologies that prioritize user privacy and safety.
Related Events
Want to learn more about AI and big data from industry leaders? Check out the following events:
- AI & Big Data Expo: Taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.
- Explore other upcoming enterprise technology events and webinars powered by TechForge here.
Conclusion
The researchers’ development of a method to enable black-box vision-language models to selectively forget specific classes of data marks an important step forward in the field of AI. By addressing both technical and ethical challenges, this innovation has far-reaching implications for various industries and applications.
As we continue to push the boundaries of what is possible with AI, it is essential that we prioritize responsible development and deployment of these technologies. The Tokyo University of Science’s black-box forgetting approach demonstrates a commitment to balancing efficiency and ethics in AI research.
Tags: AI, Artificial Intelligence, Ethics, Machine Learning, Privacy