{"id":139682,"date":"2025-07-09T09:30:20","date_gmt":"2025-07-09T07:30:20","guid":{"rendered":"https:\/\/diaridigital.urv.cat\/?p=139682"},"modified":"2025-07-28T13:03:46","modified_gmt":"2025-07-28T11:03:46","slug":"removing-information-from-ai","status":"publish","type":"post","link":"https:\/\/diaridigital.urv.cat\/en\/removing-information-from-ai\/","title":{"rendered":"Removing information from AI models is more difficult than it sounds"},"content":{"rendered":"<p>Researchers at the Universitat Rovira i Virgili have studied the effectiveness of <em>unlearning <\/em>techniques in artificial intelligence models. These strategies seek to eliminate personal, incorrect or discriminatory data from large language models such as ChatGPT, Mixtral, Bard or Copilot, among others. The analysis reveals that there is currently no method that guarantees total and irreversible erasure, other than retraining the model without the data in question, a very costly and inefficient process. This creates a conflict with the right to be forgotten, which is enshrined in European legislation and which obliges data controllers to delete people&#8217;s personal data upon request. The solution to this incompatibility, they argue, is to design new ways of training models that facilitate <em>unlearning <\/em>with guarantees.<\/p>\n<p>The performance of artificial intelligence (AI) models &#8211; also called large language models (LLMs) &#8211; depends on the data they are trained on. The companies that run them feed them with as much information as they can from as wide a range of sources as possible to make them more powerful and, above all, better than their competitors. These are huge models with billions of parameters and they know many, many things. &#8220;In some cases, they even know things that, for various reasons, they should not know,&#8221; explains Josep Domingo, researcher at the Department of Computer Engineering and Mathematics and co-author of the research.<\/p>\n<figure id=\"attachment_139669\" aria-labelledby=\"figcaption_attachment_139669\" class=\"wp-caption alignnone\" style=\"width: 1024px\"><a href=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg\"><img loading=\"lazy\" decoding=\"async\" class=\"size-large wp-image-139669\" src=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg\" alt=\"\" width=\"1024\" height=\"682\" srcset=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg 1024w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-300x200.jpg 300w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-768x512.jpg 768w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1536x1024.jpg 1536w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f.jpg 2000w\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" \/><\/a><figcaption id=\"figcaption_attachment_139669\" class=\"wp-caption-text\">David S\u00e1nchez i Josep Domingo, investigadors del Departament d&#8217;Enginyeria Inform\u00e0tica i Matem\u00e0tiques, autors de la recerca.<\/figcaption><\/figure>\n<p>So what happens when an AI model has been fed with copyrighted works &#8211; could it mimic the style of a particular writer and write a sequel to the latest bestseller? And, if the model has personal information, does it know if someone has been ill, taken sick leave or just bought, say, a flat? Fortunately, we have legal mechanisms to protect all this data, such as the Spanish Intellectual Property Law or the European Union&#8217;s General Data Protection Regulation (GDPR).<\/p>\n<p>The GDPR regulates the processing of personal data of any natural person in the European Union and includes, among other aspects, the right to be forgotten. Therefore, when data controllers receive any request to remove personal data from their systems, they have to comply. And that includes all companies with AI models operating in Europe. However, the way these models have been configured makes deleting specific data a much more complex technical challenge than it might seem.<\/p>\n<p>Against this backdrop, researchers from the CRISES Research Group have studied the ability of large language models to <em>unlearn <\/em>and the computational cost of doing so. There are mainly two approaches to removing knowledge from an AI model. The first option is the most rudimentary and involves removing all the knowledge and training the model again without the data that you want to eliminate. &#8220;This is an impractical and computationally very costly process, but it is currently the only way to guarantee one hundred per cent removal,&#8221; says David S\u00e1nchez, researcher at the Department of Computer Engineering and Mathematics and co-author of the study.<\/p>\n<p>The other approach to <em>unlearning <\/em>involves getting the model to forget specific information, which avoids the need to retrain the model from scratch every time some information needs to be removed. The underlying problem, according to the researchers, is that nobody fully understands how LLMs work, not even the people who have designed them. Although it is known how to train them and how to make them more efficient and accurate, there is no way of knowing in which region of the model a particular piece of information resides. S\u00e1nchez warns, however, that the aforementioned methods, although much more efficient, do not fully ensure <em>unlearning <\/em>and reminds us that the Regulation is very clear and requires absolute <em>safeguards<\/em> to be put in place.<\/p>\n<aside class=\"perfil_persona\">There is a method of <em>unlearning <\/em>that works by filtering the responses of the artificial intelligence model: the<strong> apparent forgetting<\/strong>. According to this method, the system works in a routine way but, when it comes to providing information, it keeps back information that is private, sensitive or inappropriate. However, the researchers say that, although it may seem to the user that the method works, the model still stores the problematic information and there is no guarantee that it will not provide it on another occasion if it receives the right instruction.<\/aside>\n<h5>A conflict between law and technology<\/h5>\n<p>The results of the study show that there is currently a conflict between the legislation and the available technology: that is, while it is possible to remove personal data from AI models, the only techniques that guarantee its removal are &#8220;frighteningly expensive&#8221;. In this regard, Domingo points out that the administrators of these models will only implement guaranteed <em>unlearning <\/em>if so requested by lots of users: &#8220;If people see that these models contain their personal data and they start to request to be forgotten, the companies could have problems&#8221;. LLM owners operating in Europe will need to take into account the GDPR and its right to be forgotten, and that means making <em>unlearning <\/em>affordable and cost-effective from a computational and economic point of view.<\/p>\n<p>The URV researchers believe that in order to find more efficient ways of <em>unlearning<\/em>, the models must first be trained with <em>unlearning <\/em>in mind. Currently, LLMs are trained by feeding them all the data at once, but there are several alternatives, which still need to be developed. Some, for example, involve fragmenting the data and feeding it piecemeal to successive versions of the model so that, if a request to be forgotten is received, it is possible to retrieve an earlier version of the model that does not have the knowledge question. Others have to do with the structure of the system and are based on modular learning, which allows parts of the model with specific information to be extracted without affecting the rest of the information stored or the capabilities that the model can legitimately retain.<\/p>\n<div class=\"col-12\" role=\"main\">\n<article id=\"post-139668\" class=\"single_page_post post-139668 post type-post status-publish format-standard has-post-thumbnail hentry category-ciencia-tecnologia category-comciencia category-enginyeria-informatica-matematiques category-escola-enginyeria category-notes-premsa category-recerca\">\n<div class=\"col-8 left\">\n<header>\n<div class=\"hgroup\">\n<h6>09\/07\/2025<\/h6>\n<h1 class=\"entry-title\">Eliminar informaci\u00f3 dels models d\u2019IA \u00e9s m\u00e9s dif\u00edcil del que sembla<\/h1>\n<\/div>\n<h2>Un estudi de la URV evidencia les limitacions del desaprenentatge en intel\u00b7lig\u00e8ncia artificial i alerta que cal desenvolupar nous m\u00e8todes que facilitin l\u2019eliminaci\u00f3 de dades sensibles<\/h2>\n<p><a href=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-1024x768.jpg\" data-slb-group=\"139668\" data-slb-active=\"1\" data-slb-internal=\"0\"><img loading=\"lazy\" decoding=\"async\" class=\"attachment-large size-large wp-post-image\" src=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-1024x768.jpg\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-1024x768.jpg 1024w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-300x225.jpg 300w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-768x576.jpg 768w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-1536x1152.jpg 1536w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-384x287.jpg 384w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850-800x600.jpg 800w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/pexels-markusspiske-113850.jpg 2000w\" alt=\"\" width=\"1024\" height=\"768\" \/><\/a><\/header>\n<div class=\"pf-content\">\n<p>Investigadors de la Universitat Rovira i Virgili han estudiat l\u2019efectivitat de les t\u00e8cniques de <em>desaprenentatge<\/em> en models d\u2019intel\u00b7lig\u00e8ncia artificial. Aquestes estrat\u00e8gies busquen eliminar dades personals \u2014per\u00f2 tamb\u00e9 incorrectes o discriminat\u00f2ries\u2014 dels models de llenguatge extensos com ChatGPT, Mixtral, Bard o Copilot, entre d\u2019altres. L\u2019an\u00e0lisi revela que actualment no hi ha cap un m\u00e8tode que garanteixi un oblit total i irreversible, m\u00e9s enll\u00e0 d\u2019entrenar el model de nou sense les dades de qu\u00e8 es vol prescindir, un proc\u00e9s molt cost\u00f3s i ineficient. Aix\u00f2 crea un conflicte amb el dret a l\u2019oblit, garantit per la legislaci\u00f3 europea, que obliga a eliminar les dades personals de les persones si ho demanen. La soluci\u00f3 a aquesta incompatibilitat, defensen, passa per dissenyar noves formes d\u2019entrenar els models que facilitin el <em>desaprenentatge<\/em> amb garanties.<\/p>\n<p>El rendiment dels models d\u2019intel\u00b7lig\u00e8ncia artificial (IA) \u2014tamb\u00e9 anomenats models de llenguatge extensos (MLE)\u2014 dep\u00e8n de les dades amb qu\u00e8 han estat entrenats. Les companyies que els gestionen els alimenten amb tota la informaci\u00f3 que poden, tan diversa com els resulta possible, per fer-los m\u00e9s potents i, sobretot, millors que els de la compet\u00e8ncia. Es tracta de models enormes, amb milers de milions de par\u00e0metres i que saben moltes, moltes coses. \u201cEn alguns casos fins i tot saben coses que, per diversos motius, no conv\u00e9 que s\u00e0piguen\u201d, explica Josep Domingo, investigador del Departament d\u2019Enginyeria Inform\u00e0tica i Matem\u00e0tiques i coautor de la recerca.<\/p>\n<figure id=\"attachment_139669\" class=\"wp-caption alignnone\" aria-labelledby=\"figcaption_attachment_139669\"><a href=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg\" data-slb-group=\"139668\" data-slb-active=\"1\" data-slb-internal=\"0\"><img loading=\"lazy\" decoding=\"async\" class=\"size-large wp-image-139669\" src=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg\" sizes=\"auto, (max-width: 1024px) 100vw, 1024px\" srcset=\"https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1024x682.jpg 1024w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-300x200.jpg 300w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-768x512.jpg 768w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f-1536x1024.jpg 1536w, https:\/\/diaridigital.urv.cat\/wp-content\/uploads\/2025\/07\/IMG_2606f.jpg 2000w\" alt=\"\" width=\"1024\" height=\"682\" \/><\/a><figcaption id=\"figcaption_attachment_139669\" class=\"wp-caption-text\">David S\u00e1nchez i Josep Domingo, investigadors del Departament d\u2019Enginyeria Inform\u00e0tica i Matem\u00e0tiques, autors de la recerca.<\/figcaption><\/figure>\n<p>Qu\u00e8 passa, doncs, quan un model d\u2019IA s\u2019ha alimentat amb obres protegides per drets d\u2019autor? Podria imitar l\u2019estil d\u2019un escriptor determinat i escriure una seq\u00fcela del <em>best-seller<\/em> de moda? I, si el model t\u00e9 informaci\u00f3 personal, sap si alg\u00fa ha estat malalt, si ha demanat una baixa laboral o si acaba de comprar-se, per exemple, un pis? Per sort disposem de mecanismes legals per protegir totes aquestes dades, com ara la Llei de la propietat intel\u00b7lectual espanyola o el Reglament general de protecci\u00f3 de dades de la Uni\u00f3 Europea (RGPD).<\/p>\n<p>El RGPD, que regula el tractament de les dades personals de qualsevol persona f\u00edsica de la Uni\u00f3 Europea, recull, entre d\u2019altres aspectes, el dret a l\u2019oblit. Per tant, arran de qualsevol petici\u00f3 d\u2019eliminar dades personals dels seus sistemes, els gestors d\u2019aquesta informaci\u00f3 han de prescindir-ne. I aix\u00f2 inclou totes les companyies amb models d\u2019IA que operen a Europa. Tanmateix, la manera com s\u2019han configurat aquests models fa que eliminar dades espec\u00edfiques sigui un repte t\u00e8cnic molt m\u00e9s complex del que podria semblar.<\/p>\n<p>Davant d\u2019aquest escenari, investigadors del grup de recerca CRISES han estudiat la capacitat dels models de llenguatge extensos per desaprendre i el cost computacional que els suposa fer-ho. Principalment, hi ha dues aproximacions per eliminar coneixements d\u2019un model d\u2019IA. La primera opci\u00f3 \u00e9s la m\u00e9s rudiment\u00e0ria i implica suprimir tots els coneixements i entrenar el model de nou sense les dades de qu\u00e8 es vol prescindir. \u201cEs un proc\u00e9s poc pr\u00e0ctic i molt cost\u00f3s en termes computacionals per\u00f2, ara per ara, \u00e9s l\u2019\u00fanica manera de garantir l\u2019oblit al cent per cent\u201d, reconeix David S\u00e1nchez, investigador del Departament d\u2019Enginyeria Inform\u00e0tica i Matem\u00e0tiques i coautor de la recerca.<\/p>\n<p>L\u2019altra manera de tractar el desaprenentatge implica que el model oblidi certa informaci\u00f3 espec\u00edfica, i evitar comen\u00e7ar de nou cada vegada que cal eliminar informaci\u00f3. El problema de fons, segons els investigadors, \u00e9s que no se sap ben b\u00e9 com funcionen els MLE \u2014ni tan sols els qui els han dissenyat. Tot i que se sap com es poden entrenar i com es poden fers m\u00e9s eficients i precisos, no hi ha cap manera de saber en quina regi\u00f3 del model es troba una informaci\u00f3 concreta. Sanchez, per\u00f2, alerta que aquests m\u00e8todes, tot i que molt m\u00e9s eficients, no asseguren totalment el <em>desaprenentatge<\/em> i recorda que el Reglament \u00e9s molt clar i requereix <em>garanties<\/em>.<\/p>\n<aside class=\"perfil_persona\">Existeix un m\u00e8tode de desaprenentatge que funciona filtrant les respostes del model d\u2019intel\u00b7lig\u00e8ncia artificial: l<strong>\u2019oblit aparent.<\/strong> D\u2019aquesta manera, el sistema treballa de forma rutin\u00e0ria, per\u00f2 a l\u2019hora de proporcionar la informaci\u00f3, en ret\u00e9 aquella de car\u00e0cter privat, sensible o inadequat. Segons els investigadors,\u00a0 malgrat que a ulls de l\u2019usuari \u00e9s un m\u00e8tode aparentment v\u00e0lid, la informaci\u00f3 problem\u00e0tica continua emmagatzemada i no es pot garantir que\u00a0 el model no ofereixi les dades restringides amb la instrucci\u00f3 adequada.<\/aside>\n<h5>Un conflicte entre llei i tecnologia<\/h5>\n<p>Els resultats de l\u2019estudi posen de manifest que existeix un conflicte entre la legislaci\u00f3 i la tecnologia disponible: si b\u00e9 \u00e9s possible eliminar dades personals dels models d\u2019IA, les t\u00e8cniques que ofereixen garanties s\u00f3n \u201cespantosament costoses\u201d. En aquest sentit, Domingo puntualitza que l\u2019inter\u00e8s dels gestors d\u2019aquests models per implementar amb garanties el desaprenentatge dep\u00e8n de les peticions dels usuaris: \u201cSi la gent veu que aquests models contenen dades personals i comen\u00e7a a haver-hi peticions perqu\u00e8 les oblidin, podrien tenir problemes\u201d. Els propietaris dels MLE que operen a Europa hauran de tenir en compte el RGPD i el seu dret a l\u2019oblit, i aix\u00f2 implica que el desaprenentatge ha de ser assumible i rendible des d\u2019un punt de vista computacional i econ\u00f2mic.<\/p>\n<p>Segons els investigadors de la URV, per trobar formes de desaprenentatge m\u00e9s eficients, cal entrenar els models pensant en aquest proc\u00e9s . Actualment, s\u2019entrena els MLE alimentant-los amb totes les dades de cop, per\u00f2 hi ha diverses alternatives que encara s\u2019han de desenvolupar. Algunes, per exemple, impliquen fragmentar les dades i alimentar versions successives del model, de manera que, arran d\u2019una petici\u00f3 d\u2019oblit sigui possible recuperar una versi\u00f3 anterior sense coneixement determinat i entrenar-la. D\u2019altres tenen a veure amb l\u2019estructura del sistema i es basen en l\u2019aprenentatge modular, que permet extreure parts del model on hi ha informaci\u00f3 concreta sense afectar la resta de la informaci\u00f3 guardada ni les capacitats que el model pot retenir leg\u00edtimament.<\/p>\n<p><iframe loading=\"lazy\" title=\"YouTube video player\" src=\"https:\/\/www.youtube.com\/embed\/ykwEQpmYF9w?si=AMrgEvT-9gKVYVI3\" width=\"560\" height=\"315\" frameborder=\"0\" allowfullscreen=\"allowfullscreen\" data-mce-fragment=\"1\"><\/iframe><\/p>\n<p>This research is part of the HERMES project, an initiative financed by the National Institute of Cybersecurity (INCIBE) with European Next Generation funds and by the URV.<\/p>\n<\/div>\n<\/div>\n<\/article>\n<\/div>\n<p><strong>Reference: <\/strong>Blanco-Justicia, A., Domingo-Ferrer, J., Jebreel, N. M., Manzanares-Salor, B., &amp; S\u00e1nchez, D. (2025). Unlearning in Large Language Models: We Are Not There Yet. IEEE Computer Society. <a href=\"https:\/\/doi.org\/10.1109\/MC.2024.3468588\">https:\/\/doi.org\/10.1109\/MC.2024.3468588<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>A URV study highlights the limitations of unlearning in artificial intelligence and warns that new methods need to be developed to facilitate the removal of sensitive data<\/p>\n","protected":false},"author":139,"featured_media":139676,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[806,3462,818,195,786,178],"tags":[],"class_list":["post-139682","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-communicating-science","category-computer-engineering-mathematics","category-press-releases","category-research","category-school-engineering","category-science-technology"],"acf":[],"_links":{"self":[{"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/posts\/139682","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/users\/139"}],"replies":[{"embeddable":true,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/comments?post=139682"}],"version-history":[{"count":0,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/posts\/139682\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/media\/139676"}],"wp:attachment":[{"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/media?parent=139682"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/categories?post=139682"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/diaridigital.urv.cat\/en\/wp-json\/wp\/v2\/tags?post=139682"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}