The young talent in 萝莉社-Dearborn鈥檚 College of Engineering and Computer Science continues to impress big national funders. Along with fellow Assistant Professors DeLean Tolbert Smith and Fred Feng, can now count himself among the CECS faculty who鈥檝e recently landed the , a prestigious nationwide grant given annually to about 500 early-career faculty who are emerging leaders in their fields.
You don鈥檛 get the nod for a CAREER Award with boilerplate ideas, and Eshete will be going after some big fish with his . Specifically, he鈥檒l be exploring a new approach to machine learning鈥檚 鈥渞obustness鈥 problem, which continues to be one of the biggest challenges in artificial intelligence, primarily because it鈥檚 so intimately tied up with the safety and security aspects of AI systems. So what exactly is robustness? Within AI disciplines, robustness can actually refer to a few different things. If we鈥檙e talking about an autonomous vehicle, for example, we鈥檇 say its navigation system is robust if it can respond correctly to slightly novel situations it鈥檚 never seen before, like stopping at a stop sign even if there鈥檚 a minor glare or a little snow on it.
Alongside safety, the other area where robustness is a big deal is cybersecurity, which is Eshete鈥檚 specialty. In this space, robustness primarily refers to whether a machine learning system can protect itself from a range of attacks. For example, if you鈥檙e a medical researcher building an AI-powered tool that can diagnose cancer by looking at a patient鈥檚 medical scans, your system was almost surely trained on real people鈥檚 medical information. Because of this, protecting people鈥檚 privacy is a big concern, especially as other doctors and hospitals start using your tool in the real world. Eshete says if your model isn鈥檛 sufficiently protected, it could be susceptible to an 鈥渋nference attack鈥 in which the model can be probed to identify a particular person whose data was used to train your model. In another kind of attack, Eshete says a malicious actor could provide scans that look legitimate, but contain 鈥渘oise鈥 specifically designed to manipulate the algorithm鈥檚 鈥渄ecision boundary.鈥 By intentionally 鈥減oisoning鈥 the system with bad data, the model鈥檚 high-stakes medical diagnoses could become inaccurate.
Eshete says AI practitioners typically use two kinds of strategies to make models robust against attacks. You can make the model harder to attack; or, you can try to 鈥渃lean鈥 the inputs of malicious 鈥渘oise鈥 before they get into the system, sort of like an antivirus pre-scan of email attachments. But Eshete thinks both of these approaches are ultimately limited, as attackers will invariably find novel ways to thwart new defenses. Instead, he鈥檚 proposing a more fundamental solution that borrows insights from key parts of his work on nation-state cyberattacks. Such highly sophisticated attacks evolve over longer periods of time, beginning with an infiltration point that lets a hacker gradually probe more deeply into other parts of the network, where real damage can be done. Eshete developed techniques for describing these pathways of attack, which is known in the cybersecurity world as attack provenance. 鈥淚t basically gives you a narrative of how the attack unfolded, and within that you鈥檒l find all kinds of juicy information that can help with attack detection and forensic analysis,鈥 he says. For example, such information helped Eshete develop anomaly detection systems, which thwart attacks by using machine learning to develop models of what normal network activity looks like, so unusual activities stick out as potential threats.
For his CAREER award project, he鈥檒l be using this idea of provenance capture and analysis as the basis for a robust cybersecurity defense strategy for AI models. Here鈥檚 how it would work: First, as a machine learning model is trained on its initial data, Eshete would capture a narrative of the model鈥檚 evolution as 鈥渢raining provenance,鈥 which helps establish the typical trajectory of what happens during training. Then, as the model goes live and users are feeding it new real-world inputs, he would similarly track what happens as an input moves all the way through the model to a final predictive output, thus capturing the equivalent of a 鈥渢hought process鈥 that led to a decision. This is called 鈥減rediction provenance.鈥
Importantly, each kind of provenance builds a signature for how the model normally functions, and Eshete says some critical insights can be gathered when you retrain the model with new inputs, which is a typical periodic maintenance procedure for machine learning systems. 鈥淲hen you retrain a model for which training provenance has been established, but with newly acquired and hence potentially poisonous data, if there is enough deviation from the stabilized training provenance, then that is a reason to suspect a data poisoning attack,鈥 Eshete says. Similarly, if a particular prediction provenance sufficiently resembles a known malicious prediction provenance signature, Eshete knows he may be dealing with an input aimed at misleading the model. If the prediction provenance is more in line with the clean provenance signature, he鈥檚 more likely dealing with a benign input.
Interestingly, Eshete鈥檚 core idea could have implications far beyond the cybersecurity realm. As we wrote about in a story earlier this year, AI鈥檚 so-called 鈥渂lack box鈥 problem continues to be a major hurdle to building safe, trustworthy AI systems. The problem stems from the fact that most machine learning systems are completely opaque about how they come to their conclusions 鈥 we get the decision but not the 鈥渢hought process.鈥 Eshete鈥檚 methods of creating prediction provenance could therefore provide new insights into how models are making their decisions 鈥 giving us the potential to build more transparent, repairable, trustworthy and less-biased AI systems.
Eshete will be working on the project through 2028, which will also include an educational component for Taylor High School students in the final two years. Right now, though, he鈥檚 taking a moment to enjoy the big vote of confidence that comes with being a CAREER awardee. 鈥淚t does feel like a validation of your ideas and your career trajectory,鈥 Eshete says. 鈥淎nd I鈥檓 grateful to the Office of Research and my colleagues and the campus in general, who鈥檝e been very supportive and optimistic about research on our campus. It鈥檚 a very big undertaking to apply for a grant like this 鈥 at least it was for me. You can鈥檛 operate in a vacuum, and this kind of support is invaluable for moving research forward.鈥
###
Want to read more about Eshete鈥檚 work? Check out our stories 鈥A dispatch from the cybersecurity arms race鈥 and 鈥Should we view cyberattacks as acts of war?鈥 which spotlight Eshete鈥檚 research on nation-state attacks. Story by Lou Blouin.