Penn State researchers have designed a graphene-based way to make encrypted keys harder to crack, in an attempt to protect data in an age where more and more private data is stored and shared digitally. Current silicon technology exploits microscopic differences between computing components to create secure keys, but the team explains that artificial intelligence (AI) techniques can be used to predict these keys and gain access to data.
Led by Saptarshi Das, assistant professor of engineering science and mechanics, the researchers used graphene to develop a novel low-power, scalable, reconfigurable hardware security device with significant resilience to AI attacks.
There has been more and more breaching of private data recently, Das said. We developed a new hardware security device that could eventually be implemented to protect these data across industries and sectors.
The device, called a physically unclonable function (PUF), is the first demonstration of a graphene-based PUF, according to the researchers. The physical and electrical properties of graphene, as well as the fabrication process, make the novel PUF more energy-efficient, scalable, and secure against AI attacks that pose a threat to silicon PUFs.
The team fabricated nearly 2,000 identical graphene transistors, which switch current on and off in a circuit. Despite their structural similarity, the transistors’ electrical conductivity varied due to the inherent randomness arising from the production process. While such variation is typically a drawback for electronic devices, it’s a desirable quality for a PUF not shared by silicon-based devices.
After the graphene transistors were implemented into PUFs, the researchers modeled their characteristics to create a simulation of 64 million graphene-based PUFs. To test the PUFs’ security, Das and his team used machine learning, a method that allows AI to study a system and find new patterns. The researchers trained the AI with the graphene PUF simulation data, testing to see if the AI could use this training to make predictions about the encrypted data and reveal system insecurities.
Neural networks are very good at developing a model from a huge amount of data, even if humans are unable to, Das said. We found that AI could not develop a model, and it was not possible for the encryption process to be learned.
This resistance to machine learning attacks makes the PUF more secure because potential hackers could not use breached data to reverse engineer a device for future exploitation, Das said. Even if the key could be predicted, the graphene PUF could generate a new key through a reconfiguration process requiring no additional hardware or replacement of components.
Normally, once a system’s security has been compromised, it is permanently compromised, said Akhil Dodda, an engineering science and mechanics graduate student conducting research under Das’s mentorship. We developed a scheme where such a compromised system could be reconfigured and used again, adding tamper resistance as another security feature.
With these features, as well as the capacity to operate across a wide range of temperatures, the graphene-based PUF could be used in a variety of applications. Further research can open pathways for its use in flexible and printable electronics, household devices and more.
Purdue University researchers also participated in this new study, in collaboration with the Penn State team.