Privacy-Preserving AI Breakthrough in Omics Data Analysis, Saudi Arabia

Date:

Updated: [falahcoin_post_modified_date]

A research team at KAUST has created a machine-learning method that utilizes a collection of algorithms focused on preserving privacy. This approach tackles a critical issue in medical research: leveraging artificial intelligence (AI) to expedite discoveries from genomic data without compromising individual privacy.

Omics data usually contains a lot of private information, such as gene expression and cell composition, which could often be related to a person’s disease or health status, says KAUST’s Xin Gao. AI models trained on this data – particularly deep learning models – have the potential to retain private details about individuals. Our primary focus is finding an improved balance between preserving privacy and optimizing model performance.

The traditional approach to preserving privacy is to encrypt the data. However, this requires the data to be decrypted for training, which introduces a heavy computational overhead. The trained model also still retains private information and so can only be used in secure environments.

Another way to preserve privacy is to break the data into smaller packets and train the model separately on each packet using a team of local training algorithms, an approach known as local training or federated learning. However, on its own, this approach still has the potential to leak private information into the trained model. A method called differential privacy can be used to break up the data in a way that guarantees privacy, but this results in a noisy model that limits its utility for precise gene-based research.

Using the differential privacy framework, adding a shuffler can achieve better model performance while keeping the same level of privacy protection; but the previous approach of using a centralized third-party shuffler that introduces a critical security flaw in that the shuffler could be dishonest, says Juexiao Zhou, lead author of the paper and a Ph.D. student in Gao’s group. The key advance of our approach is the integration of a decentralized shuffling algorithm. He explains that the shuffler not only resolves this trust issue but achieves a better trade-off between privacy preservation and model capability, while ensuring perfect privacy protection.

The team demonstrated their privacy-preserving machine-learning approach on three challenging multi-omics tasks. Not only did this method produce optimized models efficiently, but it also stood up against state-of-the-art cyberattacks, proving its robustness.

It is important to be aware that proficiently trained deep-learning models possess the ability to retain significant amounts of private information from training data, says Gao. As deep learning is being increasingly applied to analyze biological and biomedical data, the importance of privacy protection is greater than ever.

[single_post_faqs]

Share post:

Subscribe

Popular

More like this
Related

Revolutionary Small Business Exchange Network Connects Sellers and Buyers

Revolutionary SBEN connects small business sellers and buyers, transforming the way businesses are bought and sold in the U.S.

District 1 Commissioner Race Results Delayed by Recounts & Ballot Reviews, US

District 1 Commissioner Race in Orange County faces delays with recounts and ballot reviews. Find out who will come out on top in this close election.

Fed Minutes Hint at Potential Rate Cut in September amid Economic Uncertainty, US

Federal Reserve minutes suggest potential rate cut in September amid economic uncertainty. Find out more about the upcoming policy decisions.

Baltimore Orioles Host First-Ever ‘Faith Night’ with Players Sharing Testimonies, US

Experience the powerful testimonies of Baltimore Orioles players on their first-ever 'Faith Night.' Hear how their faith impacts their lives on and off the field.