Trustworthy Machine Learning with Data in the Wild

dc.contributor.authorLu, Yiwei
dc.date.accessioned2025-06-19T20:02:48Z
dc.date.available2025-06-19T20:02:48Z
dc.date.issued2025-06-19
dc.date.submitted2025-06-13
dc.description.abstractRecent advances in machine learning (ML) have been largely fueled by models trained on extensive internet-collected datasets. While this approach has yielded remarkable capabilities, it introduces critical vulnerabilities: training data can be untrustworthy, containing harmful content or becoming susceptible to data poisoning attacks. In such scenarios, model behavior can be maliciously altered, resulting in reduced test performance for classification models or the replication of copyrighted materials for generative models. This thesis examines the influence of untrusted training data on machine learning training dynamics through two crucial perspectives: the ML developer's lens, focusing on model integrity, and the data owner's viewpoint, addressing privacy and copyright concerns. Specifically, this thesis analyzes the impact of data in the wild from both theoretical and empirical perspectives. The first part formulates data poisoning attacks (specifically, accuracy degradation attacks) as bi-level optimization problems, also known as Stackelberg games. It provides a viable algorithm to poison modern machine learning models, particularly neural networks, which demonstrate significantly greater robustness to such attacks compared to traditional linear models. The second part investigates this robustness distinction and develops a principled theoretical framework for understanding the effectiveness boundaries of data poisoning attacks across various scenarios. Given some clean training data, a target model, and malicious parameter objectives, this theoretical tool determines the minimum amount of poisoned data required to achieve these parameters, thereby quantifying the fundamental limits of data poisoning attacks. Building upon the understanding of data poisoning attacks in supervised settings (i.e., classification tasks), this thesis further examines their threats in two realistic machine learning pipelines. The third part presents the first comprehensive analysis of data poisoning attacks against pre-trained feature extractors—components frequently utilized for various downstream ML tasks, such as adapting large models to medical data. This analysis reveals that drastic domain shifts can significantly increase ML models' vulnerability to data poisoning attacks, necessitating more robust countermeasures. The final section examines the role of harmful data in generative models, specifically focusing on advanced latent diffusion models for text-to-image generation tasks. Copyright infringement concerns arise when such models produce outputs substantially similar to copyrighted training data. This section introduces a novel scenario termed "disguised copyright infringement" incurred by targeted data poisoning attacks, providing a thorough description of potential attack vectors and corresponding defensive strategies.
dc.identifier.urihttps://hdl.handle.net/10012/21882
dc.language.isoen
dc.pendingfalse
dc.publisherUniversity of Waterlooen
dc.subjecttrustworthy AI
dc.subjectmachine learning
dc.subjectsecurity and privacy
dc.subjectdata poisoning attack
dc.subjectoptimization
dc.titleTrustworthy Machine Learning with Data in the Wild
dc.typeDoctoral Thesis
uws-etd.degreeDoctor of Philosophy
uws-etd.degree.departmentDavid R. Cheriton School of Computer Science
uws-etd.degree.disciplineComputer Science
uws-etd.degree.grantorUniversity of Waterlooen
uws-etd.embargo.terms0
uws.comment.hiddenDear Sir or Madam, Thank you for reviewing my thesis submission! I am hoping to meet the 50% tuition fee refund deadline, which is this Friday (June 20th). I would greatly appreciate any assistance you can provide in accelerating the timeline. Thank you very much in advance. Best regards, Yiwei Lu
uws.contributor.advisorYu, Yaoliang
uws.contributor.advisorSun, Sun
uws.contributor.affiliation1Faculty of Mathematics
uws.peerReviewStatusUnrevieweden
uws.published.cityWaterlooen
uws.published.countryCanadaen
uws.published.provinceOntarioen
uws.scholarLevelGraduateen
uws.typeOfResourceTexten

Files

Original bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Lu_Yiwei.pdf
Size:
31.21 MB
Format:
Adobe Portable Document Format

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
6.4 KB
Format:
Item-specific license agreed upon to submission
Description: