Research Vision
DSPLab advances a provenance-centric paradigm for AI trustworthiness, with focus on security, privacy, safety, explainability, and ethical norms for real-world deployment.

Research Areas

Areas Covered Across the AI Lifecycle

Active

AI Security

Defending AI against poisoning, backdoors, evasion, and model extraction attacks.

Poisoning Backdoors Adversarial Robustness
Active

AI Privacy

Preventing leakage of sensitive training information and inference-time privacy risks.

Membership Inference Privacy Audits Data Protection
Active

AI Safety

Building reliable systems that avoid unsafe or unintended model behavior in critical settings.

Safety Validation Reliability Robust Deployment
Active

Explainability and Accountability

Making model behavior interpretable, traceable, and auditable.

Explainability Traceability AI Forensics
Active

Ethical and Responsible AI

Embedding ethical norms and accountability into lifecycle-wide AI governance.

Ethical Norms Responsible AI Accountability
Provenance-Centric AI Security and Safety
As AI systems increasingly influence high-stakes domains such as cybersecurity, finance, healthcare, and autonomous systems, the core challenge is no longer only improving accuracy but ensuring that AI systems remain robust, safe, trustworthy, and accountable throughout their lifecycle. The key lies in provenance: understanding and tracing the origins, lineage of transformations, and influence pathways that shape a model’s behavior. Traditional AI evaluation relies largely on black-box testing, observing outputs without visibility into the internal processes that produced them. This leaves critical blind spots against threats such as data poisoning, backdoor attacks, adversarial manipulation, and unsafe or unintended model behavior. DSPLab introduces fine-grained observability into the AI pipeline by tracking the lifecycle history of data, training dynamics, parameter updates, and inference-time information flows. Through this lens, we develop theoretical foundations, algorithms, and systems that make robustness and safety measurable, explainable, and auditable, enabling attack detection, forensic analysis, accountability, and automated model repair. Our vision is to establish provenance as a foundational layer for AI security and AI safety, transforming AI from opaque systems into observable and auditable infrastructures where model decisions can be traced, inspected, and verified for responsible deployment.
Research Projects

Reproducible systems and open artifacts from active lab efforts.

DeepLeak

Privacy hardening for explanation methods against membership inference leakage.

avatar
Firas Ben Hmida

DeepProv

Inference provenance graph analysis for behavioral diagnosis and targeted DNN repair.

avatar
Firas Ben Hmida

PoisonSpot

Fine-grained training provenance tracking to detect clean-label backdoor poisoning.

avatar
Philemon Hailemariam
Selected Publications
(2026). DeepLeak: Privacy Enhancing Hardening of Model Explanations Against Membership Leakage. In IEEE SaTML 2026.
(2023). DeResistor: Toward Detection-Resistant Probing for Evasion of Internet Censorship. In Proceedings of the 32nd USENIX Security Symposium (SEC'23), 2023.
(2023). Designing Secure Performance Metrics for Last-Level Cache. In Proceedings of the 28th International Workshop on High-Level Parallel Programming Models and Supportive Environments (HIPS 2023).
(2023). MIAShield: Defending Membership Inference Attacks via Preemptive Exclusion of Members. In Proceedings of the 23rd Privacy Enhancing Technologies Symposium (PETS 2023).
(2022). Adversarial Detection of Censorship Measurements. In Proceedings of the 21st ACM Workshop on Privacy in the Electronic Society (WPES'22), co-located with the 29th ACM Conference on Computer and Communications Security (CCS), 2022.
(2022). DP-UTIL: Comprehensive Utility Analysis of Differential Privacy in Machine Learning. In Proceedings of the 12th ACM Conference on Data and Application Security and Privacy (ACM CODASPY).