COLIBRI: Optimizing Multi-party Secure Neural Network Inference Time for Transformers
| Authors |
|
|---|---|
| Publication date | 2025 |
| Host editors |
|
| Book title | ICT Systems Security and Privacy Protection |
| Book subtitle | 40th IFIP International Conference, SEC 2025, Maribor, Slovenia, May 21–23, 2025 : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Series | IFIP Advances in Information and Communication Technology |
| Event | 40th IFIP International Conference on ICT Systems Security and Privacy Protection |
| Volume | Issue number | I |
| Pages (from-to) | 17-31 |
| Number of pages | 15 |
| Publisher | Cham: Springer |
| Organisations |
|
| Abstract |
Secure Neural Network Inference (SNNI) protocols enable privacy-preserving inference by ensuring the confidentiality of inputs, model weights, and outputs. However, large neural networks, particularly Transformers, face significant challenges in SNNI due to high computational costs and slow execution, as these networks are typically optimized for accuracy rather than secure inference speed. We present COLIBRI, a novel approach that optimizes neural networks for efficient SNNI using Neural Architecture Search (NAS). Unlike prior methods, COLIBRI directly incorporates SNNI execution time as an optimization objective, leveraging a prediction model to estimate execution time without repeatedly running costly SNNI protocols during NAS. Our results on Cityscapes, a complex image segmentation task, show that COLIBRI reduces SNNI execution time by 26–33% while maintaining accuracy, marking a significant advancement in secure AI deployment. |
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-031-92882-6_2 |
| Other links | https://www.scopus.com/pages/publications/105005934541 |
| Downloads |
978-3-031-92882-6_2
(Final published version)
|
| Permalink to this page | |
