SwePub
Sök i LIBRIS databas

  Extended search

onr:"swepub:oai:DiVA.org:ltu-76111"
 

Search: onr:"swepub:oai:DiVA.org:ltu-76111" > Lightweight Privacy...

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Lightweight Privacy-preserving Training and Evaluation for Discretized Neural Networks

Chen, Jialu (author)
Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai, China
Zhou, Jun (author)
Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai, China
Cao, Zhenfu (author)
Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai, China
show more...
Vasilakos, Athanasios (author)
Luleå tekniska universitet,Datavetenskap,Department of Computer Science and Technology, Fuzhou University, China
Dong, Xiaolei (author)
Shanghai Key Laboratory of Trustworthy Computing, East China Normal University, Shanghai, China.
Choo, Kim-Kwang Raymond (author)
Department of Information Systems and Cyber Security, University of Texas at San Antonio, San Antonio, TX, USA
show less...
 (creator_code:org_t)
IEEE, 2020
2020
English.
In: IEEE Internet of Things Journal. - : IEEE. - 2327-4662. ; 7:4, s. 2663-2678
  • Journal article (peer-reviewed)
Abstract Subject headings
Close  
  • Machine learning, particularly the neural network, is extensively exploited in dizzying applications. In order to reduce the burden of computing for resource-constrained clients, a large number of historical private datasets are required to be outsourced to the semi-trusted or malicious cloud for model training and evaluation. To achieve privacy preservation, most of the existing work either exploited the technique of public key fully homomorphic encryption (FHE) resulting in considerable computational cost and ciphertext expansion, or secure multiparty computation (SMC) requiring multiple rounds of interactions between user and cloud. To address these issues, in this paper, a lightweight privacy-preserving model training and evaluation scheme LPTE for discretized neural networks is proposed. Firstly, we put forward an efficient single key fully homomorphic data encapsulation mechanism (SFH-DEM) without exploiting public key FHE. Based on SFH-DEM, a series of atomic calculations over the encrypted domain including multivariate polynomial, nonlinear activation function, gradient function and maximum operations are devised as building blocks. Furthermore, a lightweight privacy-preserving model training and evaluation scheme LPTE for discretized neural networks is proposed, which can also be extended to convolutional neural network. Finally, we give the formal security proofs for dataset privacy, model training privacy and model evaluation privacy under the semi-honest environment and implement the experiment on real dataset MNIST for recognizing handwritten numbers in discretized neural network to demonstrate the high efficiency and accuracy of our proposed LPTE.

Subject headings

NATURVETENSKAP  -- Data- och informationsvetenskap -- Medieteknik (hsv//swe)
NATURAL SCIENCES  -- Computer and Information Sciences -- Media and Communication Technology (hsv//eng)

Keyword

Discretized neural networks
privacy-preserving
secure outsourced computation
efficiency
Neural networks
Training
Computational modeling
Data privacy
Public key
Pervasive Mobile Computing
Distribuerade datorsystem

Publication and Content Type

ref (subject category)
art (subject category)

Find in a library

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view