Movatterモバイル変換


[0]ホーム

URL:


CN114337987B - A privacy-preserving ship name recognition model training method using federated learning - Google Patents

A privacy-preserving ship name recognition model training method using federated learning
Download PDF

Info

Publication number
CN114337987B
CN114337987BCN202111680336.3ACN202111680336ACN114337987BCN 114337987 BCN114337987 BCN 114337987BCN 202111680336 ACN202111680336 ACN 202111680336ACN 114337987 BCN114337987 BCN 114337987B
Authority
CN
China
Prior art keywords
model
aggregation
participants
participant
aggregation server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111680336.3A
Other languages
Chinese (zh)
Other versions
CN114337987A (en
Inventor
冼允廷
赵搏文
毛茹
曾策冲
沈强
解先旭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Chuangyiyuan Intelligent Technology Co ltd
Guangdong Yousuan Technology Co ltd
Original Assignee
Guangdong Chuangyiyuan Intelligent Technology Co ltd
Guangdong Yousuan Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Chuangyiyuan Intelligent Technology Co ltd, Guangdong Yousuan Technology Co ltdfiledCriticalGuangdong Chuangyiyuan Intelligent Technology Co ltd
Priority to CN202111680336.3ApriorityCriticalpatent/CN114337987B/en
Publication of CN114337987ApublicationCriticalpatent/CN114337987A/en
Application grantedgrantedCritical
Publication of CN114337987BpublicationCriticalpatent/CN114337987B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Landscapes

Abstract

The invention discloses a privacy protection ship name recognition model training method adopting federal learning, which comprises the following steps: s1, initiating a joint training ship name recognition model task; s2, training a local model by the participant and encrypting; s3, aggregating encryption models of participants and executing a federal average algorithm; s4, the aggregation server part decrypts the aggregation model and S5, the participant part decrypts the aggregation model; the method comprises the steps that a participant locally trains a model and encrypts the training model, the locally trained encryption model is uploaded to an aggregation server, and the aggregation server executes a federal average algorithm on the encrypted model; an encryption method is adopted to prevent an attacker from obtaining training data according to model parameters. In addition, the method of dividing the private key into partial private keys is adopted, the server executes partial decryption once, and the participants execute partial decryption once, so that a plurality of participants can effectively train a ship name recognition model together under the condition of not revealing the training data of the participants.

Description

Privacy protection ship name recognition model training method adopting federal learning
Technical Field
The invention relates to the technical field of information safety, in particular to a privacy protection ship name recognition model training method adopting federal learning.
Background
Federal machine learning (FEDERATED MACHINE LEARNING/FEDERATED LEARNING), also known as federal learning. Federal machine learning is a machine learning framework that can effectively help multiple institutions perform data usage and machine learning modeling while meeting the requirements of user privacy protection, data security, and government regulations. The federal learning is used as a machine learning model of a type of joint learning, so that the problem of data island can be effectively solved, participants can be modeled jointly on the basis of not sharing data, the data island can be broken technically, and AI cooperation is realized.
At present, the federal learning is combined with the maritime field, and certain difficulty exists. Different maritime parts have different ship name data sets, but the ship name data sets cannot be shared among the different maritime parts due to information security limitations. However, the ship can span the management range of different maritime departments, how to solve the problem of data island caused by the information security problem, and how to use the ship name data set of different maritime parts to solve the problem of improving the accuracy of the model, which is faced by the artificial intelligence technology in the application of intelligent water traffic. Fortunately, federal learning provides a potential solution to the two problems described above, namely, training models locally for each maritime component, and then submitting the trained models to an aggregation server for aggregation to co-train a more efficient model. Intuitively, this can effectively solve the above-described problem. However, studies have shown that model parameters still leak training data. Therefore, if the federal learning privacy protection ship name recognition model training method is designed, the participants only upload the encrypted training model to the aggregation server, and the server executes the federal average algorithm on the encrypted model, so that the data privacy of the participants can be protected. But this has drawbacks such as the participants sharing the private key and the participants being faced with a high decryption overhead.
Therefore, in order to solve the problems in the prior art, it is important to provide a privacy-preserving ship name recognition model training technique using federal learning, which does not leak training data and reduces decryption overhead.
Disclosure of Invention
The invention aims to provide a privacy protection ship name recognition model training method adopting federal learning, and aims to solve the problems of ship name data island, model parameter data training data privacy and high decryption expense of participants in all maritime parts. Specifically, in order to solve the problem of marine part data island, the invention adopts a federal learning mode to realize the joint training of a ship name recognition model by a plurality of marine parts with ship name data sets; in order to prevent model parameters from revealing training data, the method adopts a Paillier password system to encrypt the trained model; in order to reduce the decryption overhead of the participant, the invention proposes to use a partial decryption approach to significantly reduce the decryption overhead of the participant.
In order to achieve the above purpose, the present invention adopts the following technical scheme:
The invention provides a privacy protection ship name recognition model training method adopting federal learning, which comprises a joint training initiator, a model aggregation server and n federal model training participants (hereinafter referred to as initiator, aggregation server and n participants); the method comprises the following steps:
s1, initiating a joint training ship name recognition model task: initiator initiates a ship name recognition modelN participants are recruited, and model aggregation tasks are sent to an aggregation server;
in step S1, to protect the privacy of the data of the participant, the method further includes: the initiator initializes a Paillier cryptosystem public-private key pair (pk, sk), wherein the initiator splits the private key sk into two partial private keys sk1 and sk2; the initiator sends (pk, sk1) to the aggregation server andTo n participants.
S2, training a local model by the participant and encrypting: the participants train a new local model by utilizing the local ship name data set and the initial model or the decrypted aggregation model; the training process is formally represented as:
wherein f represents a local training function, d represents a local ship name data set of a participant, and t represents the aggregation times;
After training, the participants call the encryption model parameters of the encryption algorithm Enc of the Paillier cryptosystem, and the encryption process is expressed as follows in terms of form:
subsequently, the participant will encrypt the training modelAnd the local data quantity |d| of the participant is sent to the aggregation server;
S3, aggregating encryption models of participants and executing a federal average algorithm: encryption model of aggregation server when receiving n participantsThe aggregation server then performs the following calculations using the scalar multiplication homomorphism and scalar addition homomorphism of the Paillier cryptographic system:
S4, the aggregation server part decrypts the aggregation model: the aggregation server determines whether a termination condition is reached (e.g., whether the required number of times is aggregated) in the following manner:
if the set termination condition is not met, the aggregation server executes partial decryption operation on the encrypted aggregation model, and the aggregation server calculates the form as follows:
wherein N is a public key parameter; in addition, in step S4, the aggregation server calculates the sum of all participant local data volumesAnd sending D to the participant; the aggregation server willSending to the participant;
if the set termination condition has been reached, the aggregation server willAnd D is sent as an output to the initiator, wherein T represents the maximum number of aggregations;
S5, the participant part decrypts the aggregation model: the participants receive the aggregation model information of the aggregation server (i.e) Thereafter, the participant performs a partial decryption operation to obtain a decrypted aggregate model, and the participant formally performs the following calculations:
Wherein N is a public key parameter; after obtaining the decrypted model, the participants continue to calculateObtaining an average model of the aggregation
Preferably, in said step S1, an initialized ship name recognition model is providedIs a recurrent neural network CRNN, wherein,Including model structure and model parameters;
The public key pk= (n+1, N), where n=pq, p and q are two strong primes of sufficient secure length (e.g. 512 bits); private key sk= (λ, μ), where λ=pq-p-q+1, μ=λ-1 mod N, i.e. μ is the inverse of λ modulo N; the partial private key sk2 is an 80-bit random number, and the partial private key sk1=λμ-sk2 mod lambdan; when sk2 is chosen to be a smaller number, the decryption overhead for the participant is significantly reduced.
Preferably, in the step S2, the Enc encryption model isWherein r is a random positive integer less than N.
Specifically, the participant retrains a new model by taking the last trained model and the owned ship name data set as inputs; to prevent an attacker from guessing training data from the model parameters, the participants use Paillier encryption to encrypt the model parameters. Specifically, the Enc encryption model isWhere r is a random positive integer not exceeding N. It is noted that for simplicity of description, the present invention usesTo replace CRNN model, in actual operation, we need to doPerforms an operation on each parameter of the (c).
Preferably, in said step S3, said scalar multiplication homomorphism isThe scalar addition homomorphism isWhere Dec is the decryption algorithm of the Paillier cryptosystem, formally expressed asThe aggregation server gives a calculation according to scalar multiplication homomorphism and scalar addition homomorphismEasy push-out
Preferably, in said step S4, the aggregation server performs a partial decryption operation only before reaching the maximum aggregation number T, and when reaching the maximum aggregation number T, the aggregation server transmits the aggregation model to the initiator; after the aggregation server performs a partial decryption, the participant performs a partial decryption operation to obtain a decrypted model. The purpose of the aggregation server performing partial decryption is that the aggregation server performs a partial decryption operation once to reduce the decryption overhead of the participants.
Specifically, in the step S5, a ciphertext is givenCalculation ofAnd continue the calculationSince sk1+sk2 =0 mod λ and sk1+sk2 =1 mod N, it can be deduced
Also because of rλN=1mod N2, [ x ]1·[x]2=1+xN mod N2. Due to the nature described above, the participant can obtain a decrypted aggregate model by two partial decrypts. Notably, the participants perform a division after decryption (i.e) To obtain an average model after polymerization.
The present invention also provides a storage device having stored therein a plurality of instructions adapted to be loaded by a processor and to perform the step operations of the privacy preserving ship name recognition model training method as previously described.
The invention also provides an intelligent terminal which comprises a processor for executing each instruction and a storage device for storing a plurality of instructions, wherein the instructions are suitable for being loaded by the processor and executing the step operation of training the privacy preserving ship name recognition model.
Compared with the prior art, the invention has the following beneficial effects:
the privacy protection ship name recognition model training method provided by the invention has the following advantages: (1) The problem of island of marine data can be solved, and particularly, the problem of island of marine data caused by information security is solved by jointly training a ship name recognition model by a plurality of marine departments with data through adopting a federal learning mode; (2) The method can solve the problem of training data privacy, and prevents an attacker from deducing the training data according to model parameters by adopting an encryption mode, so that the privacy of the training data is protected. (3) The method can reduce the decryption cost of the participant, and the participant can reduce the decryption cost of the participant by adopting a partial decryption mode, namely the aggregation server performs decryption once, the participant performs decryption once, and the participant is prevented from obtaining the complete private key by setting a smaller partial private key for the participant.
Drawings
FIG. 1 is a flow chart of a privacy preserving ship name recognition model training method provided by the invention;
Fig. 2 is a schematic diagram of a system structure of a privacy protection ship name recognition model training method provided by the invention.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings.
As shown in fig. 1-2, this embodiment discloses a training method for protecting a ship name recognition model by using federal learning privacy, which jointly trains a ship name recognition model under the condition of not revealing the data privacy of maritime parts. In order to realize the training of a ship name recognition model for privacy protection, the method relates to an initiator, an aggregation server and n participants by combining with the accompanying figures 1 and 2; the privacy protection ship name recognition model training method provided by the embodiment specifically comprises the following steps:
S1, initiating a joint training ship name recognition model task: initializing a system, and initializing a ship name recognition model by an initiator in a federal learning modeThe ship name recognition modelA ship name recognition model was trained by using a convolutional neural network CRNN and recruiting 4 participants, 4 maritime parts. In order to simplify the description, the embodiment abstracts the model into 1 parameter, and adopts decimal places to describe the encryption and decryption process; the initiator initializes a public-private key pair (pk, sk) of the Paillier cryptosystem, where pk= (36, 35), sk= (24, 19), sk2=2,sk1 = 454.
Step S2, the participants train a local model and encrypt the local model: assuming that the 4 participants possess data amounts of 1,2 and 2, respectively, their trained model parameters are 1,2, 3 and 4, respectively. The 4 participants encrypt their trained model parameters, i.eThe participants send the encrypted model and the data volume to the aggregation server.
S3, aggregating encryption models of participants and executing a federal average algorithm: after receiving the encryption models of 4 participants { (648,1), (222,2), (44,3), (141,4) }, the aggregation server performs the following calculations using the scalar multiplication homomorphism and scalar addition homomorphism of the Paillier cryptosystem:
346←6481·2221·442·1412mod 352
S4, the aggregation server part decrypts the aggregation model: the aggregation server first judges whether a termination condition is obtained, and in this embodiment, the termination condition is whether the required times are aggregated; the judgment method is as follows:
if the set termination condition is not met, the aggregation server executes partial decryption operation on the encrypted aggregation model, and the aggregation server calculates the form as follows:
431←268454mod 352
In addition, the aggregation server also calculates d=1+1+2+2. The aggregation server sends (431,346,6) to the participant.
If the set termination condition has been reached, the participant sends 346 and 6 as output to the initiator.
S5, the participant part decrypts the aggregation model: after receiving the aggregate model information (431,346,6) of the aggregate server, the participant performs a partial decryption operation to obtain a decrypted aggregate model, the participant formally performing the following calculations:
After obtaining the decrypted model, the participants continue to calculateAn average model of the polymerization was obtained as 2.833.
Variations and modifications to the above would be obvious to persons skilled in the art to which the invention pertains from the foregoing description and teachings. Therefore, the invention is not limited to the specific embodiments disclosed and described above, but some modifications and changes of the invention should be also included in the scope of the claims of the invention. In addition, although specific terms are used in the present specification, these terms are for convenience of description only and do not limit the present invention in any way.

Claims (4)

CN202111680336.3A2021-12-302021-12-30 A privacy-preserving ship name recognition model training method using federated learningActiveCN114337987B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202111680336.3ACN114337987B (en)2021-12-302021-12-30 A privacy-preserving ship name recognition model training method using federated learning

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202111680336.3ACN114337987B (en)2021-12-302021-12-30 A privacy-preserving ship name recognition model training method using federated learning

Publications (2)

Publication NumberPublication Date
CN114337987A CN114337987A (en)2022-04-12
CN114337987Btrue CN114337987B (en)2024-09-10

Family

ID=81022474

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202111680336.3AActiveCN114337987B (en)2021-12-302021-12-30 A privacy-preserving ship name recognition model training method using federated learning

Country Status (1)

CountryLink
CN (1)CN114337987B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN115438355A (en)*2022-07-012022-12-06上海大学Privacy protection federal learning system and method in unmanned aerial vehicle auxiliary Internet of vehicles
CN115549888B (en)*2022-09-292025-08-26南京邮电大学 A privacy protection method for federated learning based on blockchain and homomorphic encryption
CN115731549A (en)*2022-11-182023-03-03广东优算科技有限公司Water ship name joint identification method and system, electronic device and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110572253A (en)*2019-09-162019-12-13济南大学 A method and system for enhancing the privacy of federated learning training data
CN113037460A (en)*2021-03-032021-06-25北京工业大学Federal learning privacy protection method based on homomorphic encryption and secret sharing

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110490738A (en)*2019-08-062019-11-22深圳前海微众银行股份有限公司A kind of federal learning method of mixing and framework

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN110572253A (en)*2019-09-162019-12-13济南大学 A method and system for enhancing the privacy of federated learning training data
CN113037460A (en)*2021-03-032021-06-25北京工业大学Federal learning privacy protection method based on homomorphic encryption and secret sharing

Also Published As

Publication numberPublication date
CN114337987A (en)2022-04-12

Similar Documents

PublicationPublication DateTitle
CN114337987B (en) A privacy-preserving ship name recognition model training method using federated learning
CN108712260B (en) A Privacy-Preserving Multi-Party Deep Learning Computational Agent Approach in Cloud Environment
CN104168108B (en)It is a kind of to reveal the traceable attribute base mixed encryption method of key
US9571274B2 (en)Key agreement protocol
CN101977112B (en) A Public Key Cryptography Encryption and Decryption Method Based on Neural Network Chaotic Attractor
CN109309569A (en)The method, apparatus and storage medium of collaboration signature based on SM2 algorithm
CN102970143B (en)Method for securely computing index of sum of held data of both parties by adopting addition homomorphic encryption
CN105406967A (en)Hierarchical attribute encryption method
CN111783129A (en)Data processing method and system for protecting privacy
CN111159727B (en)Multi-party cooperation oriented Bayes classifier safety generation system and method
CN110299987A (en)A kind of millionaires' problem solution based on homomorphic cryptography
CN109543434A (en)Block chain information encryption method, decryption method, storage method and device
CN112491529A (en)Data file encryption and integrity verification method and system used in untrusted server environment
CN108880782A (en)The secrecy calculation method of minimum value under a kind of cloud computing platform
US20160352689A1 (en)Key agreement protocol
CN106878322A (en) An Encryption and Decryption Method Based on Attribute-Based Fixed-length Ciphertext and Key
CN112953700A (en)Method, system and storage medium for improving safe multiparty computing efficiency
CN114785510A (en)Verifiable lightweight privacy protection federal learning system and method
EP4262134A1 (en)Secure multi-party computation methods and apparatuses
CN107465508A (en)A kind of method, system and the equipment of software and hardware combining construction true random number
CN116822661B (en) Privacy-preserving verifiable federated learning method based on dual-server architecture
CN117375815A (en)Anti-leakage anonymous inner product predicate encryption method
CN115189950B (en)Verifiable gradient security aggregation method and system based on multiparty security calculation
WO2016187690A1 (en)Key agreement protocol
CN114866312B (en)Shared data determining method and device for protecting data privacy

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp