Movatterモバイル変換


[0]ホーム

URL:


CN112506673A - Intelligent edge calculation-oriented collaborative model training task configuration method - Google Patents

Intelligent edge calculation-oriented collaborative model training task configuration method
Download PDF

Info

Publication number
CN112506673A
CN112506673ACN202110153068.3ACN202110153068ACN112506673ACN 112506673 ACN112506673 ACN 112506673ACN 202110153068 ACN202110153068 ACN 202110153068ACN 112506673 ACN112506673 ACN 112506673A
Authority
CN
China
Prior art keywords
training
edge
time slot
model
global
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110153068.3A
Other languages
Chinese (zh)
Other versions
CN112506673B (en
Inventor
邹昊东
张明明
俞俊
陈海洋
夏飞
王鹏飞
范磊
陶晔波
许明杰
王琳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NARI Technology Co Ltd
Information and Telecommunication Branch of State Grid Jiangsu Electric Power Co Ltd
Original Assignee
NARI Technology Co Ltd
Information and Telecommunication Branch of State Grid Jiangsu Electric Power Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NARI Technology Co Ltd, Information and Telecommunication Branch of State Grid Jiangsu Electric Power Co LtdfiledCriticalNARI Technology Co Ltd
Priority to CN202110153068.3ApriorityCriticalpatent/CN112506673B/en
Publication of CN112506673ApublicationCriticalpatent/CN112506673A/en
Application grantedgrantedCritical
Publication of CN112506673BpublicationCriticalpatent/CN112506673B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Landscapes

Abstract

The invention discloses a collaborative model training task configuration method facing intelligent edge computing, which is used for edge computing nodes and comprises one or more training time slots, wherein each training time slot comprises the following steps: sending a model training request to one or more mobile devices; receiving an availability status and a user data size of a current time slot reported from one or more mobile devices; determining the number of training small wheels required by the training of the mobile equipment and the interaction model participating in the training based on the previously obtained task configuration result and the current available state of each mobile equipment; and performing interactive model training with the mobile equipment participating in training until the number of the determined training small rounds is reached, constructing and solving an optimization problem aiming at minimizing the use of edge training resources according to the training effect and the user data scale reported by each mobile equipment, and obtaining a new task configuration result. Compared with other methods, the method has the advantages of much less training resource consumption and little difference in precision.

Description

Translated fromChinese
面向智能边缘计算的协同模型训练任务配置方法A collaborative model training task configuration method for intelligent edge computing

技术领域technical field

本发明涉及一种协同模型训练任务配置方法,具体涉及一种面向智能边缘计算的协同模型训练任务配置方法。The invention relates to a collaborative model training task configuration method, in particular to a collaborative model training task configuration method oriented to intelligent edge computing.

背景技术Background technique

在用户使用移动设备,如手机、平板电脑等的过程中会产生大量用户数据,包括浏览记录、打字记录以及各类日志信息等。这些数据在被分析处理后,能够帮助服务提供商进行更好的服务部署与提供。这类分析处理手段往往借助于机器学习模型。具体来说,一个机器学习模型包含模型结构和模型参数,以及该机器学习模型在某一特定数据集上所体现出的精度,如用分类模型对某一数据集进行分类,得到的正确分类比例作为模型精度。那么,服务提供商的目标,就是要利用各用户分布式产生的用户数据,对某一特定的机器学习模型进行训练,以求能够得到最好的模型精度。这样,服务提供商就可以利用这些机器学习模型对外提供更好的推断类服务。例如,在用户浏览商品的时候结合商品分类为用户进行商品推荐;在用户打字的时候,结合上下文进行热词推荐;或者在导航的时候,利用更精确的导航模型进行路线制定。When users use mobile devices, such as mobile phones and tablet computers, a large amount of user data will be generated, including browsing records, typing records, and various log information. After these data are analyzed and processed, they can help service providers to deploy and provide better services. Such analytical processing methods often rely on machine learning models. Specifically, a machine learning model includes the model structure and model parameters, as well as the accuracy of the machine learning model on a specific data set, such as the correct classification ratio obtained by classifying a data set with a classification model as model accuracy. Then, the goal of the service provider is to use the user data distributed by each user to train a specific machine learning model in order to obtain the best model accuracy. In this way, service providers can use these machine learning models to provide better inference-like services to the outside world. For example, when users browse products, they can recommend products for users in combination with product categories; when users are typing, they can recommend hot words in combination with context; or when navigating, use more accurate navigation models to formulate routes.

虽然把所有的用户数据汇聚到数据中心进行处理,就可以按照如上的训练方式得到机器学习模型。但是,在边缘环境下,这样的原始数据汇聚是被禁止的。其原因是:1)出于隐私的保护,用户往往不愿将自己的原始数据进行上传。2)服务提供商往往是租借运营商的边缘设备进行计算和传输。将各移动设备上的用户数据传输至数据中心,会带来高昂的跨域传输。这里的跨域同时包含两个含义:跨地域范围传输以及跨运营商至数据中心传输。Although all user data is aggregated into the data center for processing, the machine learning model can be obtained according to the above training method. However, in edge environments, such aggregation of raw data is prohibited. The reasons are: 1) For the protection of privacy, users are often reluctant to upload their own original data. 2) Service providers often rent operators' edge devices for computing and transmission. Transferring user data from various mobile devices to a data center can result in expensive cross-domain transfers. The cross-domain here also includes two meanings: cross-regional transmission and cross-operator-to-data center transmission.

由于边缘场景下用户使用移动设备的习惯不同,即使用设备的时间、频度不同以及使用过程中产生的数据规模、内容不同,以至于在进行分布式机器学习训练的时候存在着不确定性。即使某一时段内设备固定,所有用户数据均已生成,如何利用移动设备和边缘计算节点进行分布式机器学习训练,以能够在保证模型训练精度下尽可能节省边缘训练资源是关键问题。Due to the different habits of users using mobile devices in edge scenarios, that is, the time and frequency of using the devices, and the scale and content of data generated during use, there is uncertainty when conducting distributed machine learning training. Even if the device is fixed for a certain period of time and all user data has been generated, how to use mobile devices and edge computing nodes for distributed machine learning training to save edge training resources as much as possible while ensuring model training accuracy is a key issue.

发明内容SUMMARY OF THE INVENTION

发明目的:为了克服现有技术中存在的不足,本发明一方面提供一种面向智能边缘计算的协同模型训练任务配置方法,以解决分布式机器学习训练数据难以共享的问题,且在保证精度的情况下,尽量节省资源消耗。Purpose of the invention: In order to overcome the deficiencies in the prior art, on the one hand, the present invention provides a collaborative model training task configuration method oriented to intelligent edge computing, so as to solve the problem that the distributed machine learning training data is difficult to share, and ensure the accuracy. In this case, try to save resource consumption as much as possible.

技术方案:本发明的一种面向智能边缘计算的协同模型训练任务配置方法,用于边缘计算节点且包括一或多个训练时隙。该方法的每一训练时隙包括如下步骤:向一或多个边缘设备发送模型训练请求;接收来自所述一或多个边缘设备响应于所述模型训练请求而汇报的当前时隙的可用状态和用户数据规模;基于上一训练时隙得到的任务配置结果,从当前可用边缘设备中选定参与训练的边缘设备,并确定交互模型训练所需的训练小轮数目;与参与训练的边缘设备进行交互模型训练直至达到确定的训练小轮数目;根据训练效果和各边缘设备汇报的当前时隙的用户数据规模,构建以最小化边缘训练资源的使用为目标的优化问题并求解,得到用于下一训练时隙的新的任务配置结果。Technical solution: A collaborative model training task configuration method for intelligent edge computing of the present invention is used for edge computing nodes and includes one or more training time slots. Each training time slot of the method includes the steps of: sending a model training request to one or more edge devices; receiving the available status of the current time slot reported by the one or more edge devices in response to the model training request and user data scale; based on the task configuration results obtained in the previous training time slot, select edge devices participating in training from the currently available edge devices, and determine the number of training rounds required for interactive model training; Carry out interactive model training until the number of training rounds is reached; according to the training effect and the user data scale of the current time slot reported by each edge device, construct and solve the optimization problem aiming at minimizing the use of edge training resources, and obtain the The new task configuration result for the next training slot.

进一步地,任务配置结果包括:用于决策是否选择第i个边缘设备在训练时隙t内参与训练的参与者决策量

Figure 791043DEST_PATH_IMAGE001
和用于决策训练时隙t内训练小轮数目的辅助决策量
Figure 887044DEST_PATH_IMAGE002
。Further, the task configuration result includes: a participant decision amount for deciding whether to select thei -th edge device to participate in the training in the training time slott
Figure 791043DEST_PATH_IMAGE001
and the auxiliary decision amount used to decide the number of training rounds in the training time slott
Figure 887044DEST_PATH_IMAGE002
.

进一步地,训练时隙t内所需训练小轮数目

Figure 159894DEST_PATH_IMAGE003
通过下式计算:Further, the number of training rounds required in the training time slott
Figure 159894DEST_PATH_IMAGE003
Calculated by the following formula:

Figure 577231DEST_PATH_IMAGE003
=K
Figure 328149DEST_PATH_IMAGE002
Figure 577231DEST_PATH_IMAGE003
=K
Figure 328149DEST_PATH_IMAGE002
,

其中,K为常数。whereK is a constant.

进一步地,与参与训练的边缘设备进行交互模型训练时,训练时隙t内每一训练小轮具体包括:Further, when performing interactive model training with edge devices participating in the training, each training wheel in the training time slott specifically includes:

(1)边缘计算节点将先前训练所得的全局训练模型参数

Figure 809815DEST_PATH_IMAGE004
、各边缘设备的局部精度修正梯度
Figure 987986DEST_PATH_IMAGE005
以及全局精度修正梯度
Figure 938625DEST_PATH_IMAGE006
发送至所有可用边缘设备;参与训练的边缘设备根据接收到的数据和自身的局部精度损失函数
Figure 40704DEST_PATH_IMAGE007
分别计算各自对全局训练模型参数的更新
Figure 65292DEST_PATH_IMAGE008
t为当前交互训练的序数,j为当前训练小轮的序数,i为各边缘设备的序数;
Figure 742261DEST_PATH_IMAGE009
=0;(1) The edge computing node uses the previously trained global training model parameters
Figure 809815DEST_PATH_IMAGE004
, the local accuracy correction gradient of each edge device
Figure 987986DEST_PATH_IMAGE005
and the global accuracy correction gradient
Figure 938625DEST_PATH_IMAGE006
Sent to all available edge devices; the edge device participating in the training uses the received data and its own local accuracy loss function
Figure 40704DEST_PATH_IMAGE007
Calculate the respective updates to the global training model parameters separately
Figure 65292DEST_PATH_IMAGE008
;t is the ordinal number of the current interactive training,j is the ordinal number of the current training round, andi is the ordinal number of each edge device;
Figure 742261DEST_PATH_IMAGE009
=0;

(2)边缘计算节点在接收到所有参与训练的边缘设备发送的对全局训练模型参数的更新

Figure 101567DEST_PATH_IMAGE008
的基础上,计算得到新的全局模型参数
Figure 256605DEST_PATH_IMAGE010
并发送给所有参与训练的边缘设备进行验证;所有参与训练的边缘设备基于
Figure 169589DEST_PATH_IMAGE010
分别进行计算,各自得到新的局部精度
Figure 751880DEST_PATH_IMAGE011
、新的局部精度修正梯度
Figure 349215DEST_PATH_IMAGE012
、新的局部汇聚性能
Figure 557211DEST_PATH_IMAGE013
,并发送给边缘计算节点进行更新记录;(2) The edge computing node receives the update of the parameters of the global training model sent by all the edge devices participating in the training
Figure 101567DEST_PATH_IMAGE008
On the basis of , the new global model parameters are calculated
Figure 256605DEST_PATH_IMAGE010
and sent to all edge devices participating in training for verification; all edge devices participating in training are based on
Figure 169589DEST_PATH_IMAGE010
Calculated separately, each obtains a new local accuracy
Figure 751880DEST_PATH_IMAGE011
, new local accuracy correction gradient
Figure 349215DEST_PATH_IMAGE012
, the new local convergence performance
Figure 557211DEST_PATH_IMAGE013
, and send it to the edge computing node to update the record;

(3)边缘计算节点基于接收到的各边缘设备的局部精度修正梯度

Figure 556391DEST_PATH_IMAGE012
计算得到新的全局精度修正梯度
Figure 309583DEST_PATH_IMAGE014
;(3) The edge computing node corrects the gradient based on the received local accuracy of each edge device
Figure 556391DEST_PATH_IMAGE012
Calculate the new global accuracy correction gradient
Figure 309583DEST_PATH_IMAGE014
;

(4)若当前训练小轮达到当前训练时隙t所需的训练小轮数目

Figure 207263DEST_PATH_IMAGE003
,边缘计算节点还将最新的全局模型参数
Figure 641787DEST_PATH_IMAGE015
发送给未参与训练的边缘设备;未参与训练的边缘设备基于
Figure 557790DEST_PATH_IMAGE015
计算得到各自的第
Figure 934414DEST_PATH_IMAGE003
个训练小轮后新的局部精度
Figure 568657DEST_PATH_IMAGE016
,并发送给边缘计算节点进行更新记录。(4) If the current training round reaches the number of training rounds required for the current training time slott
Figure 207263DEST_PATH_IMAGE003
, the edge computing node will also update the latest global model parameters
Figure 641787DEST_PATH_IMAGE015
Sent to edge devices not participating in training; edge devices not participating in training are based on
Figure 557790DEST_PATH_IMAGE015
Calculate the respective
Figure 934414DEST_PATH_IMAGE003
New local accuracy after training epochs
Figure 568657DEST_PATH_IMAGE016
, and send it to the edge computing node to update the record.

进一步地,步骤(1)中,参与训练的边缘设备根据接收到的数据分别计算各自对全局训练模型参数的更新

Figure 557604DEST_PATH_IMAGE008
,具体包括:各参与训练的边缘设备利用获取的
Figure 328114DEST_PATH_IMAGE004
Figure 626372DEST_PATH_IMAGE005
和自身的局部精度损失函数
Figure 997179DEST_PATH_IMAGE017
构建优化函数
Figure 773505DEST_PATH_IMAGE018
,并以最小化所述优化函数
Figure 664101DEST_PATH_IMAGE018
的方式得到
Figure 883992DEST_PATH_IMAGE008
;所述优化函数
Figure 227248DEST_PATH_IMAGE018
表示为:Further, in step (1), the edge devices participating in the training calculate their respective updates to the parameters of the global training model according to the received data.
Figure 557604DEST_PATH_IMAGE008
, specifically including: each edge device participating in the training uses the acquired data
Figure 328114DEST_PATH_IMAGE004
,
Figure 626372DEST_PATH_IMAGE005
and its own local accuracy loss function
Figure 997179DEST_PATH_IMAGE017
Build optimization functions
Figure 773505DEST_PATH_IMAGE018
, and to minimize the optimization function
Figure 664101DEST_PATH_IMAGE018
way to get
Figure 883992DEST_PATH_IMAGE008
; the optimization function
Figure 227248DEST_PATH_IMAGE018
Expressed as:

Figure 322112DEST_PATH_IMAGE019
Figure 322112DEST_PATH_IMAGE019
,

其中

Figure 67214DEST_PATH_IMAGE020
Figure 504012DEST_PATH_IMAGE021
均为确定的参数。in
Figure 67214DEST_PATH_IMAGE020
,
Figure 504012DEST_PATH_IMAGE021
are definite parameters.

进一步地,步骤(2)中,新的全局模型参数

Figure 17121DEST_PATH_IMAGE010
通过下式计算得到:Further, in step (2), the new global model parameters
Figure 17121DEST_PATH_IMAGE010
It is calculated by the following formula:

Figure 197566DEST_PATH_IMAGE022
Figure 197566DEST_PATH_IMAGE022
,

其中,

Figure 437DEST_PATH_IMAGE023
为当前训练时隙t内可用边缘设备的集合,
Figure 122983DEST_PATH_IMAGE024
为训练时隙t中用于指示第i个边缘设备是否参与训练的变量,
Figure 83242DEST_PATH_IMAGE024
等于0或1。in,
Figure 437DEST_PATH_IMAGE023
is the set of available edge devices in the current training time slott ,
Figure 122983DEST_PATH_IMAGE024
is the variable used to indicate whether thei -th edge device participates in the training in the training time slott ,
Figure 83242DEST_PATH_IMAGE024
Equal to 0 or 1.

进一步地,步骤(2)中,新的局部精度

Figure 270641DEST_PATH_IMAGE011
是由各边缘设备将
Figure 442865DEST_PATH_IMAGE010
代入自身的局部精度损失函数
Figure 424728DEST_PATH_IMAGE017
而得到;新的局部精度修正梯度
Figure 229873DEST_PATH_IMAGE012
是基于新的局部精度
Figure 706115DEST_PATH_IMAGE011
而得到;新的局部汇聚性能
Figure 280316DEST_PATH_IMAGE025
是通过下式得到:Further, in step (2), the new local precision
Figure 270641DEST_PATH_IMAGE011
by each edge device
Figure 442865DEST_PATH_IMAGE010
Substitute into its own local precision loss function
Figure 424728DEST_PATH_IMAGE017
and get; the new local accuracy correction gradient
Figure 229873DEST_PATH_IMAGE012
is based on the new local precision
Figure 706115DEST_PATH_IMAGE011
and get; new local convergence performance
Figure 280316DEST_PATH_IMAGE025
is obtained by:

Figure 682348DEST_PATH_IMAGE026
Figure 682348DEST_PATH_IMAGE026
.

进一步地,步骤(3)中,新的全局精度修正梯度

Figure 443630DEST_PATH_IMAGE027
通过下式得到:Further, in step (3), the new global accuracy correction gradient
Figure 443630DEST_PATH_IMAGE027
It is obtained by the following formula:

Figure 35149DEST_PATH_IMAGE028
Figure 35149DEST_PATH_IMAGE028
,

其中,

Figure 146412DEST_PATH_IMAGE023
为当前训练时隙t内可用边缘设备的集合。in,
Figure 146412DEST_PATH_IMAGE023
is the set of available edge devices in the current training time slott .

进一步地,所述训练效果包括:训练时隙t内达到确定的训练小轮数目

Figure 532394DEST_PATH_IMAGE003
后的全局模型参数
Figure 46552DEST_PATH_IMAGE015
、各边缘设备实际观测到的局部汇聚性能
Figure 363133DEST_PATH_IMAGE029
和各边缘设备在各训练小轮中更新的局部精度
Figure 646347DEST_PATH_IMAGE011
;其中,
Figure 406492DEST_PATH_IMAGE030
。Further, the training effect includes: reaching a certain number of training small rounds in the training time slott
Figure 532394DEST_PATH_IMAGE003
global model parameters after
Figure 46552DEST_PATH_IMAGE015
, the actual observed local convergence performance of each edge device
Figure 363133DEST_PATH_IMAGE029
and the local accuracy updated by each edge device in each training epoch
Figure 646347DEST_PATH_IMAGE011
;in,
Figure 406492DEST_PATH_IMAGE030
.

进一步地,所述优化问题表示为:Further, the optimization problem is expressed as:

目标函数:

Figure 158679DEST_PATH_IMAGE031
,Objective function:
Figure 158679DEST_PATH_IMAGE031
,

约束条件:Restrictions:

1)

Figure 764103DEST_PATH_IMAGE032
,1)
Figure 764103DEST_PATH_IMAGE032
,

2)

Figure 167403DEST_PATH_IMAGE033
,2)
Figure 167403DEST_PATH_IMAGE033
,

3)

Figure 347717DEST_PATH_IMAGE034
,3)
Figure 347717DEST_PATH_IMAGE034
,

4)

Figure 570888DEST_PATH_IMAGE035
,4)
Figure 570888DEST_PATH_IMAGE035
,

5)

Figure 42321DEST_PATH_IMAGE036
,5)
Figure 42321DEST_PATH_IMAGE036
,

其中,

Figure 988542DEST_PATH_IMAGE037
为用于决策训练时隙t+1内训练小轮数目的辅助决策量;
Figure 887228DEST_PATH_IMAGE038
为训练时隙t内可用边缘设备的集合,由汇报的当前可用状态确定;
Figure 50225DEST_PATH_IMAGE039
为用于决策是否选择第i个边缘设备在训练时隙t+1内参与训练的参与者决策量;m为移动网络可并发传输的容量上限;
Figure 13764DEST_PATH_IMAGE040
为在当前边缘网络中对模型参数和梯度进行一次传输的规模;
Figure 543054DEST_PATH_IMAGE041
为训练时隙t内边缘网络中的可用带宽;
Figure 330750DEST_PATH_IMAGE042
为训练时隙t内第i个边缘设备针对单个数据样本的计算代价;
Figure 466196DEST_PATH_IMAGE043
为训练时隙t内第i个边缘设备的用户数据规模;
Figure 545011DEST_PATH_IMAGE044
为全局精度损失函数,且
Figure 465825DEST_PATH_IMAGE044
=
Figure 706313DEST_PATH_IMAGE045
Figure 843902DEST_PATH_IMAGE046
Figure 460828DEST_PATH_IMAGE047
为设定的全局精度损失;
Figure 485416DEST_PATH_IMAGE048
为当前训练时隙t内交互训练后所有边缘设备的局部汇聚性能的最大值,且
Figure 647538DEST_PATH_IMAGE048
=
Figure 554314DEST_PATH_IMAGE049
=
Figure 522402DEST_PATH_IMAGE050
Figure 401496DEST_PATH_IMAGE029
为训练时隙t内达到确定的训练小轮数目
Figure 701896DEST_PATH_IMAGE003
后第i个边缘设备实际观测到的局部汇聚性能。in,
Figure 988542DEST_PATH_IMAGE037
is the auxiliary decision amount used to decide the number of training rounds in the training time slott +1;
Figure 887228DEST_PATH_IMAGE038
is the set of available edge devices in the training time slott , determined by the current available state reported;
Figure 50225DEST_PATH_IMAGE039
is the decision amount of the participants used to decide whether to select thei -th edge device to participate in the training in the training time slott +1;m is the upper limit of the concurrent transmission capacity of the mobile network;
Figure 13764DEST_PATH_IMAGE040
is the size of one transfer of model parameters and gradients in the current edge network;
Figure 543054DEST_PATH_IMAGE041
is the available bandwidth in the edge network within the training time slott ;
Figure 330750DEST_PATH_IMAGE042
is the computational cost of thei -th edge device for a single data sample in the training time slott ;
Figure 466196DEST_PATH_IMAGE043
is the user data size of thei -th edge device in the training time slott ;
Figure 545011DEST_PATH_IMAGE044
is the global precision loss function, and
Figure 465825DEST_PATH_IMAGE044
=
Figure 706313DEST_PATH_IMAGE045
,
Figure 843902DEST_PATH_IMAGE046
;
Figure 460828DEST_PATH_IMAGE047
is the set global precision loss;
Figure 485416DEST_PATH_IMAGE048
is the maximum local convergence performance of all edge devices after interactive training in the current training time slott , and
Figure 647538DEST_PATH_IMAGE048
=
Figure 554314DEST_PATH_IMAGE049
=
Figure 522402DEST_PATH_IMAGE050
,
Figure 401496DEST_PATH_IMAGE029
A certain number of training epochs is reached within the training time slott
Figure 701896DEST_PATH_IMAGE003
The actual observed local convergence performance of the lasti -th edge device.

有益效果:与现有技术相比,本发明具有以下优点:Beneficial effect: Compared with the prior art, the present invention has the following advantages:

1、通过在交互训练过程中构建优化函数进行更新,边缘设备无需直接发送自身模型参数的原始数据至边缘计算节点,而是发送与自身原始数据相关的更新数据。这可以解决由于隐私的保护,用户往往不愿将自己的原始数据进行上传的问题,适用于分布式机器学习训练。1. By constructing an optimization function for updating in the interactive training process, the edge device does not need to directly send the original data of its own model parameters to the edge computing node, but sends the updated data related to its own original data. This can solve the problem that users are often reluctant to upload their own original data due to privacy protection, and is suitable for distributed machine learning training.

2、基于交互训练的效果构建最小化边缘训练资源的使用为目标的优化问题并求解,从而决策得到下一训练时隙的任务配置,包括下一训练时隙各边缘设备的选择偏好和训练小轮数目。根据决策结果,从可用边缘设备中选取参与训练的边缘设备进行训练,而不需要使所有边缘设备均参与训练,这使本申请的协同训练任务配置方法至少减少27%的资源消耗开销,训练精度则最多降低4%。换言之,本发明能够在保证训练精度的同时,大幅度减少训练资源消耗。2. Based on the effect of interactive training, an optimization problem with the goal of minimizing the use of edge training resources is constructed and solved, so as to obtain the task configuration of the next training time slot, including the selection preference and training time of each edge device in the next training time slot. number of rounds. According to the decision result, the edge devices participating in the training are selected from the available edge devices for training, and all edge devices do not need to participate in the training, which reduces the resource consumption overhead by at least 27% in the collaborative training task configuration method of the present application, and the training accuracy It is reduced by up to 4%. In other words, the present invention can greatly reduce the consumption of training resources while ensuring the training accuracy.

附图说明Description of drawings

图1是面向智能边缘计算的协同模型训练系统的结构示意图;Figure 1 is a schematic structural diagram of a collaborative model training system for intelligent edge computing;

图2是应用任务配置方法后实际使用各边缘计算资源作训练的训练资源花费变化情况;Figure 2 shows the change in the cost of training resources actually using each edge computing resource for training after applying the task configuration method;

图3是应用任务配置方法后全局精度的变化情况;Figure 3 shows the change of global accuracy after applying the task configuration method;

图4是应用任务配置方法后最大局部汇聚性能的变化情况。Figure 4 shows the change of the maximum local convergence performance after applying the task configuration method.

具体实施方式Detailed ways

下面结合附图对本发明所公开的方法做进一步的详细介绍。The method disclosed in the present invention will be further described in detail below with reference to the accompanying drawings.

本发明的面向智能边缘计算的协同模型训练任务配置方法用于边缘计算节点,且包括一或多个训练时隙。每一训练时隙包括如下步骤:The intelligent edge computing-oriented collaborative model training task configuration method of the present invention is used for edge computing nodes, and includes one or more training time slots. Each training slot includes the following steps:

S1:向一或多个边缘设备发送模型训练请求。这里的边缘设备可以是连接边缘计算节点的移动设备、笔记本电脑等。S1: Send a model training request to one or more edge devices. The edge devices here can be mobile devices, laptops, etc. connected to edge computing nodes.

S2:接收来自所述一或多个边缘设备响应于所述模型训练请求而汇报的当前时隙的可用状态和用户数据规模。S2: Receive the available status and user data scale of the current time slot reported by the one or more edge devices in response to the model training request.

S3:基于上一训练时隙得到的任务配置结果,从当前可用边缘设备中选定参与训练的边缘设备,并确定交互模型训练所需的训练小轮数目。其中,该步骤中上一训练时隙得到的任务配置结果包括:用于决策是否选择第i个边缘设备在训练时隙t内参与训练的参与者决策量

Figure 564810DEST_PATH_IMAGE001
和用于决策训练时隙t内训练小轮数目的辅助决策量
Figure 523539DEST_PATH_IMAGE002
。S3: Based on the task configuration result obtained in the previous training time slot, select edge devices participating in training from currently available edge devices, and determine the number of training rounds required for training the interactive model. Wherein, the task configuration result obtained in the previous training time slot in this step includes: the decision amount of the participants used to decide whether to select thei -th edge device to participate in the training in the training time slott
Figure 564810DEST_PATH_IMAGE001
and the auxiliary decision amount used to decide the number of training rounds in the training time slott
Figure 523539DEST_PATH_IMAGE002
.

其中,当前训练时隙t内所需训练小轮数目

Figure 556608DEST_PATH_IMAGE003
通过下式计算:Among them, the number of training rounds required in the current training time slott
Figure 556608DEST_PATH_IMAGE003
Calculated by the following formula:

Figure 513063DEST_PATH_IMAGE003
=K
Figure 643699DEST_PATH_IMAGE002
Figure 513063DEST_PATH_IMAGE003
=K
Figure 643699DEST_PATH_IMAGE002
,

其中,K为常数。whereK is a constant.

当前训练时隙t内参与训练的边缘设备则基于参与者决策量

Figure 343802DEST_PATH_IMAGE001
从处于可用状态的边缘设备中选定。参与者决策量
Figure 213800DEST_PATH_IMAGE051
实际上用于表征对各边缘设备选择的偏好程度,该偏好程度作为概率用于在各边缘设备间进行选择以确定参与训练的边缘设备,确定时具体可以采用Pair-wise Rounding算法或者DepRound算法求解得到。各训练时隙的参与者决策量
Figure 403473DEST_PATH_IMAGE001
由上一训练时隙决策得到,具体计算方法会在后续内容中进行介绍。对于初始训练时隙,由于没有上一训练时隙的训练效果作参考,因此将选择所有可用边缘设备作为参与者。The edge devices participating in the training in the current training time slott are based on the decision volume of the participants
Figure 343802DEST_PATH_IMAGE001
Select from edge devices that are available. Participant decision volume
Figure 213800DEST_PATH_IMAGE051
In fact, it is used to characterize the degree of preference for the selection of each edge device. The degree of preference is used as a probability to select among the edge devices to determine the edge device participating in the training. When determining, the Pair-wise Rounding algorithm or the DepRound algorithm can be used to solve the problem. get. Participant decision volume for each training slot
Figure 403473DEST_PATH_IMAGE001
It is obtained from the previous training time slot decision, and the specific calculation method will be introduced in the subsequent content. For the initial training slot, since there is no training effect from the previous training slot for reference, all available edge devices will be selected as participants.

S4:与参与训练的边缘设备进行交互模型训练直至达到确定的训练小轮数目。S4: Perform interactive model training with edge devices participating in training until a certain number of training rounds is reached.

在步骤S4中,与参与训练的边缘设备进行交互模型训练时,每一训练小轮具体包括如下过程:In step S4, when performing interactive model training with edge devices participating in the training, each training round specifically includes the following processes:

S41:边缘计算节点将先前训练所得的全局训练模型参数

Figure 975400DEST_PATH_IMAGE004
、各边缘设备的局部精度修正梯度
Figure 10352DEST_PATH_IMAGE005
以及全局精度修正梯度
Figure 780862DEST_PATH_IMAGE006
发送至所有可用边缘设备;参与训练的边缘设备根据接收到的数据和自身的局部精度损失函数
Figure 79119DEST_PATH_IMAGE007
分别计算各自对全局训练模型参数的更新
Figure 685812DEST_PATH_IMAGE008
t为当前交互训练的序数,j为当前训练小轮的序数,i为各边缘设备的序数;
Figure 727717DEST_PATH_IMAGE009
=0。S41: The edge computing node uses the previously trained global training model parameters
Figure 975400DEST_PATH_IMAGE004
, the local accuracy correction gradient of each edge device
Figure 10352DEST_PATH_IMAGE005
and the global accuracy correction gradient
Figure 780862DEST_PATH_IMAGE006
Sent to all available edge devices; the edge device participating in the training uses the received data and its own local accuracy loss function
Figure 79119DEST_PATH_IMAGE007
Calculate the respective updates to the global training model parameters separately
Figure 685812DEST_PATH_IMAGE008
;t is the ordinal number of the current interactive training,j is the ordinal number of the current training round, andi is the ordinal number of each edge device;
Figure 727717DEST_PATH_IMAGE009
=0.

其中,参与训练的边缘设备根据接收到的数据分别计算各自对全局训练模型参数的更新

Figure 618313DEST_PATH_IMAGE008
,具体包括:Among them, the edge devices participating in the training calculate their respective updates to the parameters of the global training model according to the received data.
Figure 618313DEST_PATH_IMAGE008
, including:

各参与训练的边缘设备利用获取的

Figure 336739DEST_PATH_IMAGE004
Figure 679996DEST_PATH_IMAGE005
和自身的局部精度损失函数
Figure 4886DEST_PATH_IMAGE017
构建优化函数
Figure 749988DEST_PATH_IMAGE018
,并以最小化所述优化函数
Figure 390048DEST_PATH_IMAGE018
的方式得到
Figure 204289DEST_PATH_IMAGE008
;其中,优化函数
Figure 587997DEST_PATH_IMAGE018
表示为:Each edge device participating in the training uses the obtained
Figure 336739DEST_PATH_IMAGE004
,
Figure 679996DEST_PATH_IMAGE005
and its own local accuracy loss function
Figure 4886DEST_PATH_IMAGE017
Build optimization functions
Figure 749988DEST_PATH_IMAGE018
, and to minimize the optimization function
Figure 390048DEST_PATH_IMAGE018
way to get
Figure 204289DEST_PATH_IMAGE008
; where the optimization function
Figure 587997DEST_PATH_IMAGE018
Expressed as:

Figure 141600DEST_PATH_IMAGE019
Figure 141600DEST_PATH_IMAGE019
,

其中

Figure 14878DEST_PATH_IMAGE020
Figure 254098DEST_PATH_IMAGE021
均为确定的参数。in
Figure 14878DEST_PATH_IMAGE020
,
Figure 254098DEST_PATH_IMAGE021
are definite parameters.

通过构建优化函数

Figure 503814DEST_PATH_IMAGE018
,边缘设备无需直接发送自身模型参数的原始数据至边缘计算节点,而是发送与自身原始数据相关的更新数据。这可以解决由于隐私的保护,用户往往不愿将自己的原始数据进行上传的问题,适用于分布式机器学习训练。By building an optimization function
Figure 503814DEST_PATH_IMAGE018
, the edge device does not need to directly send the original data of its own model parameters to the edge computing node, but sends updated data related to its own original data. This can solve the problem that users are often reluctant to upload their own original data due to privacy protection, and is suitable for distributed machine learning training.

S42:边缘计算节点在接收到所有参与训练的边缘设备发送的对全局训练模型参数的更新

Figure 161192DEST_PATH_IMAGE008
的基础上,计算得到新的全局模型参数
Figure 956103DEST_PATH_IMAGE010
并发送给所有参与训练的边缘设备进行验证;所有参与训练的边缘设备基于
Figure 433352DEST_PATH_IMAGE010
分别进行计算,各自得到新的局部精度
Figure 221180DEST_PATH_IMAGE011
、新的局部精度修正梯度
Figure 779069DEST_PATH_IMAGE012
、新的局部汇聚性能
Figure 197412DEST_PATH_IMAGE013
,并发送给边缘计算节点进行更新记录。S42: The edge computing node receives the update of the parameters of the global training model sent by all the edge devices participating in the training
Figure 161192DEST_PATH_IMAGE008
On the basis of , the new global model parameters are calculated
Figure 956103DEST_PATH_IMAGE010
and sent to all edge devices participating in training for verification; all edge devices participating in training are based on
Figure 433352DEST_PATH_IMAGE010
Calculated separately, each obtains a new local accuracy
Figure 221180DEST_PATH_IMAGE011
, new local accuracy correction gradient
Figure 779069DEST_PATH_IMAGE012
, the new local convergence performance
Figure 197412DEST_PATH_IMAGE013
, and send it to the edge computing node to update the record.

该步骤中:新的全局模型参数

Figure 178268DEST_PATH_IMAGE010
通过下式计算得到:In this step: New global model parameters
Figure 178268DEST_PATH_IMAGE010
It is calculated by the following formula:

Figure 238628DEST_PATH_IMAGE022
Figure 238628DEST_PATH_IMAGE022
,

其中,

Figure 932915DEST_PATH_IMAGE023
为当前训练时隙t内可用边缘设备的集合,
Figure 505848DEST_PATH_IMAGE024
为训练时隙t中用于指示第i个边缘设备是否参与训练的变量,
Figure 20006DEST_PATH_IMAGE024
等于0或1且基于参与者决策量
Figure 87319DEST_PATH_IMAGE052
得到。一般来说,
Figure 380985DEST_PATH_IMAGE024
等于0时指示对应边缘设备i未参与训练,
Figure 875551DEST_PATH_IMAGE024
等于1时指示对应边缘设备i参与训练,或者也可以反过来使用。in,
Figure 932915DEST_PATH_IMAGE023
is the set of available edge devices in the current training time slott ,
Figure 505848DEST_PATH_IMAGE024
is the variable used to indicate whether thei -th edge device participates in the training in the training time slott ,
Figure 20006DEST_PATH_IMAGE024
Equal to 0 or 1 and based on participant decision volume
Figure 87319DEST_PATH_IMAGE052
get. Generally speaking,
Figure 380985DEST_PATH_IMAGE024
When it is equal to 0, it indicates that the corresponding edge devicei does not participate in the training,
Figure 875551DEST_PATH_IMAGE024
When it is equal to 1, it indicates that the corresponding edge devicei participates in the training, or it can be used in reverse.

新的局部精度

Figure 877005DEST_PATH_IMAGE011
是由各边缘设备将
Figure 794014DEST_PATH_IMAGE010
代入自身的局部精度损失函数
Figure 134997DEST_PATH_IMAGE017
而得到;新的局部精度修正梯度
Figure 862782DEST_PATH_IMAGE012
是基于新的局部精度
Figure 39947DEST_PATH_IMAGE011
而得到;新的局部汇聚性能
Figure 776959DEST_PATH_IMAGE025
是通过下式得到:new local precision
Figure 877005DEST_PATH_IMAGE011
by each edge device
Figure 794014DEST_PATH_IMAGE010
Substitute into its own local precision loss function
Figure 134997DEST_PATH_IMAGE017
and get; the new local accuracy correction gradient
Figure 862782DEST_PATH_IMAGE012
is based on the new local precision
Figure 39947DEST_PATH_IMAGE011
and get; new local convergence performance
Figure 776959DEST_PATH_IMAGE025
is obtained by:

Figure 221716DEST_PATH_IMAGE026
Figure 221716DEST_PATH_IMAGE026
.

S43:边缘计算节点基于接收到的各局部精度修正梯度

Figure 120402DEST_PATH_IMAGE053
计算得到新的全局精度修正梯度
Figure 830869DEST_PATH_IMAGE054
。S43: The edge computing node corrects the gradient based on the received local accuracy
Figure 120402DEST_PATH_IMAGE053
Calculate the new global accuracy correction gradient
Figure 830869DEST_PATH_IMAGE054
.

该步骤中,新的全局精度修正梯度

Figure 794408DEST_PATH_IMAGE027
通过下式得到:In this step, the new global accuracy correction gradient
Figure 794408DEST_PATH_IMAGE027
It is obtained by the following formula:

Figure 906720DEST_PATH_IMAGE028
Figure 906720DEST_PATH_IMAGE028
,

其中,

Figure 179570DEST_PATH_IMAGE023
为当前训练时隙t内可用边缘设备的集合。in,
Figure 179570DEST_PATH_IMAGE023
is the set of available edge devices in the current training time slott .

若当前训练小轮达到当前训练时隙t所需的训练小轮数目

Figure 626601DEST_PATH_IMAGE003
,边缘计算节点还将最新的全局模型参数
Figure 643098DEST_PATH_IMAGE015
发送给未参与训练的边缘设备;未参与训练的边缘设备基于
Figure 609917DEST_PATH_IMAGE015
计算得到各自的第
Figure 601138DEST_PATH_IMAGE003
个训练小轮后新的局部精度
Figure 223880DEST_PATH_IMAGE016
,并发送给边缘计算节点进行更新记录。If the current training round reaches the number of training rounds required for the current training time slott
Figure 626601DEST_PATH_IMAGE003
, the edge computing node will also update the latest global model parameters
Figure 643098DEST_PATH_IMAGE015
Sent to edge devices not participating in training; edge devices not participating in training are based on
Figure 609917DEST_PATH_IMAGE015
Calculate the respective
Figure 601138DEST_PATH_IMAGE003
New local accuracy after training epochs
Figure 223880DEST_PATH_IMAGE016
, and send it to the edge computing node to update the record.

S5:根据训练效果和各边缘设备汇报的当前时隙的用户数据规模,构建以最小化边缘训练资源的使用为目标的优化问题并求解,得到用于下一训练时隙的新的任务配置结果。S5: According to the training effect and the user data scale of the current time slot reported by each edge device, construct and solve an optimization problem aiming at minimizing the use of edge training resources, and obtain a new task configuration result for the next training time slot .

在步骤S5中,训练效果包括:训练时隙t内达到确定的训练小轮数目

Figure 840806DEST_PATH_IMAGE003
后的全局模型参数
Figure 114662DEST_PATH_IMAGE015
、各边缘设备实际观测到的局部汇聚性能
Figure 791631DEST_PATH_IMAGE029
和各边缘设备在各训练小轮中更新的局部精度
Figure 646542DEST_PATH_IMAGE011
;其中,
Figure 67159DEST_PATH_IMAGE030
。In step S5, the training effect includes: reaching a certain number of training small rounds within the training time slott
Figure 840806DEST_PATH_IMAGE003
global model parameters after
Figure 114662DEST_PATH_IMAGE015
, the actual observed local convergence performance of each edge device
Figure 791631DEST_PATH_IMAGE029
and the local accuracy updated by each edge device in each training epoch
Figure 646542DEST_PATH_IMAGE011
;in,
Figure 67159DEST_PATH_IMAGE030
.

边缘计算节点的总体目标旨在为所有训练(

Figure 742991DEST_PATH_IMAGE055
次训练时隙的训练),在各自训练满足精度的条件下最小化边缘训练资源的使用,因此,建立的优化问题如下:The overall goal of edge computing nodes is to provide
Figure 742991DEST_PATH_IMAGE055
training time slots), and minimize the use of edge training resources under the condition that the respective training satisfies the accuracy. Therefore, the established optimization problem is as follows:

优化目标:

Figure 777812DEST_PATH_IMAGE056
;optimize the target:
Figure 777812DEST_PATH_IMAGE056
;

约束条件:Restrictions:

1) 对于边缘传输限制:

Figure 437464DEST_PATH_IMAGE057
1) For edge transfer restrictions:
Figure 437464DEST_PATH_IMAGE057

2) 对于参与者选择控制:

Figure 333876DEST_PATH_IMAGE058
2) For participant selection control:
Figure 333876DEST_PATH_IMAGE058

3) 对于全局后验精度要求:

Figure 146105DEST_PATH_IMAGE059
3) For global posterior accuracy requirements:
Figure 146105DEST_PATH_IMAGE059

4) 对于决策的定义域限制:

Figure 836980DEST_PATH_IMAGE060
4) Domain restrictions for decision making:
Figure 836980DEST_PATH_IMAGE060

式中,

Figure 983928DEST_PATH_IMAGE003
为训练时隙t内训练所需的小轮数目,
Figure 933298DEST_PATH_IMAGE061
为训练所得的全局汇聚精度,
Figure 583723DEST_PATH_IMAGE062
为全局精度损失;
Figure 773395DEST_PATH_IMAGE023
为训练时隙t内可用边缘设备的集合,由汇报的当前可用状态确定;
Figure 830475DEST_PATH_IMAGE001
为用于决策是否选择第i个边缘设备在训练时隙t内参与训练的参与者决策量;
Figure 131007DEST_PATH_IMAGE040
为在当前边缘网络中对模型参数和梯度进行一次传输的规模;
Figure 354047DEST_PATH_IMAGE041
为训练时隙t内边缘网络中的可用带宽;
Figure 449041DEST_PATH_IMAGE042
为训练时隙t内第i个边缘设备针对单个数据样本的计算代价;
Figure 305002DEST_PATH_IMAGE043
为训练时隙t内第i个边缘设备的用户数据规模;m为移动网络可并发传输的容量上限;
Figure 97640DEST_PATH_IMAGE048
为当前训练时隙t内交互训练后所有边缘设备的局部汇聚性能的最大值,且
Figure 988235DEST_PATH_IMAGE048
=
Figure 457394DEST_PATH_IMAGE049
=
Figure 49918DEST_PATH_IMAGE050
Figure 629935DEST_PATH_IMAGE029
为训练时隙t内达到确定的训练小轮数目
Figure 109458DEST_PATH_IMAGE003
后第i个边缘设备实际观测到的局部汇聚性能;
Figure 580145DEST_PATH_IMAGE015
为训练时隙t
Figure 348381DEST_PATH_IMAGE003
个训练小轮后的模型参数;
Figure 794406DEST_PATH_IMAGE044
为全局精度损失函数,且
Figure 580965DEST_PATH_IMAGE044
=
Figure 454244DEST_PATH_IMAGE045
Figure 460508DEST_PATH_IMAGE046
。In the formula,
Figure 983928DEST_PATH_IMAGE003
is the number of epochs required for training in training time slott ,
Figure 933298DEST_PATH_IMAGE061
is the global pooling accuracy obtained from training,
Figure 583723DEST_PATH_IMAGE062
is the global precision loss;
Figure 773395DEST_PATH_IMAGE023
is the set of available edge devices in the training time slott , determined by the current available state reported;
Figure 830475DEST_PATH_IMAGE001
is the decision amount of the participants used to decide whether to select thei -th edge device to participate in the training in the training time slott ;
Figure 131007DEST_PATH_IMAGE040
is the size of one transfer of model parameters and gradients in the current edge network;
Figure 354047DEST_PATH_IMAGE041
is the available bandwidth in the edge network within the training time slott ;
Figure 449041DEST_PATH_IMAGE042
is the computational cost of thei -th edge device for a single data sample in the training time slott ;
Figure 305002DEST_PATH_IMAGE043
is the user data scale of thei -th edge device in the training time slott ;m is the upper limit of the concurrent transmission capacity of the mobile network;
Figure 97640DEST_PATH_IMAGE048
is the maximum local convergence performance of all edge devices after interactive training in the current training time slott , and
Figure 988235DEST_PATH_IMAGE048
=
Figure 457394DEST_PATH_IMAGE049
=
Figure 49918DEST_PATH_IMAGE050
,
Figure 629935DEST_PATH_IMAGE029
A certain number of training epochs is reached within the training time slott
Figure 109458DEST_PATH_IMAGE003
The actual observed local convergence performance of the lasti -th edge device;
Figure 580145DEST_PATH_IMAGE015
for the training time slott
Figure 348381DEST_PATH_IMAGE003
model parameters after a training round;
Figure 794406DEST_PATH_IMAGE044
is the global precision loss function, and
Figure 580965DEST_PATH_IMAGE044
=
Figure 454244DEST_PATH_IMAGE045
,
Figure 460508DEST_PATH_IMAGE046
.

由于是在线场景,决策时无法准确观测到实际的训练效果,因此上述优化问题在实际中只能分解到每一训练时隙中,进行每一次子问题的求解。再者,每一次子问题的求解中,并不能提前观测到该次训练的局部汇聚精度

Figure 444644DEST_PATH_IMAGE029
=
Figure 898760DEST_PATH_IMAGE050
,以及全局汇聚精度
Figure 129890DEST_PATH_IMAGE061
。因此,需要利用上一训练时隙各边缘设备的训练效果作为参考,以近似替代当前还未训练无法得到的全局和局部汇聚精度。综上,对于训练时隙t内的训练,子问题实际为:Because it is an online scene, the actual training effect cannot be accurately observed during decision-making, so the above optimization problem can only be decomposed into each training time slot in practice, and each sub-problem can be solved. Furthermore, in the solution of each sub-problem, the local convergence accuracy of the training cannot be observed in advance.
Figure 444644DEST_PATH_IMAGE029
=
Figure 898760DEST_PATH_IMAGE050
, and the global pooling accuracy
Figure 129890DEST_PATH_IMAGE061
. Therefore, it is necessary to use the training effect of each edge device in the previous training time slot as a reference to approximate the global and local convergence accuracy that cannot be obtained without training currently. In summary, for the training in the training time slott , the sub-problems are actually:

目标函数:

Figure 935035DEST_PATH_IMAGE031
,Objective function:
Figure 935035DEST_PATH_IMAGE031
,

约束条件:Restrictions:

1)

Figure 660545DEST_PATH_IMAGE032
,1)
Figure 660545DEST_PATH_IMAGE032
,

2)

Figure 985478DEST_PATH_IMAGE033
,2)
Figure 985478DEST_PATH_IMAGE033
,

3)

Figure 138242DEST_PATH_IMAGE034
,3)
Figure 138242DEST_PATH_IMAGE034
,

4)

Figure 430683DEST_PATH_IMAGE035
,4)
Figure 430683DEST_PATH_IMAGE035
,

5)

Figure 943573DEST_PATH_IMAGE036
,5)
Figure 943573DEST_PATH_IMAGE036
,

其中,

Figure 372280DEST_PATH_IMAGE037
为用于决策训练时隙t+1内训练小轮数目的辅助决策量;
Figure 712257DEST_PATH_IMAGE039
为用于决策是否选择第i个边缘设备在训练时隙t+1内参与训练的参与者决策量。in,
Figure 372280DEST_PATH_IMAGE037
is the auxiliary decision amount used to decide the number of training rounds in the training time slott +1;
Figure 712257DEST_PATH_IMAGE039
is the amount of participant decision-making used to decide whether to select thei -th edge device to participate in training in training slott +1.

在上述子问题求解中,虽然目标函数里

Figure 226415DEST_PATH_IMAGE039
是二次,但含义和没有二次一样,二次是为了说明在实数域上是凸函数。上述子问题中
Figure 293728DEST_PATH_IMAGE039
可利用诸如IPOPT+AMPL等成熟的求解工具进行求解。In the above sub-problem solution, although the objective function
Figure 226415DEST_PATH_IMAGE039
It is quadratic, but the meaning is the same as no quadratic. The quadratic is to illustrate that it is a convex function in the real number field. in the above sub-problems
Figure 293728DEST_PATH_IMAGE039
It can be solved using mature solving tools such as IPOPT+AMPL.

图1以四个边缘设备的选择为例展示了本发明的面向智能边缘计算的协同模型训练系统的结构,其中所有的边缘设备均与同一个边缘计算节点相连并进行数据交互,且边缘网络能够允许传输的最大容量可以包含四个边缘设备;下面以两次全局模型训练为例,对本发明的面向智能边缘计算的协同模型训练任务配置方法作进一步的说明:Fig. 1 shows the structure of the intelligent edge computing-oriented collaborative model training system of the present invention by taking the selection of four edge devices as an example, wherein all edge devices are connected to the same edge computing node and perform data exchange, and the edge network can The maximum capacity allowed for transmission can include four edge devices; the following takes two global model training as an example to further illustrate the method for configuring the collaborative model training task for intelligent edge computing of the present invention:

(1)在第一次模型训练请求到达时,需要训练的数据分布在三个可用的边缘设备上;由于没有之前的已训练效果作为参考,因此将这三个可用边缘设备都认为是分布式机器学习训练的参与者,三个参与者向边缘计算节点汇报其用户数据规模;(1) When the first model training request arrives, the data to be trained is distributed on the three available edge devices; since there is no previous training effect as a reference, the three available edge devices are considered distributed. Participants in machine learning training, three participants report their user data scale to edge computing nodes;

(2)边缘计算节点初始化全局模型(边缘计算节点维护)、各边缘设备的精度修正梯度以及全局精度修正梯度;(2) The edge computing node initializes the global model (edge computing node maintenance), the accuracy correction gradient of each edge device, and the global accuracy correction gradient;

(3)边缘计算节点将全局模型参数、各边缘设备的精度修正梯度和全局精度修正梯度下发至三个边缘设备;(3) The edge computing node sends the global model parameters, the accuracy correction gradient of each edge device, and the global accuracy correction gradient to the three edge devices;

(4)各边缘设备接收到来自边缘计算设备的信息后,利用自身设备上的用户数据构造精度损失函数,并以最小化

Figure 91789DEST_PATH_IMAGE063
的形式得到
Figure 586355DEST_PATH_IMAGE064
,获得的过程为不断迭代更新
Figure 587809DEST_PATH_IMAGE064
;(4) After each edge device receives the information from the edge computing device, it uses the user data on its own device to construct an accuracy loss function, and minimizes the loss function.
Figure 91789DEST_PATH_IMAGE063
obtained in the form of
Figure 586355DEST_PATH_IMAGE064
, the obtained process is continuous iterative update
Figure 587809DEST_PATH_IMAGE064
;

(5)各边缘设备利用

Figure 203686DEST_PATH_IMAGE064
更新自己的局部模型,并利用自身的精度损失函数进行一次验证,得到局部精度、局部汇聚性能和局部精度修正梯度;(5) Utilization of edge devices
Figure 203686DEST_PATH_IMAGE064
Update your own local model, and use your own accuracy loss function to perform a verification to obtain the local accuracy, local convergence performance and local accuracy correction gradient;

(6)各边缘设备将

Figure 606986DEST_PATH_IMAGE064
、局部精度、局部汇聚性能和局部精度修正梯度发送给边缘计算节点;(6) Each edge device will
Figure 606986DEST_PATH_IMAGE064
, local accuracy, local convergence performance and local accuracy correction gradient are sent to edge computing nodes;

(7)边缘计算节点根据各边缘设备发送的

Figure 69191DEST_PATH_IMAGE064
的基础上进行全局模型更新;利用各边缘设备发送的局部精度修正梯度进行全局精度修正梯度更新;并记录局部汇聚性能;(7) The edge computing node sends the data sent by each edge device according to the
Figure 69191DEST_PATH_IMAGE064
The global model is updated on the basis of ; the global accuracy correction gradient is updated by using the local accuracy correction gradient sent by each edge device; and the local convergence performance is recorded;

(8)由于当前参与的为所有边缘设备,所以全局精度仅为各边缘设备局部精度的加权平均;(8) Since all edge devices are currently participating, the global accuracy is only the weighted average of the local accuracy of each edge device;

(9)不断进行步骤(3)到步骤(8),直到训练小轮数达到由

Figure 744892DEST_PATH_IMAGE048
所确定的
Figure 216325DEST_PATH_IMAGE003
;(9) Continue to perform steps (3) to (8) until the number of training rounds reaches from
Figure 744892DEST_PATH_IMAGE048
determined
Figure 216325DEST_PATH_IMAGE003
;

(10)观察到三个边缘设备的局部训练效果,即每次小轮的局部汇聚性能,并根据此进行三个边缘设备的偏好修正;(10) Observe the local training effect of the three edge devices, that is, the local convergence performance of each small round, and carry out the preference correction of the three edge devices according to this;

(11)第二次分布式机器学习模型训练请求到达,当前可用的是四个边缘设备;(11) The second distributed machine learning model training request arrives, and four edge devices are currently available;

(12)由于第一个边缘设备在上一次训练中的局部汇聚性能不好,因此边缘计算节点结合各边缘设备的选择偏好,选择除第一个边缘设备外的其他边缘设备作为参与者;(12) Since the local convergence performance of the first edge device in the last training is not good, the edge computing node selects other edge devices except the first edge device as participants in combination with the selection preferences of each edge device;

(13)为第二次分布式机器学习训练进行步骤(2)到步骤(10);(13) Perform steps (2) to (10) for the second distributed machine learning training;

(14)在第二次分布式机器学习训练的步骤(8)中,虽然第一个边缘设备没有进行分布式机器学习训练的参与,但是在验证的时候,仍然需要从边缘计算节点获取最新的模型参数,并利用自身的精度损失函数进行一次验证,得到局部精度,并发送给边缘计算节点。(14) In step (8) of the second distributed machine learning training, although the first edge device does not participate in the distributed machine learning training, it is still necessary to obtain the latest data from the edge computing node during verification. model parameters, and use its own accuracy loss function to perform a verification to obtain the local accuracy and send it to the edge computing node.

实验的效果如图2至图4所示,图2展示了在应用动态任务调整方法后,在不断进行分布式机器学习训练过程中的边缘计算资源消耗变化(已按最大值进行归一化),边缘训练资源消耗为边缘计算节点和各个边缘设备上计算资源花费和每一小轮传输花费的总和,训练资源消耗对比其他方法总是最少,至少减少27%的开销;图3展示了在应用动态任务调整方法后,在不断进行分布式机器学习训练过程中的全局精度变化,实际上对应的是建模中的全局后验精度

Figure 162546DEST_PATH_IMAGE065
,为在所有设备上验证得到的精度,所提方法最多降低4%的训练精度;图4展示了在应用动态任务调整方法后,在不断进行分布式机器学习训练过程中最大局部汇聚性能的变化,即建模中的
Figure 326811DEST_PATH_IMAGE048
=
Figure 240541DEST_PATH_IMAGE049
=
Figure 499352DEST_PATH_IMAGE050
。The effects of the experiments are shown in Figures 2 to 4. Figure 2 shows the changes in edge computing resource consumption during continuous distributed machine learning training after applying the dynamic task adjustment method (normalized by the maximum value) , the edge training resource consumption is the sum of the computing resource consumption on the edge computing node and each edge device and the cost of each small round of transmission. Compared with other methods, the training resource consumption is always the least, reducing the overhead by at least 27%; Figure 3 shows the application After the dynamic task adjustment method, the global accuracy change in the process of continuous distributed machine learning training actually corresponds to the global posterior accuracy in modeling
Figure 162546DEST_PATH_IMAGE065
, in order to verify the obtained accuracy on all devices, the proposed method reduces the training accuracy by up to 4%; Figure 4 shows the change in the maximum local convergence performance during continuous distributed machine learning training after applying the dynamic task adjustment method , that is, in the modeling
Figure 326811DEST_PATH_IMAGE048
=
Figure 240541DEST_PATH_IMAGE049
=
Figure 499352DEST_PATH_IMAGE050
.

本领域内的技术人员应明白,本发明的实施例可提供为方法、系统、或计算机程序产品。因此,本发明可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.

本发明是参照根据本发明实施例的方法、设备(系统)、计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.

这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.

这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.

最后应当说明的是:以上实施例仅用以说明本发明的技术方案而非对其限制,尽管参照上述实施例对本发明进行了详细的说明,本发明中的控制节点与边缘计算节点的交互方式,收集反馈信息内容与在线调度方法在各系统中均适用,所属领域的普通技术人员应当理解:依然可以对本发明的具体实施方式进行修改或者等同替换,而未脱离本发明精神和范围的任何修改或者等同替换,其均应涵盖在本发明的权利要求保护范围之内。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention and not to limit them. Although the present invention has been described in detail with reference to the above embodiments, the interaction mode between the control node and the edge computing node in the present invention , the content of the collected feedback information and the online scheduling method are applicable in each system. Those of ordinary skill in the art should understand that the specific embodiments of the present invention can still be modified or equivalently replaced without departing from the spirit and scope of the present invention. Any modification Or equivalent replacements, all of which should be covered within the protection scope of the claims of the present invention.

Claims (10)

1. A collaborative model training task configuration method facing intelligent edge computing is used for edge computing nodes and comprises one or more training time slots, and is characterized in that each training time slot comprises the following steps:
sending a model training request to one or more edge devices;
receiving an availability status and a user data size of a current time slot reported by the one or more edge devices in response to the model training request;
selecting edge equipment participating in training from the current available edge equipment based on a task configuration result obtained in the last training time slot, and determining the number of training small wheels required by interactive model training;
performing interactive model training with edge equipment participating in training until the number of the determined training small wheels is reached; and according to the training effect and the user data scale of the current time slot reported by each edge device, constructing and solving an optimization problem aiming at minimizing the use of edge training resources to obtain a new task configuration result for the next training time slot.
2. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 1, wherein the task configuration result includes: for deciding whether to select firstiAn edge device in a training time slottParticipant decision making for internal participation training
Figure 555126DEST_PATH_IMAGE001
And for decision training time slotstDecision-making aid for number of inner training small wheel
Figure 409950DEST_PATH_IMAGE002
3. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 2, wherein the training time slot is a training time slottNumber of small wheels for training
Figure 579900DEST_PATH_IMAGE003
Calculated by the following formula:
Figure 663525DEST_PATH_IMAGE003
=K
Figure 150001DEST_PATH_IMAGE002
wherein,Kis a constant.
4. The intelligent edge computing-oriented collaborative model training task configuration method of claim 1, wherein in interactive model training with an edge device participating in training, a training time slot is usedtEach training small wheel specifically comprises:
(1) the edge computing node compares the parameters of the previously trained global training model
Figure 362677DEST_PATH_IMAGE004
Local accuracy correction gradient of each edge device
Figure 70124DEST_PATH_IMAGE005
And global precision correction gradient
Figure 659237DEST_PATH_IMAGE006
Sending to all available edge devices; the edge device participating in training according to the received data and the local precision loss function of the edge device
Figure 220DEST_PATH_IMAGE007
Separately computing respective updates to global training model parameters
Figure 728005DEST_PATH_IMAGE008
tFor the ordinal number of the current interactive training,jfor the ordinal number of the current training small round,iordinal number for each edge device;
Figure 905170DEST_PATH_IMAGE009
=0;
(2) the edge computing node receives the update of the global training model parameters sent by all the edge devices participating in the training
Figure 376603DEST_PATH_IMAGE008
On the basis of the global model parameters, new global model parameters are obtained by calculation
Figure 821360DEST_PATH_IMAGE010
And sending the data to all the edge devices participating in training for verification; all edge devices participating in training are based on
Figure 985625DEST_PATH_IMAGE010
Respectively calculating to obtain new local precision
Figure 164933DEST_PATH_IMAGE011
New local accuracy correction gradient
Figure 862893DEST_PATH_IMAGE012
New local convergence performance
Figure 240785DEST_PATH_IMAGE013
And sending the data to the edge computing node for updating the record;
(3) the edge computing node corrects the gradient based on the received local precision of each edge device
Figure 497323DEST_PATH_IMAGE012
Calculating to obtain a new global precision correction gradient
Figure 163927DEST_PATH_IMAGE014
(4) If the current training small wheel reaches the current training time slottNumber of training wheels required
Figure 977163DEST_PATH_IMAGE003
The edge computing node also updates the global model parameters
Figure 163556DEST_PATH_IMAGE015
Sending the data to the edge device which does not participate in training; edge device not participating in training based on
Figure 607306DEST_PATH_IMAGE015
Calculate to obtain the respective second
Figure 541633DEST_PATH_IMAGE003
New local accuracy after each training small wheel
Figure 96242DEST_PATH_IMAGE016
And sending the data to the edge computing node for updating the record.
5. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 4, wherein in the step (1), the edge devices participating in training respectively calculate respective updates to global training model parameters according to the received data
Figure 183147DEST_PATH_IMAGE008
The method specifically comprises the following steps:
each edge device involved in training utilizes the obtained
Figure 542672DEST_PATH_IMAGE004
Figure 715028DEST_PATH_IMAGE005
And local loss of precision function of itself
Figure 57016DEST_PATH_IMAGE017
Constructing an optimization function
Figure 264007DEST_PATH_IMAGE018
And to minimize said optimization function
Figure 49560DEST_PATH_IMAGE018
In such a manner as to obtain
Figure 459944DEST_PATH_IMAGE008
The optimization function
Figure 356356DEST_PATH_IMAGE018
Expressed as:
Figure 152273DEST_PATH_IMAGE019
wherein
Figure 357996DEST_PATH_IMAGE020
Figure 504943DEST_PATH_IMAGE021
Are all determined parameters.
6. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 4, wherein in the step (2), new global model parameters
Figure 955778DEST_PATH_IMAGE010
Calculated by the following formula:
Figure 871782DEST_PATH_IMAGE022
wherein,
Figure 733559DEST_PATH_IMAGE023
time slot for current trainingtThe set of inner available edge devices,
Figure 617070DEST_PATH_IMAGE024
for training time slotstFor indicating the firstiA variable of whether an individual edge device is involved in training,
Figure 855284DEST_PATH_IMAGE024
equal to 0 or 1.
7. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 5, wherein in the step (2), the new local precision
Figure 625794DEST_PATH_IMAGE011
Is formed by edge devices
Figure 674784DEST_PATH_IMAGE010
Substituting into its own local loss of precision function
Figure 796324DEST_PATH_IMAGE017
And then obtaining; new local precision correction gradient
Figure 87497DEST_PATH_IMAGE012
Based on new local precisions
Figure 712513DEST_PATH_IMAGE011
And then obtaining; new local convergence performance
Figure 181672DEST_PATH_IMAGE025
Is obtained by the following formula:
Figure 473064DEST_PATH_IMAGE026
8. the intelligent edge computing-oriented collaborative model training task configuration method according to claim 4, wherein in the step (3), a new global accuracy correction gradient is adopted
Figure 380977DEST_PATH_IMAGE027
Obtained by the following formula:
Figure 313029DEST_PATH_IMAGE028
wherein,
Figure 15406DEST_PATH_IMAGE023
time slot for current trainingtA set of inner available edge devices.
9. The intelligent edge computing-oriented collaborative model training task configuration method according to claim 4, wherein the training effect includes: training time slottReach a certain number of training wheels internally
Figure 783642DEST_PATH_IMAGE003
Latter global model parameters
Figure 980399DEST_PATH_IMAGE015
Local convergence performance actually observed by each edge device
Figure 517691DEST_PATH_IMAGE029
And local accuracy of each edge device updated in each training small round
Figure 390969DEST_PATH_IMAGE011
(ii) a Wherein,
Figure 99031DEST_PATH_IMAGE030
10. the intelligent edge computing-oriented collaborative model training task configuration method according to claim 4, wherein the optimization problem is expressed as:
an objective function:
Figure 771583DEST_PATH_IMAGE031
constraint conditions are as follows:
Figure 491277DEST_PATH_IMAGE032
Figure 722407DEST_PATH_IMAGE033
Figure 199656DEST_PATH_IMAGE034
Figure 738216DEST_PATH_IMAGE035
Figure 250100DEST_PATH_IMAGE036
wherein,
Figure 465181DEST_PATH_IMAGE037
for training time slots for decision makingt+1 training small round number of assistant decision quantity;
Figure 678993DEST_PATH_IMAGE038
for training time slotstA set of inner available edge devices, determined by the current availability status of the report;
Figure 270512DEST_PATH_IMAGE039
for deciding whether to selectiAn edge device in a training time slott+1 participant decision volume for internal participation in training;
Figure 381775DEST_PATH_IMAGE040
the scale of one transmission of the model parameters and the gradient in the current edge network is determined;
Figure 767757DEST_PATH_IMAGE041
for training time slotstAvailable bandwidth in the inner edge network;man upper limit of capacity that the mobile network can concurrently transmit;
Figure 485177DEST_PATH_IMAGE042
for training time slotstInner firstiThe computational cost of each edge device for a single data sample;
Figure 864075DEST_PATH_IMAGE043
for training time slotstInner firstiThe user data size of each edge device;
Figure 84972DEST_PATH_IMAGE044
is a global loss of precision function, an
Figure 907434DEST_PATH_IMAGE044
=
Figure 597304DEST_PATH_IMAGE045
Figure 530625DEST_PATH_IMAGE046
Figure 855296DEST_PATH_IMAGE047
Is a set global precision loss;
Figure 848660DEST_PATH_IMAGE048
time slot for current trainingtMaximum value of local convergence performance of all edge devices after internal interaction training, and
Figure 275093DEST_PATH_IMAGE048
=
Figure 497258DEST_PATH_IMAGE049
=
Figure 692747DEST_PATH_IMAGE050
Figure 591433DEST_PATH_IMAGE029
for training time slotstReach a certain number of training wheels internally
Figure 551168DEST_PATH_IMAGE003
After thatiLocal convergence performance actually observed by each edge device.
CN202110153068.3A2021-02-042021-02-04 A collaborative model training task configuration method for intelligent edge computingActiveCN112506673B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202110153068.3ACN112506673B (en)2021-02-042021-02-04 A collaborative model training task configuration method for intelligent edge computing

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202110153068.3ACN112506673B (en)2021-02-042021-02-04 A collaborative model training task configuration method for intelligent edge computing

Publications (2)

Publication NumberPublication Date
CN112506673Atrue CN112506673A (en)2021-03-16
CN112506673B CN112506673B (en)2021-06-08

Family

ID=74953076

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202110153068.3AActiveCN112506673B (en)2021-02-042021-02-04 A collaborative model training task configuration method for intelligent edge computing

Country Status (1)

CountryLink
CN (1)CN112506673B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2020024585A1 (en)*2018-08-032020-02-06华为技术有限公司Method and apparatus for training object detection model, and device
CN111339554A (en)*2020-02-172020-06-26电子科技大学 User data privacy protection method based on mobile edge computing
CN111367657A (en)*2020-02-212020-07-03重庆邮电大学Computing resource collaborative cooperation method based on deep reinforcement learning
CN111459505A (en)*2020-05-222020-07-28南京大学 Method, device and system for deploying multi-version inference model in edge computing environment
CN111582016A (en)*2020-03-182020-08-25宁波送变电建设有限公司永耀科技分公司 Intelligent maintenance-free power grid monitoring method and system based on cloud-edge collaborative deep learning
CN111898763A (en)*2020-06-032020-11-06东南大学 A Robust Byzantine Fault Tolerant Distributed Gradient Descent Algorithm
CN112085321A (en)*2020-07-302020-12-15国家电网有限公司Station area state evaluation method based on edge calculation

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2020024585A1 (en)*2018-08-032020-02-06华为技术有限公司Method and apparatus for training object detection model, and device
CN111339554A (en)*2020-02-172020-06-26电子科技大学 User data privacy protection method based on mobile edge computing
CN111367657A (en)*2020-02-212020-07-03重庆邮电大学Computing resource collaborative cooperation method based on deep reinforcement learning
CN111582016A (en)*2020-03-182020-08-25宁波送变电建设有限公司永耀科技分公司 Intelligent maintenance-free power grid monitoring method and system based on cloud-edge collaborative deep learning
CN111459505A (en)*2020-05-222020-07-28南京大学 Method, device and system for deploying multi-version inference model in edge computing environment
CN111898763A (en)*2020-06-032020-11-06东南大学 A Robust Byzantine Fault Tolerant Distributed Gradient Descent Algorithm
CN112085321A (en)*2020-07-302020-12-15国家电网有限公司Station area state evaluation method based on edge calculation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHIQIANG WANG 等: "Adaptive Federated Learning in Resource constrained edge computing systems", 《IEEE》*

Also Published As

Publication numberPublication date
CN112506673B (en)2021-06-08

Similar Documents

PublicationPublication DateTitle
CN110610242B (en)Method and device for setting weights of participants in federal learning
CN112367109B (en)Incentive method for digital twin-driven federal learning in air-ground network
CN110782042B (en)Method, device, equipment and medium for combining horizontal federation and vertical federation
CN110443375B (en)Method and device for federated learning
WO2023124296A1 (en)Knowledge distillation-based joint learning training method and apparatus, device and medium
Wang et al.Reputation-enabled federated learning model aggregation in mobile platforms
KR20180081101A (en) Method and apparatus for optimizing user credit score
CN115270001B (en)Privacy protection recommendation method and system based on cloud collaborative learning
US20230145177A1 (en)Federated learning method and federated learning system based on mediation process
CN108289115B (en)Information processing method and system
CN119025291A (en) A resource collaborative scheduling method based on graph neural network in computing power network
CN106528614A (en)Method for predicting geographical location of user in mobile social network
CN115481752A (en)Model training method and device, electronic equipment and storage medium
CN119537024A (en) Edge computing task allocation method, device, system, storage medium and product
CN115016889A (en) A virtual machine optimization scheduling method for cloud computing
CN114760308A (en) Edge computing offloading method and device
Panigrahi et al.A reputation-aware hierarchical aggregation framework for federated learning
CN118070873B (en)Edge digital twin body deployment method based on transfer learning
CN113298115A (en)User grouping method, device, equipment and storage medium based on clustering
CN112506673B (en) A collaborative model training task configuration method for intelligent edge computing
CN116720592B (en) Federated learning model training method, device, non-volatile storage medium and electronic device
CN113743012B (en)Cloud-edge collaborative mode task unloading optimization method under multi-user scene
Liu et al.Mitigating bias in heterogeneous federated learning via stratified client selection
CN114912627A (en) Recommended model training method, system, computer equipment and storage medium
US20120151491A1 (en)Redistributing incomplete segments for processing tasks in distributed computing

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp