Movatterモバイル変換


[0]ホーム

URL:


CN112689303A - Edge cloud cooperative resource joint allocation method, system and application - Google Patents

Edge cloud cooperative resource joint allocation method, system and application
Download PDF

Info

Publication number
CN112689303A
CN112689303ACN202011584281.1ACN202011584281ACN112689303ACN 112689303 ACN112689303 ACN 112689303ACN 202011584281 ACN202011584281 ACN 202011584281ACN 112689303 ACN112689303 ACN 112689303A
Authority
CN
China
Prior art keywords
edge
computing
cloud
task
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011584281.1A
Other languages
Chinese (zh)
Other versions
CN112689303B (en
Inventor
唐怀玉
周雨晨
陈健
郭兰图
杨龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
CETC 22 Research Institute
Original Assignee
Xidian University
CETC 22 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University, CETC 22 Research InstitutefiledCriticalXidian University
Priority to CN202011584281.1ApriorityCriticalpatent/CN112689303B/en
Publication of CN112689303ApublicationCriticalpatent/CN112689303A/en
Application grantedgrantedCritical
Publication of CN112689303BpublicationCriticalpatent/CN112689303B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Landscapes

Abstract

The invention belongs to the technical field of wireless communication, and discloses a method, a system and an application for edge cloud cooperative resource joint allocation, wherein a determined task unloading part is forwarded to selected assisting user equipment; the method comprises the steps that a cooperative communication mode is adopted, the assisting user equipment forwards a task unloading part to an edge server, and at the moment, in order to ensure the decoding success rate of edge service, the unloading user equipment simultaneously sends the task unloading part to the edge server in a frequency division multiple access mode; the edge server further splits the offloaded tasks to the cloud servers. The invention adopts the edge cloud cooperation technology and realizes the purpose of providing calculation unloading service for the user in a mode of user equipment assistance. Under the background of edge cloud cooperation, proper assisting equipment is selected for users needing computing and unloading, unloading task quantity and communication resources are optimized, time delay and energy consumption index requirements of different services are guaranteed, and the aim of efficiently unloading computing tasks of the users to edge clouds and even clouds is fulfilled.

Description

Translated fromChinese
一种边云协同资源联合分配方法、系统及应用A method, system and application for joint allocation of edge-cloud collaborative resources

技术领域technical field

本发明属于无线通信技术领域,尤其涉及一种边云协同资源联合分配方法、系统及应用。The present invention belongs to the technical field of wireless communication, and in particular relates to a method, system and application for joint allocation of edge-cloud cooperative resources.

背景技术Background technique

目前:随着大数据及物联网技术的不断发展,越来越多的智能应用得以普及于无线通信领域,用户可经由无线网络获得丰富的智能业务服务。在此背景下,云计算技术的提出为智能应用与服务提供了技术支撑,无线网络内流经的大数据可经由云计算服务器分析并优化系统决策,从而为用户带来更高服务质量以及体验质量的业务。然而,随着通信技术和云计算技术的不断发展,人们不再满足于现有技术所带来的服务质量,为了进一步提升用户获取业务时的端到端时延,又或者进一步降低用户获取业务产生的能量消耗,边缘计算技术应运而生。边缘计算技术可将云计算拉近用户,为临近用户的接入点部署小型服务器,真正为用户提供“近距离”、“高速率”、“高可靠”的计算服务。At present: With the continuous development of big data and Internet of Things technologies, more and more intelligent applications are popularized in the field of wireless communication, and users can obtain rich intelligent business services through wireless networks. In this context, the proposal of cloud computing technology provides technical support for intelligent applications and services. The big data flowing in the wireless network can be analyzed and optimized by the cloud computing server, thereby bringing higher service quality and experience to users. quality business. However, with the continuous development of communication technology and cloud computing technology, people are no longer satisfied with the service quality brought by the existing technology. The resulting energy consumption, edge computing technology came into being. Edge computing technology can bring cloud computing closer to users, deploy small servers for access points close to users, and truly provide users with "close-range," "high-speed," and "high-reliability" computing services.

考虑到边缘计算相较于云计算而言可供驱使的资源量有限,在用户需求某些大型计算业务的时候,边缘计算服务器不足以满足其计算需求。在这一背景下,为了既保证用户能够“近距离”获得计算业务,又保证用户能够获得充足的计算资源,边云协同的概念应运而生。2019年Jinke Ren等人在“IEEE Transactions on Vehicular Technology”(《国际电气电子工程师协会车辆技术汇刊》)(2019年5月第68卷第5期)发表的“CollaborativeCloud and Edge Computing for Latency Minimization”(《最小化边云协同时延》)中研究了云计算和边缘计算之间的协作,令移动设备的部分任务可以在边缘节点和云服务器上联合处理,从而在有限的通信和计算能力下提高边缘云工作效率。2019年Junhui Zhao等人在“IEEE Transactions on Vehicular Technology”《国际电气电子工程师协会车辆技术汇刊》)(2019年8月第68卷第8期)发表的“Computation Offloading and ResourceAllocation For Cloud Assisted Mobile Edge Computing in Vehicular Networks”(《车联网下基于边云协同的计算卸载和资源分配》)中提出了一种专门针对车辆网的边云协同方案,通过联合优化计算卸载决策和计算资源分配,保证车辆计算卸载时的时延约束。Considering the limited amount of resources that can be driven by edge computing compared to cloud computing, when users demand certain large-scale computing services, edge computing servers are not enough to meet their computing needs. In this context, in order to ensure that users can obtain computing services at a "close distance" and that users can obtain sufficient computing resources, the concept of edge-cloud collaboration came into being. In 2019, "Collaborative Cloud and Edge Computing for Latency Minimization" published by Jinke Ren et al. in "IEEE Transactions on Vehicular Technology" ("IEEE Transactions on Vehicular Technology") (Volume 68,Issue 5, May 2019) ("Minimizing Edge-Cloud Collaboration Delay"), the cooperation between cloud computing and edge computing is studied, so that some tasks of mobile devices can be jointly processed on edge nodes and cloud servers, so that under limited communication and computing capabilities Improve edge cloud productivity. In 2019, "Computation Offloading and ResourceAllocation For Cloud Assisted Mobile Edge" published by Junhui Zhao et al. in "IEEE Transactions on Vehicular Technology" (Institute of Electrical and Electronics Engineers Transactions on Vehicle Technology) (Volume 68, Issue 8, August 2019) In "Computing in Vehicular Networks" ("Computing Offloading and Resource Allocation Based on Edge-Cloud Collaboration in the Internet of Vehicles"), an edge-cloud collaboration solution specifically for vehicle networks is proposed. Calculate the delay constraint when unloading.

然而,上述方案并未考虑到,实际网络环境中用户业务类型存在差异化,不同业务类型对传输和计算的时延以及能耗要求不同。此外,用户设备卸载计算任务时很容易被临近的其他用户设备感知,充分利用各用户设备实现计算任务的协作转发,可进一步提高边缘计算服务器处的任务可达性。However, the above solution does not take into account that there are differences in user service types in an actual network environment, and different service types have different requirements for transmission and calculation delay and energy consumption. In addition, when a user equipment offloads a computing task, it is easily perceived by other neighboring user equipments, making full use of each user equipment to realize the cooperative forwarding of computing tasks, which can further improve the task accessibility at the edge computing server.

通过上述分析,现有技术存在的问题及缺陷为:现有技术未考虑到,实际网络环境中用户业务类型存在差异化,不同业务类型对传输和计算的时延以及能耗要求不同。此外,用户设备卸载计算任务时很容易被临近的其他用户设备感知,然而现有技术并未对这些临近设备加以利用。Through the above analysis, the existing problems and defects are as follows: the existing technology does not take into account that the user service types in the actual network environment are differentiated, and different service types have different requirements for transmission and calculation delay and energy consumption. In addition, when a user equipment offloads computing tasks, it is easy to be perceived by other nearby user equipments, but the prior art does not make use of these adjacent equipments.

解决以上问题及缺陷的难度为:权重系数可以很好地表征不同用户业务对传输和计算的时延以及能耗指标要求,但是如何基于该权重系数,通过合理地资源分配和任务分割,实现时延和能耗之间的均衡,是本发明的难点之一。此外,通过协作通信技术,对临近的用户设备加以利用,一方面可加强边缘服务器卸载效率,但也进一步带来的问题求解的难度。The difficulty of solving the above problems and defects is as follows: the weight coefficient can well represent the delay and energy consumption requirements of different user services for transmission and calculation, but how to realize the time and The balance between delay and energy consumption is one of the difficulties of the present invention. In addition, using the adjacent user equipment through the cooperative communication technology can enhance the unloading efficiency of the edge server on the one hand, but it also brings the difficulty of solving the problem further.

解决以上问题及缺陷的意义为:可以基于用户任务对传输和计算的时延以及能耗指标要求,制定合理的优化方案确定用户本地计算量、上传任务量、本地传输功率、本地计算时延、协作传输功率、边缘计算任务量、边缘计算时延及云计算任务量、云计算时延等关键性能参数,从而保证时延和能耗的性能均衡。The significance of solving the above problems and defects is that a reasonable optimization scheme can be formulated to determine the user's local computing volume, upload task volume, local transmission power, local computing delay, Key performance parameters such as collaborative transmission power, edge computing task volume, edge computing delay, cloud computing task volume, cloud computing delay, etc., to ensure the performance balance between delay and energy consumption.

发明内容SUMMARY OF THE INVENTION

针对现有技术存在的问题,本发明提供了一种边云协同资源联合分配方法、系统及应用。Aiming at the problems existing in the prior art, the present invention provides a method, system and application for joint allocation of edge-cloud collaborative resources.

本发明是这样实现的,一种边云协同资源联合分配方法,所述边云协同资源联合分配方法卸载用户首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。The present invention is implemented in the following way, a method for joint allocation of edge-cloud collaborative resources. The method for joint allocation of edge-cloud collaborative resources firstly determines the delay and energy consumption weight of the offloading user based on its own service type; the offloading user selects the assistance according to the link state. equipment, and determine the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight, and note that in order to avoid interference between users, users within the coverage of adjacent edge servers use frequency division multiple access; The user first transmits the upload task to the selected assisting device, and then uses cooperative communication to offload the uploading task to the edge server with the assistance of the assisting device, so as to ensure that the edge server receives the offloading task accurately. At this time, the collaborative communication process needs to be determined. The edge server determines the amount of edge computing tasks, edge computing delay, cloud computing task volume, and cloud computing delay according to the user's weight index for computing delay and computing energy consumption, and further offloads cloud computing tasks. to the cloud server to complete the entire uninstallation process.

进一步,所述边云协同资源联合分配方法具体包括:Further, the method for joint allocation of edge-cloud collaborative resources specifically includes:

(1)卸载用户基于自身的业务类型确定时延和能耗权重,令

Figure BDA0002865158830000031
Figure BDA0002865158830000032
分别表示用户u对于通信能耗、通信时延、计算能耗和计算时延的权重指标,其中
Figure BDA0002865158830000033
Figure BDA0002865158830000034
权重值越大,表明该用户的卸载任务对于该项指标越敏感;(1) Offloading users determine the delay and energy consumption weights based on their own service types, so that
Figure BDA0002865158830000031
and
Figure BDA0002865158830000032
respectively represent the weight indicators of user u for communication energy consumption, communication delay, computing energy consumption and computing delay, where
Figure BDA0002865158830000033
and
Figure BDA0002865158830000034
The larger the weight value, the more sensitive the user's uninstallation task is to this indicator;

(2)卸载用户根据链路状态选择协助设备,根据权重确定本地计算量、上传任务量、本地传输功率及本地计算时延:(2) The offloading user selects the assisting device according to the link status, and determines the local calculation amount, upload task amount, local transmission power and local calculation delay according to the weight:

(2.1)假设用户u的备选协助设备集合为Uu,令备选设备u'作为其协助设备,则

Figure BDA0002865158830000035
其中gu,u'
Figure BDA0002865158830000036
Figure BDA0002865158830000037
分别表示用户u到备选设备u'的功率增益、用户u到用户u关联的边缘服务器nu的功率增益和备选设备u'到用户u关联的边缘服务器nu的功率增益;(2.1) Assuming that the set of alternative assisting devices for user u is Uu , let the alternative device u' be its assisting device, then
Figure BDA0002865158830000035
where gu,u' ,
Figure BDA0002865158830000036
and
Figure BDA0002865158830000037
respectively represent the power gain from user u to the candidate device u', the power gain from user u to the edge server nu associated with useru , and the power gain from the candidate deviceu ' to the edge server nu associated with user u;

(2.2)用户u可根据(1)中设定的权重优化本地计算量au、上传任务量1-au、本地传输功率pu,1及本地计算时延

Figure BDA0002865158830000038
优化问题可建模为:(2.2) The user u can optimize the local computation amount au , the upload task amount 1-au , the local transmission power pu,1 and the local computation delay according to the weight set in (1)
Figure BDA0002865158830000038
The optimization problem can be modeled as:

Figure BDA0002865158830000039
Figure BDA0002865158830000039

其中,Iu、Bu、c、

Figure BDA0002865158830000041
ξUser分别表示用户u的总任务量、卸载频带宽度(、计算每比特任务所需消耗的CPU数、本地计算时延、用户设备计算能量系数;此外,优化变量pu,1
Figure BDA0002865158830000042
有上下界限制,若超过了边界值,则直接设定为边界值;Among them, Iu , Bu , c,
Figure BDA0002865158830000041
ξUser represents the total task amount of user u, the offload bandwidth (, the number of CPUs required to calculate each bit task, the local computing delay, and the user equipment computing energy coefficient; in addition, the optimization variables pu,1 and
Figure BDA0002865158830000042
There are upper and lower bounds, if the boundary value is exceeded, it is directly set as the boundary value;

(2.3)卸载用户u采用交替迭代的方式确定变量au、pu,1

Figure BDA0002865158830000043
(2.3) Unloading user u determines the variables au , pu,1 and
Figure BDA0002865158830000043

(3)卸载用户u首先将卸载任务的1-au部分以传输功率pu,1发送至选择的协助设备u',然后在协助设备的辅助下采用协作通信的方式将卸载任务的1-au部分卸载至边缘服务器nu,此时卸载用户u和协助设备u'的协作传输功率分别为pu,2及pu',pu,2及pu'的解由以下优化问题获得:(3) The offloading user u first sends the 1-au part of the offloading task to the selected assisting device u' with the transmission power pu,1 , and then uses cooperative communication with the assistance of the assisting device to send the 1-a u part of the offloading task to the selected assisting device u'. The part of au is offloaded to the edge server nu . At this time, the cooperative transmission power of the offloaded user u and the assisting device u' is pu,2 and pu ' respectively. The solutions of pu,2 and pu' are obtained by the following optimization problem :

Figure BDA0002865158830000044
Figure BDA0002865158830000044

此处优化变量pu,2和pu'有上下界限制,若超过了边界值,则直接设定为边界值;此时优化问题OP3的目标函数前两项分别为关于pu,2和pu'的非凸函数,后两项分别为关于pu,2和pu'的凹函数,因此需要对前两项进行连续凸逼近,此时变量pu,2和pu'的解可由以下优化问题获得:Here, the optimization variables pu,2 and pu' have upper and lower bounds. If they exceed the boundary value, they are directly set as the boundary value; at this time, the first two items of the objective function of the optimization problem OP3 are about pu,2 and The non-convex function of pu' , the last two terms are concave functions about pu,2 and pu' respectively, so it is necessary to carry out continuous convex approximation to the first two terms. At this time, the solutions of the variables pu,2 and pu' It can be obtained by the following optimization problem:

Figure BDA0002865158830000045
Figure BDA0002865158830000045

其中,角标i表示迭代次数;

Figure BDA0002865158830000046
Figure BDA0002865158830000047
表示函数h(pu,2,pu')在pu,2的一阶导数并令
Figure BDA0002865158830000048
Figure BDA0002865158830000049
表示第i次迭代得到的pu,2值;
Figure BDA00028651588300000410
表示函数h(pu,2,pu')在pu'的一阶导数并代入
Figure BDA00028651588300000411
Figure BDA00028651588300000412
表示第i次迭代得到的pu'值;上述优化问题OP4为关于pu,2和pu'的凸优化问题,通过搜索H'(pu,2)=0和H'(pu')=0的根即可得到优化问题OP3中pu,2和pu'的解;基于连续凸逼近的思想,反复更新的pu,2和pu'值直至相临两次迭代的pu,2和pu'的差值分别小于一定精度,即可认为算法收敛,收敛时得到的pu,2和pu'即为(3)中卸载用户u和协助设备u'的协作传输功率pu,2和pu'的最终解;Among them, the index i represents the number of iterations;
Figure BDA0002865158830000046
Figure BDA0002865158830000047
Denote the first derivative of the function h(pu,2 ,pu' ) at pu,2 and let
Figure BDA0002865158830000048
Figure BDA0002865158830000049
Indicates the pu,2 value obtained by the i-th iteration;
Figure BDA00028651588300000410
Represent the first derivative of the function h(pu,2 ,pu' ) at pu' and substitute it into
Figure BDA00028651588300000411
Figure BDA00028651588300000412
represents the value of pu' obtained by the ith iteration; the above optimization problem OP4 is a convex optimization problem about pu,2 and pu' , by searching H'(pu,2 )=0 and H'(pu' )=0, the solutions of pu, 2 and pu' in optimization problem OP3 can be obtained; based on the idea of continuous convex approximation, the values of p u, 2 and p u' are updated repeatedly until the pu, 2 and pu' values of two adjacent iterations are reached. If the difference betweenu,2 and pu' is less than a certain precision, it can be considered that the algorithm converges, and the pu,2 and pu' obtained during convergence are the cooperative transmission of the unloaded user u and the assisting device u' in (3). The final solution of power pu,2 and pu' ;

(4)边缘服务器nu根据卸载用户u对于计算时延和计算能耗的权重指标,确定边缘计算任务量bu、边缘计算时延

Figure BDA0002865158830000051
及云计算任务量1-bu、云计算时延
Figure BDA0002865158830000052
并将云计算任务1-bu进一步卸载至云服务器,完成整个卸载过程:(4) The edge server nu determines the edge computing task amount bu and the edge computing delay according to the weight index of the offloading user u on the computing delay and computing energy consumption
Figure BDA0002865158830000051
and cloud computing task volume 1-bu , cloud computing delay
Figure BDA0002865158830000052
And further uninstall the cloud computing task 1-bu to the cloud server to complete the entire uninstallation process:

(4.1)边缘服务器nu根据(1)中设定的权重优化边缘计算任务量bu、边缘计算时延

Figure BDA0002865158830000053
及云计算任务量1-bu、云计算时延
Figure BDA0002865158830000054
优化问题可建模为:(4.1) The edge server nu optimizes the edge computing task volume bu and the edge computing delay according to the weight set in (1).
Figure BDA0002865158830000053
and cloud computing task volume 1-bu , cloud computing delay
Figure BDA0002865158830000054
The optimization problem can be modeled as:

Figure BDA0002865158830000055
Figure BDA0002865158830000055

其中ξMEC和ξCloud表示边缘服务器计算能量系数和云服务器计算能量系数;此处优化变量

Figure BDA0002865158830000056
Figure BDA0002865158830000057
有上下界限制,若超过了边界值,则直接设定为边界值;where ξMEC and ξCloud represent the computing energy coefficient of edge server and the computing energy coefficient of cloud server; the variables are optimized here
Figure BDA0002865158830000056
and
Figure BDA0002865158830000057
There are upper and lower bounds, if the boundary value is exceeded, it is directly set as the boundary value;

(4.2)边缘服务器nu采用交替迭代的方式确定变量bu

Figure BDA0002865158830000058
Figure BDA0002865158830000059
(4.2) The edge server nu determines the variables bu ,
Figure BDA0002865158830000058
and
Figure BDA0002865158830000059

(5)卸载用户u、边缘服务器nu及云服务器基于优化结果分别以计算速率

Figure BDA00028651588300000510
Figure BDA00028651588300000511
处理卸载用户u的任务中au部分、bu(1-au)部分及(1-bu)(1-au)部分,计算完毕后卸载用户u汇总各部分计算结果,并生成卸载任务的最终计算结果。(5) The offloading user u, the edge server nu and the cloud server respectively calculate the speed based on the optimization result.
Figure BDA00028651588300000510
and
Figure BDA00028651588300000511
Process the au part, bu (1-au ) part and (1-bu )(1-au ) part of the task of uninstalling user u, after the calculation is completed, the uninstall user u summarizes the calculation results of each part, and generates the uninstallation The final calculation result of the task.

进一步,所述步骤(2.3)按如下步骤进行:Further, the step (2.3) is carried out as follows:

(2.3a)对每个卸载用户u初始化一组pu,1

Figure BDA00028651588300000512
的值;(2.3a) Initialize a set of pu,1 and
Figure BDA00028651588300000512
the value of;

(2.3b)基于得到的pu,1

Figure BDA00028651588300000513
优化变量au,此时构造的优化问题被转换为关于au的凸优化问题,基于凸优化理论可得au的解为:(2.3b) Based on the obtained pu,1 and
Figure BDA00028651588300000513
Optimize the variable au , the optimization problem constructed at this time is transformed into a convex optimization problem about au , and the solution of au can be obtained based on the convex optimization theory as:

Figure BDA00028651588300000514
Figure BDA00028651588300000514

如果得到的au*大于1,则取值为1;If the obtained au* is greater than 1, the value is 1;

(2.3c)基于得到的au优化变量

Figure BDA00028651588300000515
此时构造的优化问题被转换为关于
Figure BDA00028651588300000516
的凸优化问题,基于凸优化理论可得
Figure BDA00028651588300000517
的解为
Figure BDA00028651588300000518
(2.3c) Optimization variables based on the obtained au
Figure BDA00028651588300000515
The optimization problem constructed at this point is transformed into
Figure BDA00028651588300000516
The convex optimization problem of , based on the convex optimization theory, we can get
Figure BDA00028651588300000517
The solution is
Figure BDA00028651588300000518

(2.3d)基于得到的au优化变量pu,1,此时目标函数的第一项为关于pu,1的非凸函数,第二项为关于pu,1的凹函数,因此需要对第一项进行连续凸逼近,此时变量pu,1的解可由以下优化问题获得:(2.3d) Optimize the variable pu,1 based on the obtained au . At this time, the first item of the objective function is a non-convex function about p u,1 , and the second item is a concave function about pu,1 . Therefore, it is necessary to A continuous convex approximation is performed for the first term, and the solution to the variable pu,1 can be obtained by the following optimization problem:

Figure BDA0002865158830000061
Figure BDA0002865158830000061

其中,角标i表示迭代次数;

Figure BDA0002865158830000062
Figure BDA0002865158830000063
表示函数f(pu,1)在pu,1的一阶导数并代入
Figure BDA0002865158830000064
Figure BDA0002865158830000065
表示第i次迭代得到的pu,1值;上述优化问题OP2为关于pu,1的凸优化问题,通过搜索F'(pu,1)=0的根即可得到优化问题OP2中pu,1的解;基于连续凸逼近的思想,反复更新的pu,1值直至相临两次迭代的pu,1差值小于一定精度,即可认为算法收敛,最新得到的pu,1值为优化问题OP1中pu,1的解;Among them, the index i represents the number of iterations;
Figure BDA0002865158830000062
Figure BDA0002865158830000063
Represent the first derivative of the function f(pu,1 ) at pu,1 and substitute it into
Figure BDA0002865158830000064
Figure BDA0002865158830000065
Represents the value of pu,1 obtained by the i-th iteration; the above optimization problem OP2 is a convex optimization problem about pu,1 , and p in the optimization problem OP2 can be obtained by searching for the root of F'(pu,1 )=0 The solution ofu,1 ; based on the idea of continuous convex approximation, the value of pu,1 is repeatedly updated until the difference of pu,1 between two adjacent iterations is less than a certain precision, the algorithm can be considered to be convergent, and the latest obtained pu, 1 is the solution of pu,1 in optimization problem OP1;

(2.3e)基于步骤(2.3c)和(2.3d)中得到的

Figure BDA0002865158830000066
Figure BDA0002865158830000067
更新au,再基于得到的au,刷新
Figure BDA0002865158830000068
Figure BDA0002865158830000069
不断循环直至算法收敛,收敛时得到的au、pu,1
Figure BDA00028651588300000610
即为步骤(2.2)中本地计算量au、本地传输功率pu,1及本地计算
Figure BDA00028651588300000611
的最终解。(2.3e) based on the results obtained in steps (2.3c) and (2.3d)
Figure BDA0002865158830000066
and
Figure BDA0002865158830000067
Update au , then based on the obtained au , refresh
Figure BDA0002865158830000068
and
Figure BDA0002865158830000069
Keep looping until the algorithm converges, and the au , pu,1 and
Figure BDA00028651588300000610
That is, in step (2.2), the local calculation amount au , the local transmission power pu,1 and the local calculation
Figure BDA00028651588300000611
the final solution.

进一步,所述步骤(4.2)按如下步骤进行:Further, the step (4.2) is carried out as follows:

(4.2a)对边缘服务器nu初始化一组

Figure BDA00028651588300000612
Figure BDA00028651588300000613
的值;(4.2a) Initialize a set of edge servers nu
Figure BDA00028651588300000612
and
Figure BDA00028651588300000613
the value of;

(4.2b)基于得到的

Figure BDA00028651588300000614
Figure BDA00028651588300000615
优化变量bu,此时构造的优化问题被转换为关于bu的凸优化问题,基于凸优化理论可得bu的解为:(4.2b) Based on the obtained
Figure BDA00028651588300000614
and
Figure BDA00028651588300000615
Optimize the variable bu , the optimization problem constructed at this time is transformed into a convex optimization problem about bu , and the solution of bu can be obtained based on the convex optimization theory as:

Figure BDA00028651588300000616
Figure BDA00028651588300000616

(4.2c)基于得到的bu优化变量

Figure BDA00028651588300000617
Figure BDA00028651588300000618
此时构造的优化问题被转换为关于
Figure BDA00028651588300000619
Figure BDA00028651588300000620
的凸优化问题,基于凸优化理论可得
Figure BDA00028651588300000621
Figure BDA00028651588300000622
的解分为:(4.2c) Optimization variables based on the obtained bu
Figure BDA00028651588300000617
and
Figure BDA00028651588300000618
The optimization problem constructed at this point is transformed into
Figure BDA00028651588300000619
and
Figure BDA00028651588300000620
The convex optimization problem of , based on the convex optimization theory, we can get
Figure BDA00028651588300000621
and
Figure BDA00028651588300000622
The solution is divided into:

Figure BDA00028651588300000623
Figure BDA00028651588300000624
Figure BDA00028651588300000623
and
Figure BDA00028651588300000624

(4.2d)基于步骤(4.2c)中得到的

Figure BDA00028651588300000625
Figure BDA00028651588300000626
更新bu,再基于得到的bu,刷新
Figure BDA00028651588300000627
Figure BDA00028651588300000628
不断循环直至算法收敛,收敛时得到的bu
Figure BDA00028651588300000629
Figure BDA00028651588300000630
即为步骤(4.1)中边缘计算任务量bu、边缘计算时延
Figure BDA00028651588300000631
及云计算时延
Figure BDA00028651588300000632
的最终解。(4.2d) based on the result obtained in step (4.2c)
Figure BDA00028651588300000625
and
Figure BDA00028651588300000626
update bu , and then based on the obtained bu , refresh
Figure BDA00028651588300000627
and
Figure BDA00028651588300000628
Continuously loop until the algorithm converges, and thebu ,
Figure BDA00028651588300000629
and
Figure BDA00028651588300000630
That is, the amount of edge computing tasks bu and the edge computing delay in step (4.1)
Figure BDA00028651588300000631
and cloud computing latency
Figure BDA00028651588300000632
the final solution.

本发明的另一目的在于提供一种计算机设备,所述计算机设备包括存储器和处理器,所述存储器存储有计算机程序,所述计算机程序被所述处理器执行时,使得所述处理器执行如下步骤:首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。Another object of the present invention is to provide a computer device, the computer device includes a memory and a processor, the memory stores a computer program, and when the computer program is executed by the processor, the processor executes the following Steps: First, determine the delay and energy consumption weights based on their own business types; the offload user selects the assisting device according to the link status, and determines the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight. To avoid interference between users, users within the coverage of adjacent edge servers use frequency division multiple access; the user first transmits the upload task to the selected assisting device, and then uses cooperative communication with the assistance of the assisting device. The task is offloaded to the edge server to ensure that the offload task is accurately received by the edge server. At this time, the cooperative transmission power in the cooperative communication process needs to be determined; the edge server determines the edge computing task according to the user's weighted indicators of computing delay and computing energy consumption The workload, edge computing delay, cloud computing task volume, cloud computing delay, and the cloud computing tasks are further offloaded to the cloud server to complete the entire offloading process.

本发明的另一目的在于提供一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行如下步骤:首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。Another object of the present invention is to provide a computer-readable storage medium storing a computer program. When the computer program is executed by a processor, the processor causes the processor to perform the following steps: first, determine the delay and the delay based on its own service type. Energy consumption weight; offloading users select assisting devices according to the link status, and determine the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight. Note that in order to avoid interference between users, the coverage area of adjacent edge servers The users in the network use the frequency division multiple access method; the user first transmits the upload task to the selected assisting device, and then uses the cooperative communication method to offload the uploading task to the edge server with the assistance of the assisting device, so as to ensure that the edge server is accurate. After receiving the offloading task, the cooperative transmission power in the cooperative communication process needs to be determined at this time; the edge server determines the amount of edge computing tasks, edge computing delay and cloud computing task amount, Cloud computing delays, and the cloud computing tasks are further offloaded to the cloud server to complete the entire offloading process.

本发明的另一目的在于提供一种信息数据处理终端,所述信息数据处理终端用于实现所述的边云协同资源联合分配方法。Another object of the present invention is to provide an information data processing terminal, the information data processing terminal is used to realize the joint allocation method for edge-cloud collaborative resources.

本发明的另一目的在于提供一种边缘计算服务器,所述边缘计算服务器用于实现所述的边云协同资源联合分配方法。Another object of the present invention is to provide an edge computing server, which is used to implement the method for joint allocation of edge-cloud collaborative resources.

本发明的另一目的在于提供一种云计算服务器,所述云计算服务器用于实现所述的边云协同资源联合分配方法。Another object of the present invention is to provide a cloud computing server, which is used for implementing the method for joint allocation of edge-cloud collaborative resources.

本发明的另一目的在于提供一种无线通信系统,所述无线通信系统用于实现所述的边云协同资源联合分配方法。Another object of the present invention is to provide a wireless communication system, which is used for implementing the method for joint allocation of edge-cloud cooperative resources.

结合上述的所有技术方案,本发明所具备的优点及积极效果为:本发明通过用户协助边云任务的卸载,能够提升卸载任务的可达性,同时基于卸载用户对时延、能耗的权重指标要求,实现通信、计算两方面的时延能效加权和的最小化。本发明采用边云协同技术,通过用户设备协助的方式,实现为用户提供计算卸载服务的目的。在边云协同的背景下,为需要计算卸载的用户选择合适的协助设备、优化卸载任务量及通信资源,保证不同业务的时延及能耗指标要求,实现用户计算任务高效卸载至边缘云甚至是云的目标。Combining all the above technical solutions, the advantages and positive effects of the present invention are: the present invention can improve the accessibility of the offloading task by assisting the offloading of the edge cloud task by the user, and at the same time, based on the weight of the offloading user on the delay and energy consumption The index requirements are to minimize the weighted sum of delay and energy efficiency in both communication and computation. The invention adopts the edge-cloud collaboration technology, and realizes the purpose of providing computing offloading services for users by means of user equipment assistance. In the context of edge-cloud collaboration, select appropriate assistance devices for users who need computing offloading, optimize offloading tasks and communication resources, ensure the latency and energy consumption requirements of different services, and realize efficient offloading of user computing tasks to edge cloud or even is the goal of the cloud.

本发明具体描述了边云协同技术在多用户计算任务卸载及无线通信场景中的应用,并通过联合分配计算资源和通信资源(包括协助设备选择、计算任务拆分比例、传输功率、计算时延)的方式,保证不同卸载业务在通信和计算两方面的时延及能耗指标要求,实现用户计算任务高效卸载至边缘云甚至是云的目标。The present invention specifically describes the application of edge-cloud collaboration technology in multi-user computing task offloading and wireless communication scenarios, and through joint allocation of computing resources and communication resources (including assisting device selection, computing task split ratio, transmission power, computing delay ) method to ensure the latency and energy consumption index requirements of different offloading services in terms of communication and computing, and achieve the goal of efficiently offloading user computing tasks to the edge cloud or even the cloud.

本发明提出了一种基于用户协助的边云协同资源联合分配方法,方法具体描述了边云协同技术在多用户计算任务卸载及无线通信场景中的应用,并通过联合分配计算资源和通信资源(包括协助设备选择、计算任务拆分比例、传输功率、计算时延)的方式,保证不同卸载业务在通信和计算两方面的时延及能耗指标要求,实现用户计算任务高效卸载至边缘云甚至是云的目标。The present invention proposes a method for joint allocation of edge-cloud collaborative resources based on user assistance. The method specifically describes the application of the edge-cloud collaborative technology in multi-user computing task offloading and wireless communication scenarios, and allocates computing resources and communication resources through joint allocation ( Including assisting device selection, computing task split ratio, transmission power, computing delay), to ensure the delay and energy consumption index requirements of different offload services in terms of communication and computing, and to achieve efficient offloading of user computing tasks to the edge cloud or even is the goal of the cloud.

本发明提出了一种基于用户协助的边云协同资源联合分配方法,方法考虑到实际网络环境中用户业务类型存在差异化,不同业务类型对传输和计算的时延以及能耗要求不同。此外,用户设备卸载计算任务时很容易被临近的其他用户设备感知,充分利用各用户设备实现计算任务的协作转发,可进一步提高边缘计算服务器处的任务可达性。The present invention proposes a method for joint allocation of edge-cloud collaborative resources based on user assistance, which takes into account differences in user service types in an actual network environment, and different service types have different requirements for transmission and calculation delay and energy consumption. In addition, when a user equipment offloads a computing task, it is easily perceived by other neighboring user equipments, making full use of each user equipment to realize the cooperative forwarding of computing tasks, which can further improve the task accessibility at the edge computing server.

本发明采用边云协同技术,通过用户设备协助的方式,实现为用户提供计算卸载服务的目的。在边云协同的背景下,为需要计算卸载的用户选择合适的协助设备、优化卸载任务量及通信资源,保证不同业务的时延及能耗指标要求,实现用户计算任务高效卸载至边缘云甚至是云的目标。The invention adopts the edge-cloud collaboration technology, and realizes the purpose of providing computing offloading services for users by means of user equipment assistance. In the context of edge-cloud collaboration, select appropriate assistance devices for users who need computing offloading, optimize offloading tasks and communication resources, ensure the latency and energy consumption requirements of different services, and realize efficient offloading of user computing tasks to edge cloud or even is the goal of the cloud.

附图说明Description of drawings

为了更清楚地说明本申请实施例的技术方案,下面将对本申请实施例中所需要使用的附图做简单的介绍,显而易见地,下面所描述的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下还可以根据这些附图获得其他的附图。In order to explain the technical solutions of the embodiments of the present application more clearly, the following will briefly introduce the drawings that need to be used in the embodiments of the present application. Obviously, the drawings described below are only some embodiments of the present application. For those of ordinary skill in the art, other drawings can also be obtained from these drawings without creative effort.

图1是本发明实施例提供的边云协同资源联合分配方法流程图。FIG. 1 is a flowchart of a method for joint allocation of edge-cloud collaborative resources provided by an embodiment of the present invention.

图2是本发明实施例提供的边云协同资源联合分配系统的结构示意图;2 is a schematic structural diagram of an edge-cloud collaborative resource joint allocation system provided by an embodiment of the present invention;

图2中:1、时延和能耗权重确定模块;2、参数确定模块;3、协作通信传输模块;4、卸载处理模块;5、计算结果输出模块。In Fig. 2: 1. Delay and energy consumption weight determination module; 2. Parameter determination module; 3. Collaborative communication transmission module; 4. Offload processing module; 5. Calculation result output module.

图3是本发明实施例提供的使用的用户设备协助的边云协同模型图。FIG. 3 is a diagram of a user equipment-assisted edge-cloud collaboration model provided by an embodiment of the present invention.

图4是本发明实施例提供的边云协同资源联合分配方法的实现总流程图。FIG. 4 is a general flow chart of the implementation of the method for joint allocation of edge-cloud collaborative resources provided by an embodiment of the present invention.

图5是本发明实施例提供的在不同能耗和时延权重下的性能示例图。FIG. 5 is an example diagram of performance under different energy consumption and delay weights provided by an embodiment of the present invention.

图6是本发明实施例提供的不同技术效果性能对比图。FIG. 6 is a performance comparison diagram of different technical effects provided by an embodiment of the present invention.

具体实施方式Detailed ways

为了使本发明的目的、技术方案及优点更加清楚明白,以下结合实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本发明,并不用于限定本发明。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention.

针对现有技术存在的问题,本发明提供了一种边云协同资源联合分配方法、系统及应用。本发明所涉及的边云协同涉及资源协同和业务管理协同两方面:1.资源协同特指计算资源协同,可在用户本地资源及边缘资源不足的情况下,调用云中心的资源进行补充,满足用户侧应用对资源的需要,云中心可提供的资源包括但不局限于裸机、虚拟机和容器。2.业务管理协同可通过云服务器增值网络业务的服务质量,按能耗和时延指标需求为客户提供相关网络业务。下面结合附图对本发明作详细的描述。Aiming at the problems existing in the prior art, the present invention provides a method, system and application for joint allocation of edge-cloud collaborative resources. The edge-cloud collaboration involved in the present invention involves two aspects: resource collaboration and business management collaboration: 1. Resource collaboration specifically refers to computing resource collaboration. In the case of insufficient user local resources and edge resources, the resources of the cloud center can be called for supplementation to meet the needs of User-side applications require resources, and the resources provided by the cloud center include but are not limited to bare metal, virtual machines, and containers. 2. Business management collaboration can provide customers with relevant network services according to the requirements of energy consumption and delay indicators through the service quality of cloud server value-added network services. The present invention will be described in detail below with reference to the accompanying drawings.

如图1所示,本发明提供的边云协同资源联合分配方法包括以下步骤:As shown in Figure 1, the method for joint allocation of edge-cloud collaborative resources provided by the present invention includes the following steps:

S101:卸载用户首先基于自身的业务类型确定时延和能耗权重;S101: The offloading user first determines the delay and energy consumption weights based on the service type of the offloading user;

S102:卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延;S102: The offloading user selects the assisting device according to the link state, and determines the local computation amount, the upload task amount, the local transmission power, and the local computation delay according to the weight;

S103:用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;S103: The user first transmits the uploading task to the selected assisting device, and then uses cooperative communication to offload the uploading task to the edge server with the assistance of the assisting device, so as to ensure that the edge server receives the offloading task accurately. At this time, it is necessary to determine the collaboration Cooperative transmission power during communication;

S104:边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程;S104: The edge server determines the edge computing task volume, edge computing delay, cloud computing task volume, and cloud computing delay according to the user's weight index for computing delay and computing energy consumption, and further offloads the cloud computing task to the cloud server. So as to complete the entire uninstallation process;

S105:卸载用户、边缘服务器及云服务器基于优化结果分别以各自的计算速率处理分割后的卸载任务,计算完毕后卸载用户汇总各部分计算结果,并生成最终计算结果。S105: The offloading user, the edge server, and the cloud server process the divided offloading tasks at their respective calculation rates based on the optimization result, and after the calculation is completed, the offloading user aggregates the calculation results of each part, and generates a final calculation result.

本发明提供的边云协同资源联合分配方法业内的普通技术人员还可以采用其他的步骤实施,图1的本发明提供的边云协同资源联合分配方法仅仅是一个具体实施例而已。The method for joint allocation of edge-cloud collaborative resources provided by the present invention may also be implemented by those skilled in the art by other steps. The method for joint allocation of collaborative resources for side-cloud provided by the present invention in FIG. 1 is only a specific embodiment.

如图2所示,本发明提供的边云协同资源联合分配系统包括:As shown in Figure 2, the system for joint allocation of edge-cloud collaborative resources provided by the present invention includes:

时延和能耗权重确定模块1,用于实现卸载用户基于自身的业务类型确定时延和能耗权重;The delay and energy consumptionweight determination module 1 is used to realize that the offloading user determines the delay and energy consumption weight based on its own service type;

参数确定模块2,用于实现卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延;Theparameter determination module 2 is used to realize that the unloading user selects the assisting device according to the link state, and determines the local calculation amount, the upload task amount, the local transmission power, and the local calculation delay according to the weight;

协作通信传输模块3,用于实现用户将上传任务传输至选择的协助设备,在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器;The cooperativecommunication transmission module 3 is used to realize that the user transmits the upload task to the selected assisting device, and adopts the method of collaborative communication to offload the uploading task to the edge server with the assistance of the assisting device;

卸载处理模块4,用于实现边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,完成整个卸载过程;Theoffload processing module 4 is used to realize that the edge server determines the amount of edge computing tasks, the delay of edge computing, the task amount of cloud computing, and the delay of cloud computing according to the user's weight index of computing delay and computing energy consumption, and assigns the cloud computing tasks. Further uninstall to the cloud server to complete the entire uninstall process;

计算结果输出模块5,用于卸载用户、边缘服务器及云服务器基于优化结果分别以各自的计算速率处理分割后的卸载任务,计算完毕后卸载用户汇总各部分计算结果,并生成最终计算结果。The calculationresult output module 5 is used for the offloading user, the edge server and the cloud server to process the divided offloading tasks at their respective calculation rates based on the optimization results.

下面结合附图对本发明的技术方案作进一步的描述。The technical solutions of the present invention will be further described below with reference to the accompanying drawings.

如图3所示,本发明使用的用户设备协助的边云协同模型图,其中,基站集成云服务器,网络边缘的接入点集成了边缘服务器。卸载用户完成任务卸载需要三大步骤:首先,将确定好的任务卸载部分转发至选中的协助用户设备;然后,采用协作通信的方式,由协助用户设备将任务卸载部分转发至边缘服务器,此时为了保证边缘服务的解码成功率,卸载用户设备以频分多址接入方式同时将任务卸载部分发送至边缘服务器;最后,边缘服务器将卸载的任务进一步拆分至云服务器,由于边缘服务器至云服务器通常采用有线的方式连接,因此这一部分的通信资源可被假设为充足的,仅考虑计算能耗和计算时延。As shown in FIG. 3 , the user equipment-assisted edge-cloud collaboration model diagram used in the present invention, wherein the base station integrates the cloud server, and the access point at the network edge integrates the edge server. Three steps are required for the offloading user to complete the task offloading: first, the determined offloading part of the task is forwarded to the selected assisting user equipment; then, the assisting user equipment forwards the offloading part of the task to the edge server by means of cooperative communication. In order to ensure the decoding success rate of the edge service, the offloaded user equipment simultaneously sends the offloaded part of the task to the edge server in the frequency division multiple access mode; finally, the edge server further splits the offloaded tasks to the cloud server, since the edge server to the cloud The servers are usually connected by wire, so the communication resources of this part can be assumed to be sufficient, and only the calculation energy consumption and calculation delay are considered.

如图4所示,本发明的具体实现步骤如下:As shown in Figure 4, the specific implementation steps of the present invention are as follows:

步骤一:卸载用户基于自身的业务类型确定时延和能耗权重。令

Figure BDA0002865158830000111
Figure BDA0002865158830000112
Figure BDA0002865158830000113
分别表示用户u对于通信能耗、通信时延、计算能耗和计算时延的权重指标,其中
Figure BDA0002865158830000114
Figure BDA0002865158830000115
权重值越大,表明该用户的卸载任务对于该项指标越敏感。Step 1: The offloading user determines the delay and energy consumption weights based on its own service type. make
Figure BDA0002865158830000111
Figure BDA0002865158830000112
and
Figure BDA0002865158830000113
respectively represent the weight indicators of user u for communication energy consumption, communication delay, computing energy consumption and computing delay, where
Figure BDA0002865158830000114
and
Figure BDA0002865158830000115
The larger the weight value, the more sensitive the user's uninstallation task is to this indicator.

步骤二:卸载用户根据链路状态选择协助设备,根据权重确定本地计算量、上传任务量、本地传输功率及本地计算时延:Step 2: The offloading user selects the assisting device according to the link status, and determines the local calculation amount, upload task amount, local transmission power and local calculation delay according to the weight:

2.1)假设用户u的备选协助设备集合为Uu,令备选设备u'作为其协助设备,则

Figure BDA0002865158830000116
其中gu,u'
Figure BDA0002865158830000117
Figure BDA0002865158830000118
分别表示用户u到备选设备u'的功率增益、用户u到用户u关联的边缘服务器nu的功率增益和备选设备u'到用户u关联的边缘服务器nu的功率增益。2.1) Assuming that the set of alternative assisting devices of user u is Uu , let the alternative device u' be its assisting device, then
Figure BDA0002865158830000116
where gu,u' ,
Figure BDA0002865158830000117
and
Figure BDA0002865158830000118
respectively represent the power gain from the user u to the candidate device u', the power gain from the user u to the edge server nu associated with the useru , and the power gain from the candidate deviceu ' to the edge server nu associated with the user u.

2.2)用户u可根据步骤1中设定的权重优化本地计算量au、上传任务量1-au、本地传输功率pu,1及本地计算时延

Figure BDA0002865158830000121
优化问题可建模为:2.2) User u can optimize the local computation amount au , the upload task amount 1-au , the local transmission power pu,1 and the local computation delay according to the weight set instep 1
Figure BDA0002865158830000121
The optimization problem can be modeled as:

Figure BDA0002865158830000122
Figure BDA0002865158830000122

其中,Iu、Bu、c、

Figure BDA0002865158830000123
ξUser分别表示用户u的总任务量(单位:bits)、卸载频带宽度(单位:Hz)、计算每比特任务所需消耗的CPU数(单位:CPU/bit)、本地计算时延(单位:s)、用户设备计算能量系数(可由CPU利用率乘以能效系数表示);此外,优化变量pu,1
Figure BDA0002865158830000124
有上下界限制,若超过了边界值,则直接设定为边界值。Among them, Iu , Bu , c,
Figure BDA0002865158830000123
ξUser represents the total task amount of user u (unit: bits), the offload bandwidth (unit: Hz), the number of CPUs required to calculate each bit task (unit: CPU/bit), and the local computing delay (unit: s), the user equipment calculates the energy coefficient (which can be expressed by multiplying the CPU utilization by the energy efficiency coefficient); in addition, the optimization variables pu,1 and
Figure BDA0002865158830000124
There are upper and lower bounds. If the boundary value is exceeded, it is directly set as the boundary value.

2.3)卸载用户u采用交替迭代的方式确定变量au、pu,1

Figure BDA0002865158830000125
2.3) Unloading user u determines variables au , pu,1 and
Figure BDA0002865158830000125

2.3.1)对每个卸载用户u初始化一组pu,1

Figure BDA0002865158830000126
的值;2.3.1) Initialize a set of pu,1 and
Figure BDA0002865158830000126
the value of;

2.3.2)基于得到的pu,1

Figure BDA0002865158830000127
优化变量au,此时构造的优化问题被转换为关于au的凸优化问题,基于凸优化理论可得au的解为:2.3.2) Based on the obtained pu,1 and
Figure BDA0002865158830000127
Optimize the variable au , the optimization problem constructed at this time is transformed into a convex optimization problem about au , and the solution of au can be obtained based on the convex optimization theory as:

Figure BDA0002865158830000128
Figure BDA0002865158830000128

其中

Figure BDA0002865158830000129
表示如果x大于1,则取值为1,如果x小于0,则取值为0。in
Figure BDA0002865158830000129
Indicates that if x is greater than 1, the value is 1, and if x is less than 0, the value is 0.

2.3.3)基于得到的au优化变量

Figure BDA00028651588300001210
此时构造的优化问题被转换为关于
Figure BDA00028651588300001211
的凸优化问题,基于凸优化理论可得
Figure BDA00028651588300001212
的解为
Figure BDA00028651588300001213
2.3.3) Optimize variables based on the obtained au
Figure BDA00028651588300001210
The optimization problem constructed at this point is transformed into
Figure BDA00028651588300001211
The convex optimization problem of , based on the convex optimization theory, we can get
Figure BDA00028651588300001212
The solution is
Figure BDA00028651588300001213

2.3.4)基于得到的au优化变量pu,1,此时目标函数的第一项为关于pu,1的非凸函数,第二项为关于pu,1的凹函数,因此需要对第一项进行连续凸逼近,此时变量pu,1的解可由以下优化问题获得:2.3.4) Optimize the variable pu,1 based on the obtained au . At this time, the first item of the objective function is a non-convex function about p u,1 , and the second item is a concave function about pu,1 . Therefore, it is necessary to A continuous convex approximation is performed for the first term, and the solution to the variable pu,1 can be obtained by the following optimization problem:

Figure BDA00028651588300001214
Figure BDA00028651588300001214

其中,

Figure BDA0002865158830000131
Figure BDA0002865158830000132
表示函数f(pu,1)在pu,1的一阶导数并令
Figure BDA0002865158830000133
Figure BDA0002865158830000134
表示第i次迭代得到的pu,1值;上述优化问题OP2为关于pu,1的凸优化问题,通过搜索F'(pu,1)=0的根即可得到优化问题OP2中pu,1的解;基于连续凸逼近的思想,反复更新的pu,1值直至相临两次迭代的pu,1差值小于一定精度,即可认为算法收敛,最新得到的pu,1值为优化问题OP1中pu,1的解。in,
Figure BDA0002865158830000131
Figure BDA0002865158830000132
Denote the first derivative of the function f(pu,1 ) at pu,1 and let
Figure BDA0002865158830000133
Figure BDA0002865158830000134
Represents the value of pu,1 obtained by the i-th iteration; the above optimization problem OP2 is a convex optimization problem about pu,1 , and p in the optimization problem OP2 can be obtained by searching for the root of F'(pu,1 )=0 The solution ofu,1 ; based on the idea of continuous convex approximation, the value of pu,1 is repeatedly updated until the difference of pu,1 between two adjacent iterations is less than a certain precision, the algorithm can be considered to be convergent, and the latest obtained pu, The value of1 is the solution of pu,1 in the optimization problem OP1.

2.3.5)基于步骤2.3.3)和2.3.4)中得到的

Figure BDA0002865158830000135
Figure BDA0002865158830000136
更新au,再基于得到的au,刷新
Figure BDA0002865158830000137
Figure BDA0002865158830000138
不断循环直至算法收敛,收敛时得到的au、pu,1
Figure BDA0002865158830000139
即为步骤2.2)中本地计算量au、本地传输功率pu,1及本地计算
Figure BDA00028651588300001310
的最终解。2.3.5) Based on the results obtained in steps 2.3.3) and 2.3.4)
Figure BDA0002865158830000135
and
Figure BDA0002865158830000136
Update au , then based on the obtained au , refresh
Figure BDA0002865158830000137
and
Figure BDA0002865158830000138
Keep looping until the algorithm converges, and the au , pu,1 and
Figure BDA0002865158830000139
That is, in step 2.2), the local calculation amount au , the local transmission power pu,1 and the local calculation
Figure BDA00028651588300001310
the final solution.

步骤三:卸载用户u首先将卸载任务的1-au部分以传输功率pu,1发送至选择的协助设备u',然后在协助设备的辅助下采用协作通信的方式将卸载任务的1-au部分卸载至边缘服务器nu,此时卸载用户u和协助设备u'的协作传输功率分别为pu,2及pu',pu,2及pu'的解可由以下优化问题获得:Step 3: The offloading user u first sends the 1-au part of the offloading task to the selected assisting device u' with the transmission power pu,1 , and then uses cooperative communication with the assistance of the assisting device to transfer the 1-a u part of the offloading task. The part of au is offloaded to the edge servernu , and the cooperative transmission power of the offloading user u and the assisting device u' is pu,2 and pu' respectively. The solutions ofpu,2 andpu' can be obtained by the following optimization problem :

Figure BDA00028651588300001311
Figure BDA00028651588300001311

此处优化变量pu,2和pu'有上下界限制,若超过了边界值,则直接设定为边界值;此时优化问题OP3的目标函数前两项分别为关于pu,2和pu'的非凸函数,后两项分别为关于pu,2和pu'的凹函数,因此需要对前两项进行连续凸逼近,此时变量pu,2和pu'的解可由以下优化问题获得:Here, the optimization variables pu,2 and pu' have upper and lower bounds. If they exceed the boundary value, they are directly set as the boundary value; at this time, the first two items of the objective function of the optimization problem OP3 are about pu,2 and The non-convex function of pu' , the last two terms are concave functions about pu,2 and pu' respectively, so it is necessary to carry out continuous convex approximation to the first two terms. At this time, the solutions of the variables pu,2 and pu' It can be obtained by the following optimization problem:

Figure BDA00028651588300001312
Figure BDA00028651588300001312

其中,

Figure BDA00028651588300001313
Figure BDA00028651588300001314
表示函数h(pu,2,pu')在pu,2的一阶导数并令
Figure BDA00028651588300001315
Figure BDA00028651588300001316
表示第i次迭代得到的pu,2值,
Figure BDA00028651588300001317
表示函数h(pu,2,pu')在pu'的一阶导数并令
Figure BDA00028651588300001318
Figure BDA00028651588300001319
表示第i次迭代得到的pu'值;上述优化问题OP4为关于pu,2和pu'的凸优化问题,通过搜索H'(pu,2)=0和H'(pu')=0的根即可得到优化问题OP3中pu,2和pu'的解;基于连续凸逼近的思想,反复更新的pu,2和pu'值直至相临两次迭代的pu,2和pu'的差值分别小于一定精度,即可认为算法收敛,收敛时得到的pu,2和pu'即为步骤3中卸载用户u和协助设备u'的协作传输功率pu,2和pu'的最终解。in,
Figure BDA00028651588300001313
Figure BDA00028651588300001314
Denote the first derivative of the function h(pu,2 ,pu' ) at pu,2 and let
Figure BDA00028651588300001315
Figure BDA00028651588300001316
represents the value of pu,2 obtained by the ith iteration,
Figure BDA00028651588300001317
Denote the first derivative of the function h(pu,2 ,pu' ) at pu' and let
Figure BDA00028651588300001318
Figure BDA00028651588300001319
represents the value of pu' obtained by the ith iteration; the above optimization problem OP4 is a convex optimization problem about pu,2 and pu' , by searching H'(pu,2 )=0 and H'(pu' )=0, the solutions of pu, 2 and pu' in optimization problem OP3 can be obtained; based on the idea of continuous convex approximation, the values of p u, 2 and p u' are updated repeatedly until the pu, 2 and pu' values of two adjacent iterations are reached. If the difference betweenu,2 and pu' is less than a certain precision, the algorithm can be considered to be converged, and the obtained pu,2 and pu' during convergence are the cooperative transmission power of the unloaded user u and the assisting device u' instep 3. The final solution of pu,2 and pu' .

步骤四:边缘服务器nu根据卸载用户u对于计算时延和计算能耗的权重指标,确定边缘计算任务量bu、边缘计算时延

Figure BDA0002865158830000141
及云计算任务量1-bu、云计算时延
Figure BDA0002865158830000142
并将云计算任务1-bu进一步卸载至云服务器,从而完成整个卸载过程:Step 4: The edge server nu determines the edge computing task volume bu and the edge computing delay according to the weight index of the offloading user u on the computing delay and computing energy consumption
Figure BDA0002865158830000141
and cloud computing task volume 1-bu , cloud computing delay
Figure BDA0002865158830000142
And the cloud computing task 1-bu is further uninstalled to the cloud server to complete the entire uninstallation process:

4.1)边缘服务器nu根据步骤1中设定的权重优化边缘计算任务量bu、边缘计算时延

Figure BDA0002865158830000143
及云计算任务量1-bu、云计算时延
Figure BDA0002865158830000144
优化问题可建模为:4.1) The edge server nu optimizes the edge computing task volume bu and the edge computing delay according to the weight set instep 1
Figure BDA0002865158830000143
and cloud computing task volume 1-bu , cloud computing delay
Figure BDA0002865158830000144
The optimization problem can be modeled as:

Figure BDA0002865158830000145
Figure BDA0002865158830000145

其中ξMEC和ξCloud表示边缘服务器计算能量系数和云服务器计算能量系数(计算能量系数可由CPU利用率乘以能效系数表示);此处优化变量

Figure BDA0002865158830000146
Figure BDA0002865158830000147
有上下界限制,若超过了边界值,则直接设定为边界值。Among them, ξMEC and ξCloud represent the computing energy coefficient of edge server and cloud server computing energy coefficient (the computing energy coefficient can be represented by the CPU utilization multiplied by the energy efficiency coefficient); the optimization variables here
Figure BDA0002865158830000146
and
Figure BDA0002865158830000147
There are upper and lower bounds. If the boundary value is exceeded, it is directly set as the boundary value.

4.2)边缘服务器nu采用交替迭代的方式确定变量bu

Figure BDA0002865158830000148
Figure BDA0002865158830000149
4.2) The edge server nu determines the variables bu ,
Figure BDA0002865158830000148
and
Figure BDA0002865158830000149

4.2.1)对边缘服务器nu初始化一组

Figure BDA00028651588300001410
Figure BDA00028651588300001411
的值;4.2.1) Initialize a set of edge servers nu
Figure BDA00028651588300001410
and
Figure BDA00028651588300001411
the value of;

4.2.2)基于得到的

Figure BDA00028651588300001412
Figure BDA00028651588300001413
优化变量bu,此时构造的优化问题被转换为关于bu的凸优化问题,基于凸优化理论可得bu的解为:4.2.2) Based on the obtained
Figure BDA00028651588300001412
and
Figure BDA00028651588300001413
Optimize the variable bu , the optimization problem constructed at this time is transformed into a convex optimization problem about bu , and the solution of bu can be obtained based on the convex optimization theory as:

Figure BDA00028651588300001414
Figure BDA00028651588300001414

其中

Figure BDA00028651588300001415
表示如果x大于1,则取值为1,如果x小于0,则取值为0。in
Figure BDA00028651588300001415
Indicates that if x is greater than 1, the value is 1, and if x is less than 0, the value is 0.

4.2.3)基于得到的bu优化变量

Figure BDA00028651588300001416
Figure BDA00028651588300001417
此时构造的优化问题被转换为关于
Figure BDA00028651588300001418
Figure BDA00028651588300001419
的凸优化问题,基于凸优化理论可得
Figure BDA00028651588300001420
Figure BDA00028651588300001421
的解分为:4.2.3) Based on the obtained bu optimization variables
Figure BDA00028651588300001416
and
Figure BDA00028651588300001417
The optimization problem constructed at this point is transformed into
Figure BDA00028651588300001418
and
Figure BDA00028651588300001419
The convex optimization problem of , based on the convex optimization theory, we can get
Figure BDA00028651588300001420
and
Figure BDA00028651588300001421
The solution is divided into:

Figure BDA00028651588300001422
Figure BDA00028651588300001423
Figure BDA00028651588300001422
and
Figure BDA00028651588300001423

4.2.4)基于步骤4.2.3)中得到的

Figure BDA0002865158830000151
Figure BDA0002865158830000152
更新bu,再基于得到的bu,刷新
Figure BDA0002865158830000153
Figure BDA0002865158830000154
不断循环直至算法收敛,收敛时得到的bu
Figure BDA0002865158830000155
Figure BDA0002865158830000156
即为步骤4.1)中边缘计算任务量bu、边缘计算时延
Figure BDA0002865158830000157
及云计算时延
Figure BDA0002865158830000158
的最终解。4.2.4) Based on the result obtained in step 4.2.3)
Figure BDA0002865158830000151
and
Figure BDA0002865158830000152
update bu , and then based on the obtained bu , refresh
Figure BDA0002865158830000153
and
Figure BDA0002865158830000154
Continuously loop until the algorithm converges, and thebu ,
Figure BDA0002865158830000155
and
Figure BDA0002865158830000156
That is, the amount of edge computing tasks bu and the edge computing delay in step 4.1)
Figure BDA0002865158830000157
and cloud computing latency
Figure BDA0002865158830000158
the final solution.

步骤五:卸载用户u、边缘服务器nu及云服务器基于优化结果分别以计算速率

Figure BDA0002865158830000159
Figure BDA00028651588300001510
处理卸载用户u的任务中au部分、bu(1-au)部分及(1-bu)(1-au)部分,计算完毕后卸载用户u汇总各部分计算结果,并生成卸载任务的最终计算结果。Step 5: Uninstall the user u, the edge server nu and the cloud server based on the optimization results and calculate the speed respectively
Figure BDA0002865158830000159
and
Figure BDA00028651588300001510
Process the au part, bu (1-au ) part and (1-bu )(1-au ) part of the task of uninstalling user u, after the calculation is completed, the uninstall user u summarizes the calculation results of each part, and generates the uninstallation The final calculation result of the task.

下面结合附图对本发明的技术效果作详细的描述。The technical effects of the present invention will be described in detail below with reference to the accompanying drawings.

如图5所示,给出了本发明与传统策略对比示例图。图例横坐标为卸载用户的总任务量;(a)对比了不同能耗权重下的用户卸载时延,(b)对比了不同能耗权重下的用户卸载能耗,其中点实线和圆实线所代表的能耗权重分别为

Figure BDA00028651588300001511
Figure BDA00028651588300001512
相对应的时延权重则分别为
Figure BDA00028651588300001513
Figure BDA00028651588300001514
图例假设频谱资源归一化,各单位能量系数为1。从图中可以看出,随着总任务量的增加,用户卸载总时延和总能耗不可避免得会产生上升趋势;但根据卸载任务类型的不同,或者说卸载任务的指标需求的不同,调节能耗、时延权重因子,在能耗权重更大时候,可以通过牺牲一定卸载时延的方式,换取更低的卸载能耗,反之亦成立。因此时延能耗权重表征了卸载任务对于时延和能耗的需求,合理的设定时延能耗权重,可实现卸载时延和卸载能耗的性能折中。As shown in FIG. 5 , an example diagram of the comparison between the present invention and the traditional strategy is given. The abscissa of the legend is the total amount of unloaded users; (a) compares the user unloading delay under different energy consumption weights, (b) compares the user unloading energy consumption under different energy consumption weights, where the dotted line and the circle are solid The power consumption weights represented by the lines are
Figure BDA00028651588300001511
and
Figure BDA00028651588300001512
The corresponding delay weights are
Figure BDA00028651588300001513
and
Figure BDA00028651588300001514
The legend assumes that the spectrum resources are normalized, and the energy coefficient of each unit is 1. It can be seen from the figure that with the increase of the total amount of tasks, the total time delay and total energy consumption of user unloading will inevitably have an upward trend; Adjust the energy consumption and delay weight factors. When the energy consumption weight is larger, a certain unloading delay can be sacrificed in exchange for lower unloading energy consumption, and vice versa. Therefore, the delay and energy consumption weight represents the requirements of the offloading task for the delay and energy consumption, and a reasonable setting of the delay energy consumption weight can achieve a performance trade-off between the offloading delay and the offloading energy consumption.

如图6所示,给出了本发明与其他技术效果的性能对比图;(a)对比了不同技术下的用户卸载时延,(b)对比了不同技术的用户卸载能耗。对比方案为仅由边缘计算服务器为用户提供业务服务,不考虑云中心资源和业务管理的协同。可以看到,通过边缘协同以及计算资源和任务拆分比例的合理划分,可以进一步在时延和可靠性两方面提升系统性能。As shown in FIG. 6 , a performance comparison diagram of the present invention and other technical effects is given; (a) compares the user unloading delay under different technologies, (b) compares the user unloading energy consumption of different technologies. The comparison scheme is that only the edge computing server provides business services to users, and does not consider the coordination of cloud center resources and business management. It can be seen that through edge collaboration and a reasonable division of computing resources and task split ratios, system performance can be further improved in terms of latency and reliability.

应当注意,本发明的实施方式可以通过硬件、软件或者软件和硬件的结合来实现。硬件部分可以利用专用逻辑来实现;软件部分可以存储在存储器中,由适当的指令执行系统,例如微处理器或者专用设计硬件来执行。本领域的普通技术人员可以理解上述的设备和方法可以使用计算机可执行指令和/或包含在处理器控制代码中来实现,例如在诸如磁盘、CD或DVD-ROM的载体介质、诸如只读存储器(固件)的可编程的存储器或者诸如光学或电子信号载体的数据载体上提供了这样的代码。本发明的设备及其模块可以由诸如超大规模集成电路或门阵列、诸如逻辑芯片、晶体管等的半导体、或者诸如现场可编程门阵列、可编程逻辑设备等的可编程硬件设备的硬件电路实现,也可以用由各种类型的处理器执行的软件实现,也可以由上述硬件电路和软件的结合例如固件来实现。It should be noted that the embodiments of the present invention may be implemented by hardware, software, or a combination of software and hardware. The hardware portion may be implemented using special purpose logic; the software portion may be stored in memory and executed by a suitable instruction execution system, such as a microprocessor or specially designed hardware. Those of ordinary skill in the art will appreciate that the apparatus and methods described above may be implemented using computer-executable instructions and/or embodied in processor control code, for example on a carrier medium such as a disk, CD or DVD-ROM, such as a read-only memory Such code is provided on a programmable memory (firmware) or a data carrier such as an optical or electronic signal carrier. The device and its modules of the present invention can be implemented by hardware circuits such as very large scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, etc., or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc., It can also be implemented by software executed by various types of processors, or by a combination of the above-mentioned hardware circuits and software, such as firmware.

以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,都应涵盖在本发明的保护范围之内。The above are only specific embodiments of the present invention, but the protection scope of the present invention is not limited to this. Any person skilled in the art is within the technical scope disclosed by the present invention, and all within the spirit and principle of the present invention Any modifications, equivalent replacements and improvements made within the scope of the present invention should be included within the protection scope of the present invention.

Claims (10)

Translated fromChinese
1.一种边云协同资源联合分配方法,其特征在于,所述边云协同资源联合分配方法卸载用户首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。1. A method for joint allocation of edge-cloud collaborative resources, characterized in that, in the method for joint allocation of edge-cloud collaborative resources, the offloading user first determines delay and energy consumption weight based on its own service type; the offloading user selects an assisting device according to a link state. , and determine the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight. It is noted that in order to avoid interference between users, users within the coverage of adjacent edge servers use frequency division multiple access; users First, transfer the upload task to the selected assisting device, and then use cooperative communication to offload the uploading task to the edge server with the assistance of the assisting device, so as to ensure that the edge server receives the offloading task accurately. At this time, it is necessary to determine the process of collaborative communication. The edge server determines the amount of edge computing tasks, edge computing delay, cloud computing task amount, and cloud computing delay according to the user's weight index of computing delay and computing energy consumption, and further offloads the cloud computing tasks to Cloud server to complete the entire uninstallation process.2.如权利要求1所述的边云协同资源联合分配方法,其特征在于,所述边云协同资源联合分配方法具体包括:2. The method for joint allocation of side-cloud collaborative resources according to claim 1, wherein the method for joint allocation of collaborative resources for side-cloud specifically comprises:(1)卸载用户基于自身的业务类型确定时延和能耗权重,令
Figure FDA0002865158820000011
Figure FDA0002865158820000012
分别表示用户u对于通信能耗、通信时延、计算能耗和计算时延的权重指标,其中
Figure FDA0002865158820000013
Figure FDA0002865158820000014
权重值越大,表明该用户的卸载任务对于该项指标越敏感;(1) Offloading users determine the delay and energy consumption weights based on their own service types, so that
Figure FDA0002865158820000011
and
Figure FDA0002865158820000012
respectively represent the weight indicators of user u for communication energy consumption, communication delay, computing energy consumption and computing delay, where
Figure FDA0002865158820000013
and
Figure FDA0002865158820000014
The larger the weight value, the more sensitive the user's uninstallation task is to this indicator;(2)卸载用户根据链路状态选择协助设备,根据权重确定本地计算量、上传任务量、本地传输功率及本地计算时延:(2) The offloading user selects the assisting device according to the link status, and determines the local calculation amount, upload task amount, local transmission power and local calculation delay according to the weight:(2.1)假设用户u的备选协助设备集合为Uu,令备选设备u'作为其协助设备,则
Figure FDA0002865158820000015
其中gu,u'
Figure FDA0002865158820000016
Figure FDA0002865158820000017
分别表示用户u到备选设备u'的功率增益、用户u到用户u关联的边缘服务器nu的功率增益和备选设备u'到用户u关联的边缘服务器nu的功率增益;
(2.1) Assuming that the set of alternative assisting devices for user u is Uu , let the alternative device u' be its assisting device, then
Figure FDA0002865158820000015
where gu,u' ,
Figure FDA0002865158820000016
and
Figure FDA0002865158820000017
respectively represent the power gain from user u to the candidate device u', the power gain from user u to the edge server nu associated with useru , and the power gain from the candidate deviceu ' to the edge server nu associated with user u;
(2.2)用户u可根据(1)中设定的权重优化本地计算量au、上传任务量1-au、本地传输功率pu,1及本地计算时延
Figure FDA0002865158820000018
优化问题可建模为:
(2.2) The user u can optimize the local computation amount au , the upload task amount 1-au , the local transmission power pu,1 and the local computation delay according to the weight set in (1)
Figure FDA0002865158820000018
The optimization problem can be modeled as:
Figure FDA0002865158820000019
Figure FDA0002865158820000019
其中,Iu、Bu、c、
Figure FDA0002865158820000021
ξUser分别表示用户u的总任务量、卸载频带宽度(、计算每比特任务所需消耗的CPU数、本地计算时延、用户设备计算能量系数;此外,优化变量pu,1
Figure FDA0002865158820000022
有上下界限制,若超过了边界值,则直接设定为边界值;
Among them, Iu , Bu , c,
Figure FDA0002865158820000021
ξUser represents the total task amount of user u, the offload bandwidth (, the number of CPUs required to calculate each bit task, the local computing delay, and the user equipment computing energy coefficient; in addition, the optimization variables pu,1 and
Figure FDA0002865158820000022
There are upper and lower bounds, if the boundary value is exceeded, it is directly set as the boundary value;
(2.3)卸载用户u采用交替迭代的方式确定变量au、pu,1
Figure FDA0002865158820000023
(2.3) Unloading user u determines the variables au , pu,1 and
Figure FDA0002865158820000023
(3)卸载用户u首先将卸载任务的1-au部分以传输功率pu,1发送至选择的协助设备u',然后在协助设备的辅助下采用协作通信的方式将卸载任务的1-au部分卸载至边缘服务器nu,此时卸载用户u和协助设备u'的协作传输功率分别为pu,2及pu',pu,2及pu'的解由以下优化问题获得:(3) The offloading user u first sends the 1-au part of the offloading task to the selected assisting device u' with the transmission power pu,1 , and then uses cooperative communication with the assistance of the assisting device to send the 1-a u part of the offloading task to the selected assisting device u'. The part of au is offloaded to the edge server nu . At this time, the cooperative transmission power of the offloaded user u and the assisting device u' is pu,2 and pu ' respectively. The solutions of pu,2 and pu' are obtained by the following optimization problem :
Figure FDA0002865158820000024
Figure FDA0002865158820000024
此处优化变量pu,2和pu'有上下界限制,若超过了边界值,则直接设定为边界值;此时优化问题OP3的目标函数前两项分别为关于pu,2和pu'的非凸函数,后两项分别为关于pu,2和pu'的凹函数,因此需要对前两项进行连续凸逼近,此时变量pu,2和pu'的解可由以下优化问题获得:Here, the optimization variables pu,2 and pu' have upper and lower bounds. If they exceed the boundary value, they are directly set as the boundary value; at this time, the first two items of the objective function of the optimization problem OP3 are about pu,2 and The non-convex function of pu' , the last two terms are concave functions about pu,2 and pu' respectively, so it is necessary to carry out continuous convex approximation to the first two terms. At this time, the solutions of the variables pu,2 and pu' It can be obtained by the following optimization problem:
Figure FDA0002865158820000025
Figure FDA0002865158820000025
其中,角标i表示迭代次数;
Figure FDA0002865158820000026
Figure FDA0002865158820000027
表示函数h(pu,2,pu')在pu,2的一阶导数并令
Figure FDA0002865158820000028
Figure FDA0002865158820000029
表示第i次迭代得到的pu,2值;
Figure FDA00028651588200000210
表示函数h(pu,2,pu')在pu'的一阶导数并代入
Figure FDA00028651588200000211
Figure FDA00028651588200000212
表示第i次迭代得到的pu'值;上述优化问题OP4为关于pu,2和pu'的凸优化问题,通过搜索H'(pu,2)=0和H'(pu')=0的根即可得到优化问题OP3中pu,2和pu'的解;基于连续凸逼近的思想,反复更新的pu,2和pu'值直至相临两次迭代的pu,2和pu'的差值分别小于一定精度,即可认为算法收敛,收敛时得到的pu,2和pu'即为(3)中卸载用户u和协助设备u'的协作传输功率pu,2和pu'的最终解;
Among them, the index i represents the number of iterations;
Figure FDA0002865158820000026
Figure FDA0002865158820000027
Denote the first derivative of the function h(pu,2 ,pu' ) at pu,2 and let
Figure FDA0002865158820000028
Figure FDA0002865158820000029
Indicates the pu,2 value obtained by the i-th iteration;
Figure FDA00028651588200000210
Represent the first derivative of the function h(pu,2 ,pu' ) at pu' and substitute it into
Figure FDA00028651588200000211
Figure FDA00028651588200000212
represents the value of pu' obtained by the ith iteration; the above optimization problem OP4 is a convex optimization problem about pu,2 and pu' , by searching H'(pu,2 )=0 and H'(pu' )=0, the solutions of pu, 2 and pu' in optimization problem OP3 can be obtained; based on the idea of continuous convex approximation, the values of p u, 2 and p u' are updated repeatedly until the pu, 2 and pu' values of two adjacent iterations are reached. If the difference betweenu,2 and pu' is less than a certain precision, it can be considered that the algorithm converges, and the pu,2 and pu' obtained during convergence are the cooperative transmission of the unloaded user u and the assisting device u' in (3). The final solution of power pu,2 and pu' ;
(4)边缘服务器nu根据卸载用户u对于计算时延和计算能耗的权重指标,确定边缘计算任务量bu、边缘计算时延
Figure FDA0002865158820000031
及云计算任务量1-bu、云计算时延
Figure FDA0002865158820000032
并将云计算任务1-bu进一步卸载至云服务器,完成整个卸载过程:
(4) The edge server nu determines the edge computing task amount bu and the edge computing delay according to the weight index of the offloading user u on the computing delay and computing energy consumption
Figure FDA0002865158820000031
and cloud computing task volume 1-bu , cloud computing delay
Figure FDA0002865158820000032
And further uninstall the cloud computing task 1-bu to the cloud server to complete the entire uninstallation process:
(4.1)边缘服务器nu根据(1)中设定的权重优化边缘计算任务量bu、边缘计算时延
Figure FDA0002865158820000033
及云计算任务量1-bu、云计算时延
Figure FDA0002865158820000034
优化问题可建模为:
(4.1) The edge server nu optimizes the edge computing task volume bu and the edge computing delay according to the weight set in (1).
Figure FDA0002865158820000033
and cloud computing task volume 1-bu , cloud computing delay
Figure FDA0002865158820000034
The optimization problem can be modeled as:
Figure FDA0002865158820000035
Figure FDA0002865158820000035
其中ξMEC和ξCloud表示边缘服务器计算能量系数和云服务器计算能量系数;此处优化变量
Figure FDA0002865158820000036
Figure FDA0002865158820000037
有上下界限制,若超过了边界值,则直接设定为边界值;
where ξMEC and ξCloud represent the computing energy coefficient of edge server and the computing energy coefficient of cloud server; the variables are optimized here
Figure FDA0002865158820000036
and
Figure FDA0002865158820000037
There are upper and lower bounds, if the boundary value is exceeded, it is directly set as the boundary value;
(4.2)边缘服务器nu采用交替迭代的方式确定变量bu
Figure FDA0002865158820000038
Figure FDA0002865158820000039
(4.2) The edge server nu determines the variables bu ,
Figure FDA0002865158820000038
and
Figure FDA0002865158820000039
(5)卸载用户u、边缘服务器nu及云服务器基于优化结果分别以计算速率
Figure FDA00028651588200000310
Figure FDA00028651588200000311
处理卸载用户u的任务中au部分、bu(1-au)部分及(1-bu)(1-au)部分,计算完毕后卸载用户u汇总各部分计算结果,并生成卸载任务的最终计算结果。
(5) The offloading user u, the edge server nu and the cloud server respectively calculate the speed based on the optimization result.
Figure FDA00028651588200000310
and
Figure FDA00028651588200000311
Process the au part, bu (1-au ) part and (1-bu )(1-au ) part of the task of uninstalling user u, after the calculation is completed, the uninstall user u summarizes the calculation results of each part, and generates the uninstallation The final calculation result of the task.
3.如权利要求2所述的边云协同资源联合分配方法,其特征在于,所述步骤(2.3)按如下步骤进行:3. The method for joint allocation of edge-cloud collaborative resources as claimed in claim 2, wherein the step (2.3) is carried out as follows:(2.3a)对每个卸载用户u初始化一组pu,1
Figure FDA00028651588200000312
的值;
(2.3a) Initialize a set of pu,1 and
Figure FDA00028651588200000312
the value of;
(2.3b)基于得到的pu,1
Figure FDA00028651588200000313
优化变量au,此时构造的优化问题被转换为关于au的凸优化问题,基于凸优化理论可得au的解为:
(2.3b) Based on the obtained pu,1 and
Figure FDA00028651588200000313
Optimize the variable au , the optimization problem constructed at this time is transformed into a convex optimization problem about au , and the solution of au can be obtained based on the convex optimization theory as:
Figure FDA00028651588200000314
Figure FDA00028651588200000314
如果得到的au*大于1,则取值为1;If the obtained au* is greater than 1, the value is 1;(2.3c)基于得到的au优化变量
Figure FDA00028651588200000315
此时构造的优化问题被转换为关于
Figure FDA00028651588200000316
的凸优化问题,基于凸优化理论可得
Figure FDA00028651588200000317
的解为
Figure FDA00028651588200000318
(2.3c) Optimization variables based on the obtained au
Figure FDA00028651588200000315
The optimization problem constructed at this point is transformed into
Figure FDA00028651588200000316
The convex optimization problem of , based on the convex optimization theory, we can get
Figure FDA00028651588200000317
The solution is
Figure FDA00028651588200000318
(2.3d)基于得到的au优化变量pu,1,此时目标函数的第一项为关于pu,1的非凸函数,第二项为关于pu,1的凹函数,因此需要对第一项进行连续凸逼近,此时变量pu,1的解可由以下优化问题获得:(2.3d) Optimize the variable pu,1 based on the obtained au . At this time, the first item of the objective function is a non-convex function about p u,1 , and the second item is a concave function about pu,1 . Therefore, it is necessary to A continuous convex approximation is performed for the first term, and the solution to the variable pu,1 can be obtained by the following optimization problem:
Figure FDA0002865158820000041
Figure FDA0002865158820000041
其中,角标i表示迭代次数;
Figure FDA0002865158820000042
Figure FDA0002865158820000043
表示函数f(pu,1)在pu,1的一阶导数并代入
Figure FDA0002865158820000044
Figure FDA0002865158820000045
表示第i次迭代得到的pu,1值;上述优化问题OP2为关于pu,1的凸优化问题,通过搜索F'(pu,1)=0的根即可得到优化问题OP2中pu,1的解;基于连续凸逼近的思想,反复更新的pu,1值直至相临两次迭代的pu,1差值小于一定精度,即可认为算法收敛,最新得到的pu,1值为优化问题OP1中pu,1的解;
Among them, the index i represents the number of iterations;
Figure FDA0002865158820000042
Figure FDA0002865158820000043
Represent the first derivative of the function f(pu,1 ) at pu,1 and substitute it into
Figure FDA0002865158820000044
Figure FDA0002865158820000045
Represents the value of pu,1 obtained by the i-th iteration; the above optimization problem OP2 is a convex optimization problem about pu,1 , and p in the optimization problem OP2 can be obtained by searching for the root of F'(pu,1 )=0 The solution ofu,1 ; based on the idea of continuous convex approximation, the value of pu,1 is repeatedly updated until the difference of pu,1 between two adjacent iterations is less than a certain precision, the algorithm can be considered to be convergent, and the latest obtained pu, 1 is the solution of pu,1 in optimization problem OP1;
(2.3e)基于步骤(2.3c)和(2.3d)中得到的
Figure FDA0002865158820000046
Figure FDA0002865158820000047
更新au,再基于得到的au,刷新
Figure FDA0002865158820000048
Figure FDA0002865158820000049
不断循环直至算法收敛,收敛时得到的au、pu,1
Figure FDA00028651588200000410
即为步骤(2.2)中本地计算量au、本地传输功率pu,1及本地计算
Figure FDA00028651588200000411
的最终解。
(2.3e) based on the results obtained in steps (2.3c) and (2.3d)
Figure FDA0002865158820000046
and
Figure FDA0002865158820000047
Update au , then based on the obtained au , refresh
Figure FDA0002865158820000048
and
Figure FDA0002865158820000049
Keep looping until the algorithm converges, and the au , pu,1 and
Figure FDA00028651588200000410
That is, in step (2.2), the local calculation amount au , the local transmission power pu,1 and the local calculation
Figure FDA00028651588200000411
the final solution.
4.如权利要求2所述的边云协同资源联合分配方法,其特征在于,所述步骤(4.2)按如下步骤进行:4. The method for joint allocation of edge-cloud collaborative resources as claimed in claim 2, wherein the step (4.2) is carried out as follows:(4.2a)对边缘服务器nu初始化一组
Figure FDA00028651588200000412
Figure FDA00028651588200000413
的值;
(4.2a) Initialize a set of edge servers nu
Figure FDA00028651588200000412
and
Figure FDA00028651588200000413
the value of;
(4.2b)基于得到的
Figure FDA00028651588200000414
Figure FDA00028651588200000415
优化变量bu,此时构造的优化问题被转换为关于bu的凸优化问题,基于凸优化理论可得bu的解为:
(4.2b) Based on the obtained
Figure FDA00028651588200000414
and
Figure FDA00028651588200000415
Optimize the variable bu , the optimization problem constructed at this time is transformed into a convex optimization problem about bu , and the solution of bu can be obtained based on the convex optimization theory as:
Figure FDA00028651588200000416
Figure FDA00028651588200000416
(4.2c)基于得到的bu优化变量
Figure FDA00028651588200000417
Figure FDA00028651588200000418
此时构造的优化问题被转换为关于
Figure FDA00028651588200000419
Figure FDA00028651588200000420
的凸优化问题,基于凸优化理论可得
Figure FDA00028651588200000421
Figure FDA00028651588200000422
的解分为:
(4.2c) Optimization variables based on the obtained bu
Figure FDA00028651588200000417
and
Figure FDA00028651588200000418
The optimization problem constructed at this point is transformed into
Figure FDA00028651588200000419
and
Figure FDA00028651588200000420
The convex optimization problem of , based on the convex optimization theory, we can get
Figure FDA00028651588200000421
and
Figure FDA00028651588200000422
The solution is divided into:
Figure FDA00028651588200000423
Figure FDA00028651588200000424
Figure FDA00028651588200000423
and
Figure FDA00028651588200000424
(4.2d)基于步骤(4.2c)中得到的
Figure FDA00028651588200000425
Figure FDA00028651588200000426
更新bu,再基于得到的bu,刷新
Figure FDA0002865158820000051
Figure FDA0002865158820000052
不断循环直至算法收敛,收敛时得到的bu
Figure FDA0002865158820000053
Figure FDA0002865158820000054
即为步骤(4.1)中边缘计算任务量bu、边缘计算时延
Figure FDA0002865158820000055
及云计算时延
Figure FDA0002865158820000056
的最终解。
(4.2d) based on the result obtained in step (4.2c)
Figure FDA00028651588200000425
and
Figure FDA00028651588200000426
update bu , and then based on the obtained bu , refresh
Figure FDA0002865158820000051
and
Figure FDA0002865158820000052
Continuously loop until the algorithm converges, and thebu ,
Figure FDA0002865158820000053
and
Figure FDA0002865158820000054
That is, the amount of edge computing tasks bu and the edge computing delay in step (4.1)
Figure FDA0002865158820000055
and cloud computing latency
Figure FDA0002865158820000056
the final solution.
5.一种计算机设备,其特征在于,所述计算机设备包括存储器和处理器,所述存储器存储有计算机程序,所述计算机程序被所述处理器执行时,使得所述处理器执行如下步骤:首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。5. A computer device, characterized in that the computer device comprises a memory and a processor, the memory stores a computer program, and when the computer program is executed by the processor, the processor is caused to perform the following steps: First, determine the delay and energy consumption weights based on their own service types; the offload user selects the assisting device according to the link status, and determines the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight. Users within the coverage area of adjacent edge servers use frequency division multiple access; users first transmit the upload task to the selected assisting device, and then use cooperative communication to offload the uploading task with the assistance of the assisting device. In this case, the cooperative transmission power in the cooperative communication process needs to be determined; the edge server determines the amount of edge computing tasks, Edge computing delay and cloud computing task volume, cloud computing delay, and the cloud computing task is further offloaded to the cloud server to complete the entire offloading process.6.一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行如下步骤:首先基于自身的业务类型确定时延和能耗权重;卸载用户根据链路状态选择协助设备,并根据权重确定本地计算量、上传任务量、本地传输功率、本地计算时延,注意到为避免用户间的干扰,相临边缘服务器覆盖范围内的用户采用频分多址接入方式;用户首先将上传任务传输至选择的协助设备,然后在协助设备的辅助下采用协作通信的方式将上传任务卸载至边缘服务器,从而确保边缘服务器准确地接收到卸载任务,此时需要确定协作通信过程中的协作传输功率;边缘服务器根据用户对于计算时延和计算能耗的权重指标,确定边缘计算任务量、边缘计算时延及云计算任务量、云计算时延,并将云计算任务进一步卸载至云服务器,从而完成整个卸载过程。6. A computer-readable storage medium storing a computer program, and when the computer program is executed by a processor, the processor is caused to perform the following steps: firstly determine a time delay and an energy consumption weight based on its own service type; Select the assisting device according to the link status, and determine the local calculation amount, upload task amount, local transmission power, and local calculation delay according to the weight. It is noted that in order to avoid interference between users, users within the coverage of adjacent edge servers use frequency division Multiple access mode; the user first transmits the uploading task to the selected assisting device, and then uses cooperative communication to offload the uploading task to the edge server with the assistance of the assisting device, so as to ensure that the edge server receives the offloading task accurately. It is necessary to determine the cooperative transmission power in the cooperative communication process; the edge server determines the edge computing task volume, edge computing delay and cloud computing task volume, cloud computing delay according to the user's weight index for computing delay and computing energy consumption, and The cloud computing tasks are further offloaded to the cloud server to complete the entire offloading process.7.一种信息数据处理终端,其特征在于,所述信息数据处理终端用于实现权利要求1~4任意一项所述的边云协同资源联合分配方法。7 . An information data processing terminal, characterized in that, the information data processing terminal is used to implement the joint edge-cloud collaborative resource allocation method according to any one of claims 1 to 4 .8.一种边缘计算服务器,其特征在于,所述边缘计算服务器用于实现权利要求1~4任意一项所述的边云协同资源联合分配方法。8 . An edge computing server, characterized in that, the edge computing server is used to implement the method for joint allocation of edge-cloud collaborative resources according to any one of claims 1 to 4 .9.一种云计算服务器,其特征在于,所述云计算服务器用于实现权利要求1~4任意一项所述的边云协同资源联合分配方法。9 . A cloud computing server, characterized in that, the cloud computing server is used to implement the method for joint allocation of edge-cloud collaborative resources according to any one of claims 1 to 4 .10.一种无线通信系统,其特征在于,所述无线通信系统用于实现权利要求1~4任意一项所述的边云协同资源联合分配方法。10 . A wireless communication system, characterized in that, the wireless communication system is used to implement the method for joint edge-cloud cooperative resource allocation according to any one of claims 1 to 4 .
CN202011584281.1A2020-12-282020-12-28Edge cloud cooperative resource joint allocation method, system and applicationActiveCN112689303B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202011584281.1ACN112689303B (en)2020-12-282020-12-28Edge cloud cooperative resource joint allocation method, system and application

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202011584281.1ACN112689303B (en)2020-12-282020-12-28Edge cloud cooperative resource joint allocation method, system and application

Publications (2)

Publication NumberPublication Date
CN112689303Atrue CN112689303A (en)2021-04-20
CN112689303B CN112689303B (en)2022-07-22

Family

ID=75453989

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202011584281.1AActiveCN112689303B (en)2020-12-282020-12-28Edge cloud cooperative resource joint allocation method, system and application

Country Status (1)

CountryLink
CN (1)CN112689303B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113191654A (en)*2021-05-122021-07-30余绍祥Distributed resource scheduling system and method based on multi-objective optimization
CN113419867A (en)*2021-08-232021-09-21浙大城市学院Energy-saving service supply method in edge-oriented cloud collaborative computing environment
CN113660625A (en)*2021-07-062021-11-16山东师范大学 Method and system for resource coordination between multi-access edge computing nodes and CDN nodes
CN113784373A (en)*2021-08-242021-12-10苏州大学Combined optimization method and system for time delay and frequency spectrum occupation in cloud edge cooperative network
CN113923223A (en)*2021-11-152022-01-11安徽大学User allocation method with low time cost in edge environment
CN114448991A (en)*2021-12-282022-05-06西安电子科技大学Multi-edge server selection method, system, medium, device and terminal
CN114844900A (en)*2022-05-052022-08-02中南大学Edge cloud resource cooperation method based on uncertain demand
CN115695231A (en)*2022-10-312023-02-03中国联合网络通信集团有限公司 Method, device and storage medium for determining network energy efficiency
CN116506877A (en)*2023-06-262023-07-28北京航空航天大学 A Distributed Collaborative Computing Method for Mobile Crowd Sensing
CN117193873A (en)*2023-09-052023-12-08北京科技大学 A computing offloading method and device suitable for industrial control systems
CN118260067A (en)*2024-02-182024-06-28北京中电飞华通信有限公司Computing resource distribution method based on simulated annealing and related equipment

Citations (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN106900011A (en)*2017-02-282017-06-27重庆邮电大学Task discharging method between a kind of cellular basestation based on MEC
CN107241767A (en)*2017-06-142017-10-10广东工业大学The method and device that a kind of mobile collaboration is calculated
WO2017196249A1 (en)*2016-05-132017-11-16Telefonaktiebolaget Lm Ericsson (Publ)Network architecture, methods, and devices for a wireless communications network
US20180220432A1 (en)*2017-02-012018-08-02Deutsche Telekom AgNetwork resources brokering system and brokering entity for accessing a network resource with preferential treatment
CN108804227A (en)*2018-05-232018-11-13大连理工大学The method of the unloading of computation-intensive task and best resource configuration based on mobile cloud computing
CN109474681A (en)*2018-11-052019-03-15安徽大学 Resource allocation method, system and server system for mobile edge computing server
CN109814951A (en)*2019-01-222019-05-28南京邮电大学 A joint optimization method for task offloading and resource allocation in mobile edge computing networks
CN111464983A (en)*2020-03-102020-07-28深圳大学 A computing and communication cooperation method and system in a passive edge computing network
CN111586762A (en)*2020-04-292020-08-25重庆邮电大学Task unloading and resource allocation joint optimization method based on edge cooperation
CN111585916A (en)*2019-12-262020-08-25国网辽宁省电力有限公司电力科学研究院LTE electric power wireless private network task unloading and resource allocation method based on cloud edge cooperation
CN111913723A (en)*2020-06-152020-11-10合肥工业大学Cloud-edge-end cooperative unloading method and system based on assembly line

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2017196249A1 (en)*2016-05-132017-11-16Telefonaktiebolaget Lm Ericsson (Publ)Network architecture, methods, and devices for a wireless communications network
US20180220432A1 (en)*2017-02-012018-08-02Deutsche Telekom AgNetwork resources brokering system and brokering entity for accessing a network resource with preferential treatment
CN106900011A (en)*2017-02-282017-06-27重庆邮电大学Task discharging method between a kind of cellular basestation based on MEC
CN107241767A (en)*2017-06-142017-10-10广东工业大学The method and device that a kind of mobile collaboration is calculated
CN108804227A (en)*2018-05-232018-11-13大连理工大学The method of the unloading of computation-intensive task and best resource configuration based on mobile cloud computing
CN109474681A (en)*2018-11-052019-03-15安徽大学 Resource allocation method, system and server system for mobile edge computing server
CN109814951A (en)*2019-01-222019-05-28南京邮电大学 A joint optimization method for task offloading and resource allocation in mobile edge computing networks
CN111585916A (en)*2019-12-262020-08-25国网辽宁省电力有限公司电力科学研究院LTE electric power wireless private network task unloading and resource allocation method based on cloud edge cooperation
CN111464983A (en)*2020-03-102020-07-28深圳大学 A computing and communication cooperation method and system in a passive edge computing network
CN111586762A (en)*2020-04-292020-08-25重庆邮电大学Task unloading and resource allocation joint optimization method based on edge cooperation
CN111913723A (en)*2020-06-152020-11-10合肥工业大学Cloud-edge-end cooperative unloading method and system based on assembly line

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MENGQI YANG, JIAN CHEN, LONG YANG,LU LV ,BINGTAO HE, AND BOYANG: ""Design and Performance Analysis of Cooperative NOMA With Coordinated Direct and Relay Transmission"", 《IEEE ACCESS》*
杨龙,陈健: ""一种基于非正交多址接入技术的协作多播传输方法"", 《科技成果登记表》*

Cited By (19)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113191654A (en)*2021-05-122021-07-30余绍祥Distributed resource scheduling system and method based on multi-objective optimization
CN113660625A (en)*2021-07-062021-11-16山东师范大学 Method and system for resource coordination between multi-access edge computing nodes and CDN nodes
CN113660625B (en)*2021-07-062024-02-02山东师范大学Resource cooperation method and system for multi-access edge computing node and CDN node
CN113419867A (en)*2021-08-232021-09-21浙大城市学院Energy-saving service supply method in edge-oriented cloud collaborative computing environment
CN113419867B (en)*2021-08-232022-01-18浙大城市学院Energy-saving service supply method in edge-oriented cloud collaborative computing environment
CN113784373B (en)*2021-08-242022-11-25苏州大学 Joint optimization method and system for time delay and spectrum occupancy in cloud-edge collaborative network
CN113784373A (en)*2021-08-242021-12-10苏州大学Combined optimization method and system for time delay and frequency spectrum occupation in cloud edge cooperative network
CN113923223A (en)*2021-11-152022-01-11安徽大学User allocation method with low time cost in edge environment
CN113923223B (en)*2021-11-152024-02-06安徽大学User allocation method with low time cost in edge environment
CN114448991B (en)*2021-12-282022-10-21西安电子科技大学 A method, system, medium, device and terminal for selecting a multi-edge server
CN114448991A (en)*2021-12-282022-05-06西安电子科技大学Multi-edge server selection method, system, medium, device and terminal
CN114844900B (en)*2022-05-052022-12-13中南大学 A Collaborative Method for Edge-Cloud Resources Based on Uncertain Demand
CN114844900A (en)*2022-05-052022-08-02中南大学Edge cloud resource cooperation method based on uncertain demand
CN115695231A (en)*2022-10-312023-02-03中国联合网络通信集团有限公司 Method, device and storage medium for determining network energy efficiency
CN116506877A (en)*2023-06-262023-07-28北京航空航天大学 A Distributed Collaborative Computing Method for Mobile Crowd Sensing
CN116506877B (en)*2023-06-262023-09-26北京航空航天大学 A distributed collaborative computing method for mobile crowd sensing
CN117193873A (en)*2023-09-052023-12-08北京科技大学 A computing offloading method and device suitable for industrial control systems
CN117193873B (en)*2023-09-052024-06-25北京科技大学 A method and device for offloading calculations in industrial control systems
CN118260067A (en)*2024-02-182024-06-28北京中电飞华通信有限公司Computing resource distribution method based on simulated annealing and related equipment

Also Published As

Publication numberPublication date
CN112689303B (en)2022-07-22

Similar Documents

PublicationPublication DateTitle
CN112689303B (en)Edge cloud cooperative resource joint allocation method, system and application
Cui et al.A novel offloading scheduling method for mobile application in mobile edge computing
CN111586762B (en)Task unloading and resource allocation joint optimization method based on edge cooperation
CN111586696B (en)Resource allocation and unloading decision method based on multi-agent architecture reinforcement learning
CN110351754B (en)Industrial Internet machine equipment user data calculation unloading decision method based on Q-learning
US11831708B2 (en)Distributed computation offloading method based on computation-network collaboration in stochastic network
CN112492626A (en)Method for unloading computing task of mobile user
CN114189892A (en)Cloud-edge collaborative Internet of things system resource allocation method based on block chain and collective reinforcement learning
CN107995660A (en) Joint Task Scheduling and Resource Allocation Method Supporting D2D-Edge Server Offloading
CN112105062B (en)Mobile edge computing network energy consumption minimization strategy method under time-sensitive condition
CN111885147A (en)Dynamic resource pricing method in edge calculation
CN107295109A (en)Task unloading and power distribution joint decision method in self-organizing network cloud computing
CN105656999B (en)A kind of cooperation task immigration method of energy optimization in mobile cloud computing environment
CN109714382B (en)Multi-user multi-task migration decision method of unbalanced edge cloud MEC system
CN112822707B (en) A method for task offloading and resource allocation in MEC with limited computing resources
CN111949409B (en) A method and system for offloading computing tasks in power wireless heterogeneous networks
CN111163143B (en)Low-delay task unloading method for mobile edge calculation
CN114265631A (en) A method and device for intelligent offloading of mobile edge computing based on federated meta-learning
CN118113484B (en)Resource scheduling method, system, storage medium and vehicle
CN112988347A (en)Edge computing unloading method and system for reducing system energy consumption and cost sum
CN114449490A (en)Multi-task joint computing unloading and resource allocation method based on D2D communication
CN113747450A (en)Service deployment method and device in mobile network and electronic equipment
CN114980160B (en) A UAV-assisted terahertz communication network joint optimization method and device
Liu et al.Joint optimization of task offloading and computing resource allocation in MEC-D2D network
Yang et al.Deep reinforcement learning-based low-latency task offloading for mobile-edge computing networks

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp