Movatterモバイル変換


[0]ホーム

URL:


CN116647604B - A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios - Google Patents

A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios

Info

Publication number
CN116647604B
CN116647604BCN202310670378.1ACN202310670378ACN116647604BCN 116647604 BCN116647604 BCN 116647604BCN 202310670378 ACN202310670378 ACN 202310670378ACN 116647604 BCN116647604 BCN 116647604B
Authority
CN
China
Prior art keywords
server
service request
edge
algorithm unit
service
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310670378.1A
Other languages
Chinese (zh)
Other versions
CN116647604A (en
Inventor
芮兰兰
高志鹏
李菡
刘茂华
陈子轩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and TelecommunicationsfiledCriticalBeijing University of Posts and Telecommunications
Priority to CN202310670378.1ApriorityCriticalpatent/CN116647604B/en
Publication of CN116647604ApublicationCriticalpatent/CN116647604A/en
Application grantedgrantedCritical
Publication of CN116647604BpublicationCriticalpatent/CN116647604B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Classifications

Landscapes

Abstract

Translated fromChinese

本发明公开了一种边端协作场景中适应动态环境的算力资源调度方法,在动态边端协作场景中,进行业务请求量预测,综合考虑多维度资源对业务请求、边缘服务器和算法单元进行评估,并进行业务请求和边缘服务器的匹配,使用深度强化学习算法解决算法单元调度问题,使其适应动态变化的环境,同时对时延、边缘层贡献度、边缘服务器资源利用均衡程度进行了优化,优化时延,提高边缘层的贡献度,使边缘服务器的资源利用率更加均衡。通过仿真实验表明,本发明提出的方法,可以在一定程度上降低时延、提高边缘层贡献度,同时提高边缘服务器资源利用均衡程度。

The present invention discloses a computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios. In dynamic edge-end collaboration scenarios, the method predicts the amount of business requests, comprehensively considers multi-dimensional resources to evaluate business requests, edge servers, and algorithm units, and matches business requests with edge servers. A deep reinforcement learning algorithm is used to solve the algorithm unit scheduling problem, making it adaptable to dynamically changing environments. At the same time, the method optimizes latency, edge layer contribution, and edge server resource utilization balance, optimizes latency, improves edge layer contribution, and makes edge server resource utilization more balanced. Simulation experiments show that the method proposed in the present invention can reduce latency, improve edge layer contribution, and improve edge server resource utilization balance to a certain extent.

Description

Computing power resource scheduling method adapting to dynamic environment in side-end cooperation scene
Technical Field
The invention relates to the technical field of resource scheduling in a side-end cooperation scene, in particular to a computing power resource scheduling method adapting to a dynamic environment in a side-end cooperation scene.
Background
With the development of intelligent devices and the internet of things, more and more applications, such as virtual reality/augmented reality technology, face recognition, intelligent driving and the like, are rapidly generated. These applications have a high demand for computational power. In traditional cloud computing, a large amount of applications and data are transmitted to a cloud data center for processing, and the cloud data center has strong computing power and storage capacity, so that the requirements of the applications can be met. At the same time, however, the terminal device is often far away from the cloud center, which may generate a long transmission delay. For some delay-sensitive services, cloud computing is difficult to meet its low-latency requirements.
Edge computing has been rapidly developed and widely used in recent years. Unlike traditional cloud computing, edge computing deploys high-performance servers near users, sinking computing power to network edges, enhancing computing power at network edges, and deploying computation-intensive, delay-sensitive services to edge servers for processing.
Edge Intelligence (EI) is an integration of Mobile Edge Computing (MEC) and artificial intelligence technology, and has recently become a promising example to support network edge computing intensive artificial intelligence applications. In the applications of the virtual reality/augmented reality technology, face recognition, intelligent driving and the like, artificial intelligent algorithms are widely applied, for example, the face recognition technology needs to use algorithms such as CNN and the like to provide services for the algorithms. In the edge intelligent technology, an artificial intelligent algorithm can be deployed on an edge server, training and reasoning processes are completed on the edge server, and more efficient services are provided for face recognition and other services. Compared with the traditional cloud center, the edge server has a short transmission delay and a short distance from the user, but the resources of the edge server are limited.
In order to reasonably utilize resources, in the existing scheduling method, CN114490018A (prior art 1) proposes a service scheduling algorithm based on resource feature matching, and a mechanism of resource feature matching is introduced into the service scheduling algorithm to solve the problems that the MQoS requirements of users cannot be met and resources cannot be fully utilized in a service scheduling scene. By introducing the concept of resource feature matching, the execution efficiency is improved, the energy consumption is reduced, and the dynamic adjustment of the optimization direction is supported.
CN113821317A (prior art 2) proposes a micro-service scheduling method, device and equipment for edge cloud cooperation, which uses service completion time and resource utilization rate of an edge server as optimization targets by decoupling an application service program into a plurality of micro-services and based on an improved particle swarm optimization algorithm, and finally obtains a micro-service scheduling strategy for realizing multi-target optimization through continuous iteration, on one hand, service scheduling enables service execution to be close to a data source as much as possible, thereby meeting the low-delay requirement of a user, and on the other hand, micro-services of partial non-delay sensitive application are scheduled to a cloud server, so that the calculation pressure of the edge server is well relieved, the problem of overload of the edge server is avoided, and the service quality is influenced.
CN113391647a (prior art 3) proposes a method and a system for deploying and scheduling edge computing services of multiple unmanned aerial vehicles, which relate to the technical field of unmanned aerial vehicle communication and service deployment. The method comprises the following steps of service dependence and topology-aware micro-service deployment, greedy micro-service scheduling, load-aware micro-service redeployment and failure-triggered service redeployment. The method has the advantages of high service response speed, capability of initializing micro service deployment according to the popular condition of the micro services, the dependency relationship among the micro services and the network topology, greedily carrying out micro service scheduling, effective reduction of application response delay, good environmental adaptability, capability of sensing queuing length and idle state of micro service operation on the edge computing nodes, capability of dynamically adjusting service function deployment when the service is congested or resources are idle, better adaptation to changing execution environments, and flexible failure adjustment, wherein when the unmanned aerial vehicle edge nodes fail, service redeployment can be carried out timely, and service continuity is ensured.
In the existing scheduling method, the scheme 1 introduces a mechanism of resource feature matching into a service scheduling algorithm, but the considered resource features are not comprehensive enough, and other resources may not be matched or reasonably utilized. The optimization targets considered in the service scheduling in the technical scheme 2 and the technical scheme 3 are time delay, resource utilization rate and the like, and the optimization targets are not comprehensive enough in the complex environment in the side-end cooperation scene. According to the technical scheme 2, the particle swarm optimization algorithm is adopted to conduct micro-service scheduling, the traditional heuristic algorithm generally gives an optimal solution based on the determined problem, and the problem solving efficiency is low in a dynamic environment. Therefore, the invention formulates a set of algorithm unit scheduling method, adapts to dynamic environment and reasonably utilizes resources.
Disclosure of Invention
Aiming at the problems of incomplete resource characteristics and optimization targets considered in the prior art in a dynamic side-end cooperation scene, the invention provides a computational power resource scheduling method adapting to a dynamic environment in the side-end cooperation scene.
In order to achieve the above object, the present invention provides the following technical solutions:
The invention provides a computing power resource scheduling method adapting to dynamic environment in a side-end cooperation scene, which comprises the following steps:
s1, constructing resource assessment models of service requests, edge servers and algorithm units, wherein each resource assessment model is modeled from four dimensions of calculation, communication, storage and memory;
s2, predicting based on the historical service request quantity to obtain future service request quantity, and taking a prediction result as a basis of algorithm unit scheduling;
S3, evaluating the service request, the edge server and the algorithm unit by using resource indexes of four dimensions of calculation, communication, storage and memory, and matching the service request and the edge server according to an evaluation result;
s4, comprehensively predicting the obtained service request quantity, the matching result of the service request and the edge server, and the type of an algorithm unit required by the service request, and formulating a multi-objective optimization problem model by considering time delay, edge layer contribution degree and edge server resource utilization balance degree;
S5, carrying out algorithm unit scheduling based on a deep reinforcement learning algorithm.
Further, in step S1, the service request is modeled as a quad requesti=(tci,tbi,tsi,tmi),0≤i<Mt, where elements in the quad respectively represent a calculation requirement, a communication requirement, a storage requirement, and a memory requirement of the service, and Mt represents the number of service requests in the current time slot t;
Modeling the edge server as a quadruple Ei=(Ci,Bi,Si,Mi), wherein i < N is more than or equal to 0, and elements in the quadruple respectively represent computing capacity, communication capacity, storage capacity and memory capacity of the server, and N represents the number of the servers;
the algorithm unit is modeled as a quadruple rk=(ck,bk,sk,mk), k is more than or equal to 0 and less than or equal to W, wherein elements in the quadruple respectively represent calculation resources, communication resources, storage resources and memory resources which are needed to be occupied by the algorithm unit, and W represents the number of the algorithm unit.
Further, in step S2, the method for predicting based on the historical service request amount is that the service request amount of the t time slot is set as X (t), the time sequence sliding window size d is adopted, and the service request amount of the past d time slots is used for predicting the service request amount of the future 1 time window by adopting a GRU deep learning algorithm.
Further, in step S3, the service request and the edge server are evaluated by using the resource indexes of four dimensions of calculation, communication, storage and memory, and the service request and the edge server data are normalized, which comprises the following steps:
Sequence of computing demands for a serviceNormalizing to obtain normalized resultSimilarly, the communication requirement, the storage requirement and the memory requirement of the service request are normalized, and the calculation, the communication, the storage and the memory resources owned by the edge server are normalized, so that the following results are obtained:
the normalized result of the resource requirement of the request i is expressed as
Representing the normalized result of the resources owned by the server i as
And then, giving different weights Q= (a, b, c, d) to the four indexes to obtain comprehensive evaluation results TRi and ERi, wherein the following formula is shown:
TRi=GTi·QT
ERi=GEi·QT
Wherein TRi is a comprehensive evaluation result of the service request i, ERi is a comprehensive evaluation result of the edge server i, GEi is a normalization result of resources owned by the edge server i, GTi is a normalization result of resource requirements of the service request i, and QT is a weight of four resource indexes.
Further, in step S3, the service request and the edge server are classified by a Rank method, and are matched by a Map method according to the classification;
the service request and the edge server are classified by a Rank method, and the method comprises the following steps:
1) Finding the maximum value and the minimum value in the comprehensive evaluation result sequence, and calculating the difference value;
2) Defining a number of levels rankcount to be divided;
3) Dividing the difference between the maximum value and the minimum value in the comprehensive evaluation result by the number of grades to obtain a grade spacing gap, namely, the difference between the maximum value and the minimum value in each grade range is gap;
4) The evaluation result corresponds to the grade interval to obtain a grading result;
The matching method by Map method according to the grade is as follows:
1) Selecting servers with the same grade for the service according to the grade of the service;
2) If the grade of a certain service is r, selecting a server with the grade of r, if the number of the servers with the grade of r is more than 1, selecting a server with the least service number from the servers to be matched with the service, and if the grade of a certain service is r but no server with the grade of r, selecting a server with the closest grade;
3) The final matching result matrix A is obtained and expressed as follows:
Where ai denotes that service i matches server ai.
Further, in step S4, the delay model construction process is as follows:
The service request needs to be transmitted to the server for processing, and is divided into two cases:
If the matched server j has an algorithm unit rk required by the service requesti, namely btj,k =1, the server directly responds to the service request;
if the matched server j does not have the algorithm unit rk required by the service requesti, namely btj,k =0, transmitting the service request to the central cloud;
In case one, the delay from the terminal device to the server isLet the case-time delay beIn the second case, if the response cannot be obtained and needs to be transmitted to the central cloud, the time delay of the second case is tauvc, the time delay of the service i is defined asIs expressed by the following formula:
the average delay is defined as:
Where M represents the number of service requests.
Further, in step S4, the edge layer contribution model construction process is as follows:
By usingIndicating whether the service request i can be responded at the edge layer, setting the server to which the service request i is matched as j, the algorithm unit required by the service request i as k,Is expressed by the following formula:
If the matched server j has an algorithm unit rk required by the service requesti, namely, btj,k =1, the server directly responds to the service request, and if the matched server j has no algorithm unit rk required by the service requesti, namely, btj,k =0, the service request is transmitted to the central cloud;
The contribution degree of the edge layer is expressed as Ht, defined as the ratio of the number of the services responded in the edge layer in the time slot t to the total number of the services, and expressed by the following formula:
Where Mt represents the number of service requests for the current slot t.
Further, in step S4, the edge server resource utilization balance degree model is constructed as follows:
let rcj(yj,zk) denote the percentage of the remaining resources of server j under a certain index, expressed as:
Wherein y epsilon { C, B, S, M } represents the current resource quantity of the server under a certain index, and z epsilon { C, B, S, M } represents the resources which the algorithm unit needs to occupy under a certain index;
The RCt={rc1,rc2,...,rcN is used for representing the current comprehensive resource quantity of the server after the algorithm unit is deployed, and the current resource quantity of the server is represented as follows:
The resource utilization balance deltat is defined as 1 minus the standard deviation of the rcj sequence, and represents the balance of the current resources of each server, and the resource utilization balance of the edge server is expressed by the following formula:
Wherein, theRepresenting the mean of the RCt sequence.
Further, the multi-objective optimization problem model of step S4 is constructed as follows:
Wherein, theThe average time delay is represented, Ht represents the contribution degree of the edge layer, deltat represents the resource utilization balance degree of the edge server, and alpha, beta and gamma represent weight coefficients of the time delay, the contribution degree of the edge layer and the resource utilization balance degree respectively.
Further, in step S5, the algorithm unit is deployed on the server by the scheduling method based on deep reinforcement learning, and the result is represented as a matrix Bt:
Wherein bij is a binary variable, if bij is 1, it indicates that algorithm unit j is present on server i, and if bij is 0, it indicates that algorithm unit j is not present on server i.
Compared with the prior art, the invention has the beneficial effects that:
According to the computing power resource scheduling method suitable for the dynamic environment in the side-end cooperation scene, in the dynamic side-end cooperation scene, the service request quantity is predicted, the service request, the edge server and the algorithm unit are evaluated by comprehensively considering multidimensional resources, the service request and the edge server are matched, the problem of scheduling the algorithm unit is solved by using a deep reinforcement learning algorithm, the algorithm unit is suitable for the dynamically changing environment, meanwhile, the time delay, the contribution degree of the edge layer and the resource utilization balance degree of the edge server are optimized, the time delay is optimized, the contribution degree of the edge layer is improved, and the resource utilization rate of the edge server is more balanced. Simulation experiments show that the method provided by the invention can reduce time delay to a certain extent, improve the contribution degree of the edge layer and improve the resource utilization balance degree of the edge server.
Drawings
Fig. 1 is a flowchart of a computing power resource scheduling method adapting to a dynamic environment in a side-end collaboration scenario provided by an embodiment of the present invention.
Fig. 2 is a schematic diagram of a matching method between a service request and an edge server according to an embodiment of the present invention.
Fig. 3 is a delay comparison chart provided in an embodiment of the present invention.
Fig. 4 is a comparison chart of contribution degrees of edge layers according to an embodiment of the present invention.
Fig. 5 is a comparison chart of resource utilization balance degrees according to an embodiment of the present invention.
Detailed Description
The resources of the edge server are limited, a reasonable algorithm unit scheduling method is formulated, and the limited resources are fully utilized to provide service for the service request. The service request in the environment is dynamically changed, and the scheduling scheme of the algorithm unit is prepared by taking the dynamic characteristics of the environment into consideration, and simultaneously taking the service quality of the user and the resource utilization condition of the edge server into consideration. Therefore, the invention formulates the computing power resource scheduling method adapting to the dynamic environment in the side-end cooperation scene.
In order to better understand the technical solution, the technical solution in the embodiment of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiment of the present application. It will be apparent that the described examples are only some embodiments, but not all embodiments, of the present application. Based on the embodiments of the present application, all other embodiments obtained by the person skilled in the art based on the present application are included in the scope of protection of the present application.
The computing power resource scheduling method adapting to the dynamic environment in the side-end cooperation scene, as shown in figure 1, comprises the following steps:
S1, constructing resource assessment models of service requests, edge servers and algorithm units, wherein each resource assessment model is modeled from four dimensions of calculation, communication, storage and memory.
Specifically, the service request is modeled as a quad requesti=(tci,tbi,tsi,tmi),0≤i<Mt, where the elements in the quad represent the computation requirement, communication requirement, storage requirement, and memory requirement of the service, and Mt represents the number of service requests in the current slot t.
Modeling the edge server as a quadruple Ei=(Ci,Bi,Si,Mi), wherein i < N is more than or equal to 0, and the elements in the quadruple respectively represent the computing capacity, communication capacity, storage capacity and memory capacity of the server, and N represents the number of the servers.
When the algorithm unit is operated, certain calculation, communication, storage and memory resources are occupied, and the resources occupied by the algorithm unit with the type of rk are represented by four-element rk=(ck,bk,sk,mk), wherein k is more than or equal to 0 and less than or equal to W. Wherein, elements in the quadruple respectively represent calculation resources, communication resources, storage resources and memory resources which are needed to be occupied by the algorithm units, and W represents the number of the algorithm units.
The service request quantity has a certain time characteristic, so that a service prediction model is firstly trained according to the historical data characteristics of the service and is used for predicting the future service request quantity. And adopting a GRU deep learning algorithm to actively predict the service request quantity, and taking a prediction result as a basis of algorithm unit scheduling. Let the traffic request of the t-th time slot be X (t), the time series sliding window size d, and use the traffic request of the past d time slots to predict the traffic request of the future 1 time window. The method comprises the steps of predicting by adopting GRU, in a training stage, a historical observation sample set is X= { X (t-d), (t-d+1),. The use of { X (t-d) }, the use of X (t-1) } as input of a GRU network model, outputting a prediction result X '(t), and adjusting model parameters by comparing X' (t) and X (t), so as to perform model training.
S2, predicting based on the historical service request quantity to obtain the future service request quantity, and taking a prediction result as a basis of algorithm unit scheduling.
The method for predicting based on the historical service request quantity is to set the service request quantity of the t-th time slot as X (t), the time sequence sliding window size d, and the service request quantity of the past d time slots is used for predicting the service request quantity of the future 1 time window by adopting a GRU deep learning algorithm.
The invention formulates a matching method of the service and the server taking the resource as a main consideration factor. Firstly, comprehensively evaluating the service requirements and the server resources to obtain a comprehensive evaluation result. And then classifying the service and the server according to the evaluation result. After grading, the service and the server are matched according to the grade. If the method is not adopted, the service with large resource requirement can be matched with the server with poor performance, so that the service requirement can not be met. By adopting the method, the resources of the server can be more reasonably utilized, and the matching result can also provide effective reference for the subsequent algorithm unit scheduling work.
S3, evaluating the service request, the edge server and the algorithm unit by using resource indexes of four dimensions of calculation, communication, storage and memory, and matching the service request and the edge server according to an evaluation result.
Specifically, the service request and the edge server are evaluated, and the service request and the edge server data are normalized, wherein the process is as follows:
computing demand sequence for all servicesNormalizing to obtain normalized resultSimilarly, the communication requirement, the storage requirement and the memory requirement of the service request are normalized, and the calculation, the communication, the storage and the memory resources owned by the edge server are normalized, so that the following results are obtained:
the normalized result of the resource requirement of the request i is expressed as
Representing the normalized result of the resources owned by the server i as
And then, giving different weights Q= (a, b, c, d) to the four indexes to obtain comprehensive evaluation results TRi and ERi, wherein the following formula is shown:
TRi=GTi·QT
ERi=GEi·QT
Wherein TRi is a comprehensive evaluation result of the service request i, ERi is a comprehensive evaluation result of the edge server i, GEi is a normalization result of resources owned by the edge server i, GTi is a normalization result of resource requirements of the service request i, and QT is a weight of four resource indexes.
Then, the service request and the server are classified by a Rank method, are matched by a Map method, and a matching result is defined as a matrix A.
The service request and the edge server are classified by a Rank method, and the method comprises the following steps:
The service and server are classified according to the comprehensive evaluation results TRi and ERi:
Rank(TRi)=RTRi
Rank(ERi)=RERi
for the hierarchical Rank, the following is defined:
1) Finding the maximum value and the minimum value in the comprehensive evaluation result sequence, and calculating the difference value;
2) Defining the number of levels rankcount to be divided, for example, to be divided into 5 levels, rankcount =5;
3) Dividing the difference between the maximum value and the minimum value in the comprehensive evaluation result by the number of grades to obtain a grade spacing gap, namely, the difference between the maximum value and the minimum value in each grade range is gap;
4) And (5) corresponding the evaluation result to the grade interval to obtain a grading result.
After grading, the service and the server are matched according to the grades by a Map method, and the method comprises the following steps:
1) Selecting servers with the same grade for the service according to the grade of the service;
2) If the grade of a certain service is r, selecting a server with the grade of r, if the number of the servers with the grade of r is more than 1, selecting a server with the least service number from the servers to be matched with the service, and if the grade of a certain service is r but no server with the grade of r, selecting a server with the closest grade;
3) The final matching result matrix A is obtained and expressed as follows:
Where ai denotes that service i matches server ai.
Different services require different algorithm units to serve them, and the mapping relation NS (requesti)=rk indicates that the algorithm unit required for service i is rk.
As shown in fig. 2. Firstly, classifying service requests and servers through a comprehensive evaluation method and a classification method to obtain classification results of the service requests and the servers, and then matching the service requests and the servers through a matching method Map according to the classification results.
S4, comprehensively predicting the obtained service request quantity, the matching result of the service request and the edge server, and the type of an algorithm unit required by the service request, and formulating a multi-objective optimization problem model by considering time delay, edge layer contribution degree and edge server resource utilization balance degree.
Specifically, the time delay model construction process is as follows:
The service request needs to be transmitted to the server for processing, and is divided into two cases:
If the matched server j has an algorithm unit rk required by the service requesti, namely btj,k =1, the server directly responds to the service request;
if the matched server j does not have the algorithm unit rk required by the service requesti, namely btj,k =0, transmitting the service request to the central cloud;
In case one, the delay from the terminal device to the server isLet the case-time delay beIn the second case, if the response cannot be obtained and needs to be transmitted to the central cloud, the time delay of the second case is tauvc, the time delay of the service i is defined asIs expressed by the following formula:
the average delay is defined as:
Where M represents the number of service requests.
The edge layer contribution degree model construction process is as follows:
One of the optimization goals of the scheme is to make the service request respond to the matched edge server to the maximum extent so as to maximize the effect of the edge layer. Thus, the edge layer contribution is set. And matching the service with the server by using the matching method, wherein if the server has an algorithm unit required by the service request, the service request can be processed, the service can be responded at the edge layer, otherwise, the service cannot be responded at the edge layer. The edge layer contribution is denoted as Ht and is defined as the ratio of the number of traffic in the edge layer response to the total number of traffic in the time slot t.
By usingIndicating whether the service request i can be responded at the edge layer, setting the server to which the service request i is matched as j, the algorithm unit required by the service request i as k,Is expressed by the following formula:
If the matched server j has an algorithm unit rk required by the service requesti, namely, btj,k =1, the server directly responds to the service request, and if the matched server j has no algorithm unit rk required by the service requesti, namely, btj,k =0, the service request is transmitted to the central cloud;
The contribution degree of the edge layer is expressed as Ht, defined as the ratio of the number of the services responded in the edge layer in the time slot t to the total number of the services, and expressed by the following formula:
Where Mt represents the number of service requests for the current slot t.
The edge server resource utilization balance degree model is constructed as follows:
The algorithm unit needs to occupy the resource space of the server. In order to keep the resource utilization conditions of the servers as balanced as possible, the invention defines the resource utilization balance degree to represent the difference degree of the resource utilization conditions of each server.
Let rcj(yj,zk) denote the percentage of the remaining resources of server j under a certain index, expressed as:
Wherein y epsilon { C, B, S, M } represents the current resource amount of the server under a certain index, z epsilon { C, B, S, M } represents the resources which the algorithm unit needs to occupy under a certain index, and rcj(Cj,ck) represents the current residual resource percentage of the server j under the calculation index.
The RCt={rc1,rc2,...,rcN is used for representing the current comprehensive resource quantity of the server after the algorithm unit is deployed, and the current resource quantity of the server is expressed as follows because four indexes have large difference in magnitude order:
The constraint is that none of the four sub-terms in the following formula is less than zero.
The resource utilization balance deltat is defined as 1 minus the standard deviation of the rcj sequence, and represents the balance of the current resources of each server, and the resource utilization balance of the edge server is expressed by the following formula:
Wherein, theThe larger deltat, the more uniform the resource utilization, which represents the mean of the RCt sequence.
From the service perspective, they need to get a response as soon as possible to get a response at the edge server, so the delay is an important evaluation index, and the invention expects a smaller delay. From the perspective of the edge server, the function exerted by the edge layer should be maximized, so that as many service requests as possible are responded at the edge layer, and meanwhile, the resource utilization of each server should be kept balanced as much as possible when the algorithm unit is scheduled. Thus, the present invention formulates a total optimization objective function G as follows.
The multi-objective optimization problem model (comprehensive optimization objective) is constructed as follows:
Wherein, theThe average time delay is represented, Ht represents the contribution degree of the edge layer, deltat represents the resource utilization balance degree of the edge server, and alpha, beta and gamma represent weight coefficients of the time delay, the contribution degree of the edge layer and the resource utilization balance degree respectively.
S5, carrying out algorithm unit scheduling based on a deep reinforcement learning algorithm.
In the deep reinforcement learning algorithm, an agent interacts with the environment to learn, and a scheduling decision is made based on the real-time state of the environment, so that the method can be well adapted to the dynamically-changing environment, and the scheduling problem is solved by adopting the deep reinforcement learning algorithm. In addition, the deployment scene of the invention comprises a plurality of servers and a plurality of algorithm units, the state space and the action space are high in dimensionality, and the high-dimensional problem is difficult to solve by the Q learning method in reinforcement learning, so that the DQN method combining reinforcement learning and deep learning is adopted. The method interacts with the environment, and realizes the dynamic scheduling of the algorithm unit according to the dynamically changed environment information. In addition, in order to prevent the algorithm unit from occupying server resources for a long time, a life cycle is set for the algorithm unit, and the deployed algorithm unit is automatically deleted and corresponding resource space is released after reaching a certain life cycle Tend. The method can dynamically schedule the algorithm units, reduce the time delay to the maximum extent, improve the service response rate and improve the utilization balance of the server resources.
Specifically, the invention deploys the algorithm unit on the server through the scheduling method based on deep reinforcement learning, and the result is represented as a matrix Bt:
Wherein bij is a binary variable, if bij is 1, it indicates that algorithm unit j is present on server i, and if bij is 0, it indicates that algorithm unit j is not present on server i.
The algorithm model training process is as follows:
firstly, defining a state space, an action space and a reward function in an algorithm unit scheduling environment, wherein a Markov model is defined as follows:
The state is that the current algorithm unit on the server has state (t) = [ s0,s1,...,sp,...,sN*W-1],sp epsilon {0,1}, if sp = 1, let p/N = i, p% N = j, wherein N is the number of the servers and W is the number of the algorithm units, namely, the element number p divided by the number of the servers is equal to i by the whole, the remainder is j, and the like, which indicates that the algorithm unit j exists on the server i;
The method comprises the actions that an algorithm unit scheduling decision a (t) = [ ac0,ac1,...,acW-1],0≤aci < N, wherein aci represents that an algorithm unit i is deployed on a server aci, if the algorithm unit exists, the deployment is not repeated, and if the residual resources of the server can not meet the resource requirements of the algorithm unit, the deployment is not performed;
the rewarding function is set as a comprehensive optimization objective function;
R(t)=Gt (16)
Then defining an algorithm unit existence time matrix:
wherein N is the number of servers, W is the number of algorithm units, di.j represents the existing time of the algorithm unit j on the server i, and is used for judging whether the algorithm unit reaches the life cycle limit or not and whether the algorithm unit needs to be deleted or not.
The algorithm unit scheduling algorithm flow based on the deep reinforcement learning is shown in table 1.
Table 1 algorithm unit scheduling algorithm flow based on deep reinforcement learning
And predicting based on the historical service request amount, randomly generating the resource amount of the edge server, the service request and the resource requirement of the algorithm unit, matching the service request and the edge server, and training a scheduling algorithm based on deep reinforcement learning. And setting the weights of the time delay, the edge layer contribution degree and the edge server resource utilization balance degree in the comprehensive optimization objective function to be alpha=1, beta=5 and gamma=1 respectively. And performing algorithm unit scheduling by using a trained algorithm unit scheduling method based on deep reinforcement learning, and comparing the algorithm unit scheduling with a random scheduling method and a greedy scheduling method only optimizing time delay. The comparison results under the three indexes are shown in fig. 3, 4 and 5 respectively. Fig. 3 shows the time delay comparison result, wherein the time delay obtained by the greedy scheduling method only optimizing the time delay is optimal, the scheduling method based on deep reinforcement learning is inferior, and the random scheduling method is worst. Fig. 4 shows the comparison result of the edge layer contribution, wherein the edge layer contribution obtained by the greedy scheduling method only optimizing the time delay is optimal, the scheduling method based on deep reinforcement learning is inferior, and the random scheduling method is worst. Fig. 5 shows the comparison of the degree of resource utilization balance, wherein the time delay obtained by the scheduling method based on deep reinforcement learning is optimal, the random scheduling method is inferior, and the greedy scheduling method only optimizing the time delay is worst. The greedy scheduling method only optimizing the time delay only considers the time delay, and meanwhile, the greedy scheduling method only optimizing the time delay under the two indexes performs optimally because the contribution degree of the edge layer and the time delay are closely related, but the greedy scheduling method ignores other indexes, so that the resource utilization balance degree is the worst. The scheduling method based on deep reinforcement learning comprehensively considers three indexes, and compared with a random scheduling method, the scheduling method based on deep reinforcement learning is optimized to a certain extent. Simulation experiments show that the method provided by the invention can reduce time delay to a certain extent, improve the contribution degree of the edge layer and improve the resource utilization balance degree of the edge server.
The foregoing is merely illustrative of the preferred embodiments and principles of the present invention, and not in limitation thereof. Any modification, equivalent replacement, improvement, etc. which are within the spirit and principle of the present invention, should be considered as the protection scope of the present invention, based on the ideas provided by the present invention, for those skilled in the art.

Claims (10)

CN202310670378.1A2023-06-072023-06-07 A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenariosActiveCN116647604B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN202310670378.1ACN116647604B (en)2023-06-072023-06-07 A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN202310670378.1ACN116647604B (en)2023-06-072023-06-07 A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios

Publications (2)

Publication NumberPublication Date
CN116647604A CN116647604A (en)2023-08-25
CN116647604Btrue CN116647604B (en)2025-08-26

Family

ID=87622844

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN202310670378.1AActiveCN116647604B (en)2023-06-072023-06-07 A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios

Country Status (1)

CountryLink
CN (1)CN116647604B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN117992230B (en)*2024-02-212024-10-25北京驭达科技有限公司Vehicle-mounted edge calculation method and system based on autonomous learning
CN118939412B (en)*2024-07-112025-07-25中国司法大数据研究院有限公司Cloud edge collaborative optimization balanced deployment method with delay assurance
CN119376955B (en)*2024-12-272025-04-18中科边缘智慧信息科技(苏州)有限公司Calculation power resource evaluation method and system for maneuvering edge application scene
CN120256150B (en)*2025-06-062025-08-05清华海峡研究院(厦门)Computer resource allocation method and device for differentiated teaching scene

Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2022027776A1 (en)*2020-08-032022-02-10威胜信息技术股份有限公司Edge computing network task scheduling and resource allocation method and edge computing system
CN116137724A (en)*2023-02-202023-05-19重庆邮电大学 A task offloading and resource allocation method based on mobile edge computing

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN113993218A (en)*2021-11-182022-01-28国网福建省电力有限公司经济技术研究院 A collaborative offloading and resource allocation method based on multi-agent DRL under MEC architecture

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2022027776A1 (en)*2020-08-032022-02-10威胜信息技术股份有限公司Edge computing network task scheduling and resource allocation method and edge computing system
CN116137724A (en)*2023-02-202023-05-19重庆邮电大学 A task offloading and resource allocation method based on mobile edge computing

Also Published As

Publication numberPublication date
CN116647604A (en)2023-08-25

Similar Documents

PublicationPublication DateTitle
CN116647604B (en) A computing resource scheduling method that adapts to dynamic environments in edge-end collaboration scenarios
CN112882815B (en)Multi-user edge calculation optimization scheduling method based on deep reinforcement learning
Tak et al.Federated edge learning: Design issues and challenges
Dong et al.Quantum particle swarm optimization for task offloading in mobile edge computing
CN108920280B (en)Mobile edge computing task unloading method under single-user scene
WO2022257348A1 (en)Task offloading and resource allocation method in uncertain network environment
CN114219074B (en) A resource allocation algorithm for wireless communication networks with dynamic adjustment on demand
CN109840154B (en) A task-dependent computing migration method in mobile cloud environment
CN107172166A (en)The cloud and mist computing system serviced towards industrial intelligentization
CN114741955B (en) A multi-objective optimization task scheduling method based on secure cloud
CN116541106B (en) Calculation task offloading method, computing device and storage medium
CN115967990B (en) A method for offloading edge collaborative services based on classification and prediction
Liang et al.Multiple service model refreshments in digital twin-empowered edge computing
Chai et al.A dynamic queuing model based distributed task offloading algorithm using deep reinforcement learning in mobile edge computing
CN117042184A (en)Calculation unloading and resource allocation method based on deep reinforcement learning
Chen et al.Multi-agent deep reinforcement learning for collaborative task offloading in mobile edge computing networks
Zhang et al.Offloading approach for mobile edge computing based on chaotic quantum particle swarm optimization strategy
Zanbouri et al.A new fog-based transmission scheduler on the Internet of multimedia things using a fuzzy-based quantum genetic algorithm
Venkatesan et al.Enhancing resource management in fog-cloud internet of things systems with deep learning-based task allocation
CN119829243A (en)Calculation power scheduling method and system applied to large-scale network simulation system
CN119583597A (en) An industrial Internet of Things edge computing offloading method and system based on distributed deep neural network
CN114077482A (en) An edge intelligent computing optimization method for industrial intelligent manufacturing
Liu et al.Task scheduling model of edge computing for AI flow computing in Internet of Things
CN114997400A (en)Neural network acceleration reasoning method
CN113157344A (en)DRL-based energy consumption perception task unloading method in mobile edge computing environment

Legal Events

DateCodeTitleDescription
PB01Publication
PB01Publication
SE01Entry into force of request for substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp