Movatterモバイル変換


[0]ホーム

URL:


CN105335236A - Distributed evidence obtaining dynamic load balanced scheduling method and device - Google Patents

Distributed evidence obtaining dynamic load balanced scheduling method and device
Download PDF

Info

Publication number
CN105335236A
CN105335236ACN201510915168.XACN201510915168ACN105335236ACN 105335236 ACN105335236 ACN 105335236ACN 201510915168 ACN201510915168 ACN 201510915168ACN 105335236 ACN105335236 ACN 105335236A
Authority
CN
China
Prior art keywords
task
computing node
module
node
tasks
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510915168.XA
Other languages
Chinese (zh)
Other versions
CN105335236B (en
Inventor
陈俊珊
苏再添
吴少华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen Meiya Pico Information Co Ltd
Original Assignee
Xiamen Meiya Pico Information Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen Meiya Pico Information Co LtdfiledCriticalXiamen Meiya Pico Information Co Ltd
Priority to CN201510915168.XApriorityCriticalpatent/CN105335236B/en
Publication of CN105335236ApublicationCriticalpatent/CN105335236A/en
Application grantedgrantedCritical
Publication of CN105335236BpublicationCriticalpatent/CN105335236B/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Landscapes

Abstract

The invention relates to the field of electronic medium evidence obtaining, in particular to a distributed evidence obtaining dynamic load balanced scheduling method and device. The distributed evidence obtaining dynamic load balanced scheduling method and device can support the multi-medium high-speed parallel evidence obtaining analysis work scene. Due to the fact that a task is decomposed to a plurality of computing nodes to be processed, a large amount of time is saved, and evidence obtaining efficiency is also greatly improved. The distributed evidence obtaining dynamic load balanced scheduling method and device can reduce or avoid the problem that overall performance reduction is caused by uneven distribution of network loads, a computing node CPU, an internal storage and the like.

Description

A kind of distributed evidence obtaining dynamic load leveling dispatching method and device
Technical field
The invention belongs to electronic media evidence obtaining field, relate to a kind of distributed evidence obtaining dynamic load leveling dispatching method particularly.
Background technology
Along with the development of society, increasing data are Electronically preserved, the data be stored in computing machine and other equipment progressively become important evidence in computer network case and Evidence in Litigation, and the constantly bringing forth new ideas of infotech and memory technology, make the memory capacity of equipment also increasing, the thing followed is that the case-involving storage medium of evidence obtaining process is many, capacity large, forensics analysis task is heavy, the inferior problem of inefficiency, and the forensics analysis how realized rapidly and efficiently also just becomes the emphasis of evidence obtaining product.
Traditional forensics analysis software and hardware equipment great majority are Evidence models of unit, as publication: CN203659010U, its synchronization can only support that sole user processes a case, need very powerful computing power just can complete when information memory capacity is very large, and the time expended is also very long, along with the development of network and memory technology, traditional forensics analysis equipment can not meet the requirement such as high-speed data processing and mass data correlation analysis, not only complicated operation but also inefficiency.For solving Problems existing in unit evidence obtaining, indivedual evidence-obtaining system starts to introduce distributed network forensics pattern, task matching is coupled together by network to multiple stage machine (compute node) by distributed evidence-obtaining system exactly, the arithmetic capability of each computing node is utilized to improve overall performance, but the effect that the problem that the performance difference existed due to each computing node and dispatching algorithm exist load imbalance makes the overall performance of system cannot reach desirable, therefore, reasonably to address this problem and just must carry out load balance scheduling.
Summary of the invention
The object of the invention is as solving the problem and providing one can carry out multimedium high-speed parallel forensics analysis, evidence obtaining efficiency is high, can reduce or avoid because of offered load, computing node CPU, internal memory equal distribution is uneven and the distributed evidence obtaining dynamic load leveling dispatching method of overall performance decline problem that causes and device.
For this reason, the invention discloses a kind of distributed evidence obtaining dynamic load leveling dispatching method, comprise the steps:
A1, if computing node is a computing node set P, and to each computing node Picarry out initialization;
A2, according to computing node Piresource situation determination task window threshold values Tm;
A3, calculates current operation node Piload Wi, and to its sequence;
A4, calculates current operation node Piidleness Ridle, and it is sorted;
A5, is set to a set of tasks T, and sorts to set of tasks T according to task priority by task;
A6, takes out least-loaded and the highest computing node P of idleness from computing node set Piif the result obtained is sky, repeat steps A 3 to A6, until obtain;
A7, obtains the highest task T of priority from set of tasks Tiif the result obtained is sky, repeat steps A 4 to A7, until obtain;
A8, obtaining in steps A 7 of task Tidistribute to the computing node P obtained in steps A 6i, upgrade set of tasks T and computing node P simultaneouslyicurrent number of tasks;
A9, repeated execution of steps A2 to A8, until all tasks are all assigned.
Further, in described steps A 3, at least one situation according to the CPU of computing node, internal memory and network condition calculates load Wi.
Further, in described steps A 3, summation is weighted to the CPU of computing node, internal memory and offered load and calculates load Wi.
Further, in described steps A 3, adopt by stages dynamic state feedback mechanism to the load W calculatedifeed back, and to feeding back the load W obtainedisort.
Further, in described steps A 4, described idleness Ridlecomputing method be: establish current operation node Pinumber of tasks be Tn, then the idleness R of each computing nodeidleaccording to formulaRidle=Tm-TnTm×100%Calculate.
Further, in described steps A 5, FCFS policing algorithm is adopted to sort when the priority of task is identical.
Further, also comprise task division step, be specially: task is split by granularity.
Further, also comprise abnormality processing step, be specially:
When computing node breaks down, being distributed on this node of task is reclaimed in time and be assigned on other normal nodes and continue to perform;
The threshold values of task process time-out is set, when a task time-out does not also complete, is re-assigned to other node and goes to perform or wait for complete at this node;
When task occurs mistake when performing, the information of record and feedback error in time.
The invention also discloses a kind of distributed evidence obtaining dynamic load leveling dispatching device, comprising:
Initialization module, for setting computing node as a computing node set P, and to each computing node Picarry out initialization;
Task threshold computing module, for according to computing node Piresource situation determination task window threshold values Tm;
Load calculates order module, for calculating current operation node Piload Wi, and to its sequence;
Idleness calculates order module, for calculating current operation node Piidleness Ridle, and it is sorted;
Task priority order module, for all tasks are set to a set of tasks T, and sorts to set of tasks T according to task priority;
Optimum computing node acquisition module, for taking out least-loaded and the highest computing node P of idleness from computing node set Piif the result obtained is sky, repeats load and calculate order module, idleness calculating order module, task priority order module and optimum computing node acquisition module, until obtain;
Limit priority task acquisition module, for obtaining the highest task T of priority from set of tasks Tiif the result obtained is sky, repeats idleness and calculate order module, task priority order module, optimum computing node acquisition module and limit priority task acquisition module, until obtain;
Task allocating module, for the task T that limit priority task acquisition module is obtainedidistribute to the computing node P that optimum computing node acquisition module obtainsi, upgrade set of tasks T and computing node P simultaneouslyicurrent number of tasks;
Loop module, for repeating task threshold computing module, load calculates order module, idleness calculates order module, task priority order module, optimum computing node acquisition module, limit priority task acquisition module and task allocating module, until all tasks are all assigned.
Further, described load calculates order module, and at least one situation for the CPU according to computing node, internal memory and network condition calculates load Wi.
Further, also comprise task division module, for task being split by granularity.
Further, also comprise abnormality processing module, specifically comprise:
Fault processing module, continues to perform for being distributed on this node of task being reclaimed in time when computing node breaks down and being assigned on other normal nodes;
Timeout treatment module, for arranging the threshold values of task process time-out, when a task time-out also not completing, being re-assigned to other node and going to perform or wait for complete at this node;
Error logging and feedback module, for there is mistake when task when performing, the information of record and feedback error in time.
Advantageous Effects of the present invention:
The present invention can support the operative scenario of multimedium high-speed parallel forensics analysis, it passes through Task-decomposing to the enterprising row relax of multiple computing nodes, not only save a large amount of time, also the efficiency of evidence obtaining is substantially increased, simultaneously, can reduce or avoid because of offered load, computing node CPU, internal memory equal distribution is uneven and the overall performance decline problem that causes, takes full advantage of the arithmetic capability of multiple stage machine.The present invention can be applied in evidence obtaining industry, the operations such as forensics analysis, search, index can not only be carried out fast and effectively to mass data, and also very strong in flexible expansion, the hot-swappable computing node when not halt system, the expansion that the system that realizes is seamless.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the embodiment of the present invention;
Fig. 2 is the evidence obtaining performance comparison figure of the embodiment of the present invention.
Embodiment
Now the present invention is further described with embodiment by reference to the accompanying drawings.
As shown in Figure 1, a kind of distributed evidence obtaining dynamic load leveling dispatching method, comprises the steps:
A1, if computing node is a computing node set P, and to each computing node Picarry out initialization.
In the present embodiment, computing node is computing machine, for performing evidence obtaining task, all computing nodes participating in evidence obtaining is set to a computing node set P{P1, P2, P3pn, and to each computing node Picarry out initialization, just can receive an assignment after initialization success and perform.
A2, according to computing node Piresource situation determination task window threshold values Tm.
Concrete, at computing node Piduring initialization, according to the resource situation determination task window threshold values T such as CPU check figure and internal memory of computing nodem, as Tm=24, then represent that computing node synchronization can only receive at most 24 tasks.
A3, calculates current operation node Piload Wi, and it is sorted from low to high, take out the computing node that load is minimum.
Concrete, at least one situation according to the CPU of computing node, internal memory and network condition calculates load Wi.In the present embodiment, according to computing node Pithe resource situation such as CPU, internal memory, network calculate current load Wi, Wiinterval is [0,100], and concrete grammar is: be located at t computing node Picpu load be Cpu (Pi), offered load Net (Pi), internal memory load is Mem (Pi), the computing formula of load value: Wi=R1* Cpu (Pi)+R2* Net (Pi)+R3* Mem (Pi), wherein R1, R2, R3value set according to the kind of task, as: be set as 0.4,0.5,0.1.
Further, traditional load-balancing mechanism uses the dispatching method of static polling schemas or dynamic realtime load feedback mostly, consider the problem such as network, system resource overhead that the inaccuracy of static scheduling and Real-time Feedback bring, the present invention adopts by stages dynamic state feedback mechanism with head it off, is specially: whole non-load balanced case is divided into different intervals as (0 according to configuration file, 33], (33,66], (66,100] three intervals, work as Wibe worth from the first interval (0,33] change to second (33,66] interval value, the W of computing nodeivalue feeds back, to sort.
In addition, a load threshold V is set in systems in which, works as Wivalue enters more than the task of being assigned to computing node during V value the state of waiting in line, and avoids the extreme case occurring that load is overweight, ensures the stability of system.
A4, calculates current operation node Piidleness Ridle, and it is sorted.
Be specially: establish current operation node Pinumber of tasks be Tn, then the idleness R of each computing nodeidleaccording to formula:calculate, the R that all compute node drawidlerendezvous value sorts from high to low, takes out the computing node P that idleness is the highesti.
A5, is set to a set of tasks T{T by all tasks1, T2, T3tn, and according to task priority, set of tasks T is sorted.
Concrete, first various types of task is divided by granularity, during division, should avoid occurring large task too consuming time, can not occur splitting too thin a large amount of little task.All tasks after dividing are set to a set of tasks T{T1, T2, T3tn, and corresponding priority is configured to it, then sort to set of tasks T according to task priority, the task that priority is identical sorts according to FCFS policing algorithm.FCFS policing algorithm is prior art, and this no longer describes in detail.
A6, takes out least-loaded and the highest computing node P of idleness from computing node set Piif the result obtained is sky, repeat steps A 3 to A6, until obtain.
A7, obtains the highest task T of priority from set of tasks Tiif the result obtained is sky, repeat steps A 4 to A7, until obtain.
A8, obtaining in steps A 7 of task Tidistribute to the computing node P obtained in steps A 6i, upgrade set of tasks T and computing node P simultaneouslyicurrent number of tasks.
A9, repeated execution of steps A2 to A8, until all tasks are all assigned.
In addition, in the present embodiment, also comprise abnormality processing step, be specially:
When computing node breaks down, being distributed on this node of task is reclaimed in time and be assigned on other normal nodes and continue to perform.
The threshold values of task process time-out is set, when a task time-out does not also complete, is re-assigned to other node and goes to perform or wait for complete at this node.
When task occurs mistake when performing, the information of record and feedback error in time.
Step number in the present embodiment is for convenience of description, and the sequencing that not exclusively ride instead of walk is rapid.
Present invention also offers a kind of distributed evidence obtaining dynamic load leveling dispatching device, comprising:
Initialization module, for setting computing node as a computing node set P, and to each computing node Picarry out initialization.
Concrete, initialization module, for being set to a computing node set P{P by all computing nodes participating in evidence obtaining1, P2, P3pn, and to each computing node Picarry out initialization, just can receive an assignment after initialization success and perform.
Task threshold computing module, for according to computing node Piresource situation determination task window threshold values Tm.
Concrete, at computing node Piduring initialization, according to the resource situation determination task window threshold values T such as CPU check figure and internal memory of computing nodem.
Load calculates order module, for calculating current operation node Piload Wi, and to its sequence.
Concrete, load calculates order module according to computing node Pithe resource situation such as CPU, internal memory, network calculate current load Wi, Wiinterval is [0,100], and concrete grammar is: be located at t computing node Picpu load be Cpu (Pi), offered load Net (Pi), internal memory load is Mem (Pi), the computing formula of load value: Wi=R1* Cpu (Pi)+R2* Net (Pi)+R3* Mem (Pi), wherein R1, R2, R3value set according to the kind of task, as: be set as 0.4,0.5,0.1.
Further, load calculates order module and comprises by stages dynamic feedback module, for whole non-load balanced case is divided into different intervals according to configuration file, as (0,33], (33,66], (66,100] three intervals, work as Wibe worth from the first interval (0,33] change to second (33,66] interval value, the W of computing nodeivalue feeds back and calculates order module, to sort to load.
Idleness calculates order module, for calculating current operation node Piidleness Ridle, and it is sorted.
Be specially: establish current operation node Pinumber of tasks be Tn, then the idleness R of each computing nodeidleaccording to formula:calculate, by the R that all compute node drawidlerendezvous value sorts from high to low, takes out the computing node P that idleness is the highesti.
Task priority order module, for all tasks are set to a set of tasks T, and sorts to set of tasks T according to task priority.Concrete, task priority order module also comprises:
Task division module, for dividing by granularity various types of task, should avoid occurring large task too consuming time, can not occur splitting too thin a large amount of little task during division.
Priority configuration module, for configuring corresponding priority to all the letting alone after division.
Optimum computing node acquisition module, for taking out least-loaded and the highest computing node P of idleness from computing node set Piif the result obtained is sky, repeats load and calculate order module, idleness calculating order module, task priority order module and optimum computing node acquisition module, until obtain.
Limit priority task acquisition module, for obtaining the highest task T of priority from set of tasks Tiif the result obtained is sky, repeats idleness and calculate order module, task priority order module, optimum computing node acquisition module and limit priority task acquisition module, until obtain.
Task allocating module, for the task T that limit priority task acquisition module is obtainedidistribute to the computing node P that optimum computing node acquisition module obtainsi, upgrade set of tasks T and computing node P simultaneouslyicurrent number of tasks.
Loop module, for repeating task threshold computing module, load calculates order module, idleness calculates order module, task priority order module, optimum computing node acquisition module, limit priority task acquisition module and task allocating module, until all tasks are all assigned.
In addition, in the present embodiment, device also comprises abnormality processing module, comprising:
Fault processing module, continues to perform for being distributed on this node of task being reclaimed in time when computing node breaks down and being assigned on other normal nodes.
Timeout treatment module, for arranging the threshold values of task process time-out, when a task time-out also not completing, being re-assigned to other node and going to perform or wait for complete at this node.
Error logging and feedback module, for there is mistake when task when performing, the information of record and feedback error in time.
The distributed evidence obtaining dynamic load leveling dispatching device of the embodiment of the present invention can be contained in server, this server is communicated to connect by network and computing node, in the present embodiment, network is provided by the cross-platform TCP communication module of the high-performance realized based on BoostAsio, the IOCP network model used under its windows platform is a kind of technology being applicable to high capacity server, not only scalability is good, and execution efficiency is very high, it is the basis that whole dispatch service carries out task scheduling and exchanges data.
To Figure 2 shows that under the test data mirror image of formed objects different number computing node and conventional individual pattern are collected evidence the performance comparison figure equipped, as can be seen from the figure, the performance of conventional individual pattern evidence obtaining equipment is the poorest, and the performance of the evidence obtaining equipment that computing node number is more is better.
Each module that embodiment disclosed herein describes and algorithm steps, can realize with electronic hardware, computer software or the combination of the two.These functions perform with hardware or software mode actually, depend on application-specific and the design constraint of technical scheme.Professional and technical personnel can use distinct methods to realize described function to each specifically should being used for, but this realization should not thought and exceeds scope of the present invention.
Although specifically show in conjunction with preferred embodiment and describe the present invention; but those skilled in the art should be understood that; not departing from the spirit and scope of the present invention that appended claims limits; can make a variety of changes the present invention in the form and details, be protection scope of the present invention.

Claims (12)

CN201510915168.XA2015-12-102015-12-10A kind of distributed dynamic load leveling dispatching method and device of collecting evidenceActiveCN105335236B (en)

Priority Applications (1)

Application NumberPriority DateFiling DateTitle
CN201510915168.XACN105335236B (en)2015-12-102015-12-10A kind of distributed dynamic load leveling dispatching method and device of collecting evidence

Applications Claiming Priority (1)

Application NumberPriority DateFiling DateTitle
CN201510915168.XACN105335236B (en)2015-12-102015-12-10A kind of distributed dynamic load leveling dispatching method and device of collecting evidence

Publications (2)

Publication NumberPublication Date
CN105335236Atrue CN105335236A (en)2016-02-17
CN105335236B CN105335236B (en)2019-03-19

Family

ID=55285791

Family Applications (1)

Application NumberTitlePriority DateFiling Date
CN201510915168.XAActiveCN105335236B (en)2015-12-102015-12-10A kind of distributed dynamic load leveling dispatching method and device of collecting evidence

Country Status (1)

CountryLink
CN (1)CN105335236B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107026907A (en)*2017-03-302017-08-08上海斐讯数据通信技术有限公司A kind of load-balancing method, load equalizer and SiteServer LBS
CN109814997A (en)*2019-01-182019-05-28创新奇智(广州)科技有限公司A kind of distributed freedom equilibrium artificial intelligence method for scheduling task and system
CN112711467A (en)*2020-12-172021-04-27北京科银京成技术有限公司Partition timeout processing method and device, computer equipment and storage medium
CN115002127A (en)*2022-06-092022-09-02方图智能(深圳)科技集团股份有限公司Distributed audio system

Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6438576B1 (en)*1999-03-292002-08-20International Business Machines CorporationMethod and apparatus of a collaborative proxy system for distributed deployment of object rendering
CN101562479A (en)*2009-05-192009-10-21中国联合网络通信集团有限公司Background service transmission method and system based on idle channel
CN103049323A (en)*2012-12-312013-04-17西安奇维科技股份有限公司Multi-interrupt balance management method implemented in FPGA (field programmable gate array)
CN103338228A (en)*2013-05-302013-10-02江苏大学Cloud calculating load balancing scheduling algorithm based on double-weighted least-connection algorithm
CN104239148A (en)*2013-06-062014-12-24腾讯科技(深圳)有限公司Distributed task scheduling method and device
CN104866373A (en)*2015-05-202015-08-26南京国电南自电网自动化有限公司Real-time operating system simulation method based on cross-platform technology

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US6438576B1 (en)*1999-03-292002-08-20International Business Machines CorporationMethod and apparatus of a collaborative proxy system for distributed deployment of object rendering
CN101562479A (en)*2009-05-192009-10-21中国联合网络通信集团有限公司Background service transmission method and system based on idle channel
CN103049323A (en)*2012-12-312013-04-17西安奇维科技股份有限公司Multi-interrupt balance management method implemented in FPGA (field programmable gate array)
CN103338228A (en)*2013-05-302013-10-02江苏大学Cloud calculating load balancing scheduling algorithm based on double-weighted least-connection algorithm
CN104239148A (en)*2013-06-062014-12-24腾讯科技(深圳)有限公司Distributed task scheduling method and device
CN104866373A (en)*2015-05-202015-08-26南京国电南自电网自动化有限公司Real-time operating system simulation method based on cross-platform technology

Cited By (6)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
CN107026907A (en)*2017-03-302017-08-08上海斐讯数据通信技术有限公司A kind of load-balancing method, load equalizer and SiteServer LBS
CN107026907B (en)*2017-03-302020-08-14广东红餐科技有限公司Load balancing method, load balancer and load balancing system
CN109814997A (en)*2019-01-182019-05-28创新奇智(广州)科技有限公司A kind of distributed freedom equilibrium artificial intelligence method for scheduling task and system
CN112711467A (en)*2020-12-172021-04-27北京科银京成技术有限公司Partition timeout processing method and device, computer equipment and storage medium
CN112711467B (en)*2020-12-172023-10-10北京科银京成技术有限公司Partition timeout processing method, partition timeout processing device, computer equipment and storage medium
CN115002127A (en)*2022-06-092022-09-02方图智能(深圳)科技集团股份有限公司Distributed audio system

Also Published As

Publication numberPublication date
CN105335236B (en)2019-03-19

Similar Documents

PublicationPublication DateTitle
CN100555230C (en)Method for providing processor cluster for system with multiple processors
CN105335236A (en)Distributed evidence obtaining dynamic load balanced scheduling method and device
CN103401947A (en)Method and device for allocating tasks to multiple servers
CN109032769B (en)Container-based continuous integrated CI (CI) task processing method and device
CN104718548A (en) Efficient pushdown of joins in heterogeneous database systems involving large-scale low-power clusters
CN110868330B (en) Evaluation method, device and evaluation system for dividing CPU resources of cloud platform
CN104793996A (en)Task scheduling method and device of parallel computing equipment
CN108122055A (en)The resource regulating method and device of a kind of Flow Shop
CN106339802A (en)Task allocation method, task allocation device and electronic equipment
CN111158904A (en)Task scheduling method, device, server and medium
CN116302574B (en)Concurrent processing method based on MapReduce
CN107273339A (en)A kind of task processing method and device
CN105045670A (en)Method and system for balancing loads of central processing units and graphic processing units
CN105227616B (en)A kind of method of the dynamic creation of remote sensing satellite Ground Processing System task and distribution
CN112148475A (en)Task scheduling method and system of loongson big data all-in-one machine integrating load and power consumption
CN104281636A (en)Concurrent distributed processing method for mass report data
CN112052087B (en)Deep learning training system and method for dynamic resource adjustment and migration
CN112468414B (en)Cloud computing multi-level scheduling method, system and storage medium
CN113656502A (en)Data synchronization method, system, electronic device and storage medium
CN111506409B (en)Data processing method and system
CN106293670B (en)Event processing method and device and server
CN112799852A (en)Multi-dimensional SBP distributed signature decision system and method for logic node
CN105808340A (en)Load balancing method and system
CN118260080A (en)Server load balancing method and device, server cluster, equipment and medium
CN103139295A (en)Cloud computing resource dispatch method and device

Legal Events

DateCodeTitleDescription
C06Publication
PB01Publication
C10Entry into substantive examination
SE01Entry into force of request for substantive examination
GR01Patent grant
GR01Patent grant

[8]ページ先頭

©2009-2025 Movatter.jp