[go: up one dir, main page]

CN114090253A - Method, apparatus, electronic device and readable medium for allocating network computing resources - Google Patents

Method, apparatus, electronic device and readable medium for allocating network computing resources Download PDF

Info

Publication number
CN114090253A
CN114090253A CN202111396820.3A CN202111396820A CN114090253A CN 114090253 A CN114090253 A CN 114090253A CN 202111396820 A CN202111396820 A CN 202111396820A CN 114090253 A CN114090253 A CN 114090253A
Authority
CN
China
Prior art keywords
requirements
computing
network computing
sub
computing resources
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111396820.3A
Other languages
Chinese (zh)
Inventor
唐静
雷波
解云鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Telecom Corp Ltd
Original Assignee
China Telecom Corp Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Telecom Corp Ltd filed Critical China Telecom Corp Ltd
Priority to CN202111396820.3A priority Critical patent/CN114090253A/en
Publication of CN114090253A publication Critical patent/CN114090253A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5005Allocation of resources, e.g. of the central processing unit [CPU] to service a request
    • G06F9/5027Allocation of resources, e.g. of the central processing unit [CPU] to service a request the resource being a machine, e.g. CPUs, Servers, Terminals
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5061Partitioning or combining of resources
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2209/00Indexing scheme relating to G06F9/00
    • G06F2209/50Indexing scheme relating to G06F9/50
    • G06F2209/5011Pool

Landscapes

  • Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

本公开提供一种网络算力资源的分配方法、装置、电子设备和可读介质,其中,网络算力资源的分配方法包括:接收终端发送的业务需求;解析业务需求中的子需求;根据子需求将网络算力资源划分为多个算力资源池;将子需求分配至对应的算力资源池进行计算。通过本公开实施例,可以提高网络算力资源的利用率,也提升了网络算力资源的可靠性和运算能力。

Figure 202111396820

The present disclosure provides a method, device, electronic device and readable medium for allocating network computing resources, wherein the method for allocating network computing resources includes: receiving a service requirement sent by a terminal; analyzing sub-requirements in the service requirement; Demand divides network computing resources into multiple computing resource pools; assigns sub-demands to corresponding computing resource pools for calculation. Through the embodiments of the present disclosure, the utilization rate of network computing resources can be improved, and the reliability and computing capability of network computing resources can also be improved.

Figure 202111396820

Description

Network computing resource allocation method and device, electronic equipment and readable medium
Technical Field
The present disclosure relates to the field of network computing technologies, and in particular, to a method and an apparatus for allocating network computing power resources, an electronic device, and a readable medium.
Background
At present, edge computing is a computing scheme at the edge side of a network close to an object or a data source, a distributed open platform with core capabilities of network, computing, storage and application is fused, edge intelligent service is provided nearby, and key requirements of industry digitization on aspects of agile connection, real-time service, data optimization, application intelligence, safety, privacy protection and the like are met.
In the related art, edge computing is applied to a 5G network, that is, a wireless access network is utilized to provide services and cloud computing functions required by IT of a telecommunication user nearby, so that a telecommunication-level service environment with high performance, low delay and high bandwidth is created, rapid downloading of various contents, services and applications in the network is accelerated, and consumers can enjoy uninterrupted high-quality network experience.
However, in the edge computing or computational network, the computing power of the nodes that can provide computational resources is not uniform, and particularly, in the logic computation with a small computation amount, if the GPU resources that can bear high-parallel computation are used, the computational resources are insufficient when high-computation traffic arrives.
It is to be noted that the information disclosed in the above background section is only for enhancement of understanding of the background of the present disclosure, and thus may include information that does not constitute prior art known to those of ordinary skill in the art.
Disclosure of Invention
The present disclosure is directed to a method, an apparatus, an electronic device, and a readable medium for allocating network computing power resources, which overcome, at least to some extent, the problems of unreasonable allocation of network computing power resources due to the limitations and disadvantages of the related art.
According to a first aspect of the embodiments of the present disclosure, there is provided a method for allocating network computing resources, including: receiving a service requirement sent by a terminal; analyzing the sub-requirements in the service requirements; dividing the network computing resources into a plurality of computing resource pools according to the sub-demands; and distributing the sub-demands to corresponding calculation resource pools for calculation.
In an exemplary embodiment of the present disclosure, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-demand includes: and if the sub-requirements are determined to be storage requirements and/or control requirements, dividing the network computing resources to obtain a logic computing resource pool.
In an exemplary embodiment of the present disclosure, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-demand further includes: and if the sub-requirements are determined to be at least one of image processing requirements, computing requirements, password cracking requirements, numerical analysis requirements, data processing requirements and financial analysis requirements, dividing the network computing resources into parallel computing resource pools.
In an exemplary embodiment of the present disclosure, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-demand further includes: and if the sub-requirements are determined to be the neural network computing requirements and/or the machine learning computing requirements, dividing the network computing resources to obtain a neural network computing resource pool.
In an exemplary embodiment of the present disclosure, resolving sub-requirements in the business requirements includes: and performing calculation analysis on the service requirement through calculation requirement, wherein an expression of the calculation analysis comprises: cbr ═ α x Σ Ai + β x Σ Bj + γ x Σ Ck + q, said Cbr characterizing said computation power requirements, said α, said β, and said γ being preset weights, said Ai characterizing logical computation requirements, said Bj characterizing parallel computation requirements, said Ck characterizing neural network computation requirements, and said q characterizing redundant computation power.
In an exemplary embodiment of the present disclosure, the business requirements include at least one of hyper-computational business requirements, AI inference training business requirements, target detection business requirements, and voice business requirements.
In an exemplary embodiment of the present disclosure, the computing resources in the computing resource pool include at least one of a CPU, a GPU, and an AI chip.
According to a second aspect of the embodiments of the present disclosure, there is provided an apparatus for allocating network computing resources, including: the receiving module is set to receive the service requirement sent by the terminal; the analysis module is arranged for analyzing the sub-requirements in the service requirements; the dividing module is used for dividing the network computing power resource into a plurality of computing power resource pools according to the sub-demands; and the computing module is configured to allocate the sub-demands to the corresponding computing power resource pools for computing.
According to a third aspect of the present disclosure, there is provided an electronic device comprising: a memory; and a processor coupled to the memory, the processor configured to perform the method of any of the above based on instructions stored in the memory.
According to a fourth aspect of the present disclosure, there is provided a computer-readable storage medium, on which a program is stored, which when executed by a processor, implements the method for allocating network computing power resources as described in any one of the above.
According to the embodiment of the disclosure, the sub-requirements in the service requirements are analyzed, the network computing resources are divided into the computing resource pools according to the sub-requirements, and then the sub-requirements are allocated to the corresponding computing resource pools for computing, so that the allocation mode of the network computing resources is optimized, and the utilization rate and the reliability of the network computing resources are improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and together with the description, serve to explain the principles of the disclosure. It is to be understood that the drawings in the following description are merely exemplary of the disclosure, and that other drawings may be derived from those drawings by one of ordinary skill in the art without the exercise of inventive faculty.
FIG. 1 is a flow chart of a method for allocation of network computing resources in an exemplary embodiment of the disclosure;
FIG. 2 is a flow chart of another method for allocating network computing resources in an exemplary embodiment of the present disclosure;
FIG. 3 is a flow chart of another method for allocating network computing resources in an exemplary embodiment of the present disclosure;
FIG. 4 is a flow chart of another method for allocation of network computing resources in an exemplary embodiment of the disclosure;
FIG. 5 is a flow chart of another method for allocation of network computing resources in an exemplary embodiment of the present disclosure;
FIG. 6 is a flow chart of another method for allocation of network computing resources in an exemplary embodiment of the present disclosure;
FIG. 7 is an architecture diagram of an allocation scheme for network computing resources in an exemplary embodiment of the present disclosure;
FIG. 8 is a block diagram of an apparatus for allocating network computing resources in an exemplary embodiment of the present disclosure;
fig. 9 is a block diagram of an electronic device in an exemplary embodiment of the disclosure.
Detailed Description
Example embodiments will now be described more fully with reference to the accompanying drawings. Example embodiments may, however, be embodied in many different forms and should not be construed as limited to the examples set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the concept of example embodiments to those skilled in the art. The described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments. In the following description, numerous specific details are provided to give a thorough understanding of embodiments of the disclosure. One skilled in the relevant art will recognize, however, that the subject matter of the present disclosure can be practiced without one or more of the specific details, or with other methods, components, devices, steps, and the like. In other instances, well-known technical solutions have not been shown or described in detail to avoid obscuring aspects of the present disclosure.
Further, the drawings are merely schematic illustrations of the present disclosure, in which the same reference numerals denote the same or similar parts, and thus, a repetitive description thereof will be omitted. Some of the block diagrams shown in the figures are functional entities and do not necessarily correspond to physically or logically separate entities. These functional entities may be implemented in the form of software, or in one or more hardware modules or integrated circuits, or in different networks and/or processor devices and/or microcontroller devices.
The following describes exemplary embodiments of the present disclosure in detail with reference to fig. 1 to 9.
Fig. 1 is a flowchart of a method for allocating network computing resources in an exemplary embodiment of the present disclosure.
Referring to fig. 1, the method for allocating network computing resources may include:
and step S102, receiving the service requirement sent by the terminal.
And step S104, analyzing the sub-requirements in the service requirements.
And step S106, dividing the network computing resources into a plurality of computing resource pools according to the sub-demands.
And S108, distributing the sub-demands to corresponding computing power resource pools for computing.
According to the embodiment of the disclosure, the sub-requirements in the service requirements are analyzed, the network computing resources are divided into the computing resource pools according to the sub-requirements, and then the sub-requirements are distributed to the corresponding computing resource pools for computing, so that the distribution mode of the network computing resources is optimized, and the utilization rate and the reliability of the network computing resources are improved.
Next, each step of the network computing power resource allocation method will be described in detail.
In an exemplary embodiment of the present disclosure, as shown in fig. 2, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-requirements includes:
step S202, if the sub-requirements are determined to be storage requirements and/or control requirements, dividing the network computing resources to obtain a logic computing resource pool.
In an exemplary embodiment of the present disclosure, as shown in fig. 3, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-requirements further includes:
step S302, if the sub-requirements are determined to be at least one of image processing requirements, computing requirements, password cracking requirements, numerical analysis requirements, data processing requirements and financial analysis requirements, dividing the network computing resources into parallel computing resource pools.
In an exemplary embodiment of the present disclosure, as shown in fig. 4, dividing the network computing power resource into a plurality of computing power resource pools according to the sub-requirements further includes:
step S402, if the sub-requirements are determined to be neural network computing requirements and/or machine learning computing requirements, dividing the network computing resources into neural network computing resource pools.
In an exemplary embodiment of the present disclosure, as shown in fig. 5, resolving sub-requirements in the service requirement includes:
step S502, performing calculation force analysis on the service requirement according to calculation force requirements, wherein an expression of the calculation force analysis comprises:
Cbr=α×∑Ai+β×∑Bj+γ×∑Ck+q,
the Cbr represents the computing power requirement, the alpha, the beta and the gamma are preset weights, the Ai represents a logic computing requirement, the Bj represents a parallel computing requirement, the Ck represents a neural network computing requirement, and the q represents a redundant computing power.
In an exemplary embodiment of the present disclosure, the business requirements include at least one of hyper-computational business requirements, AI inference training business requirements, target detection business requirements, and voice business requirements.
In an exemplary embodiment of the present disclosure, the computing resources in the computing resource pool include at least one of a CPU, a GPU, and an AI chip.
In an exemplary embodiment of the present disclosure, as shown in fig. 6, the step of allocating the network computing power resource includes:
step S602: in terms of service input, in the existing services at present, various services have different demands on computing power, such as a super-computation service, an AI inference training service, an inference type service for target detection, and the like, and speech semantic translation.
Step S604: the computing power network global orchestrator receives the service request.
Step S606: analyzing the calculation power requirement of the service requirement through a calculation power requirement analysis module:
computing power demand Cbr=α·∑Ai+β·∑Bj+γ·∑Ck+q。
Wherein Ai represents a logic operation requirement, Bj represents a parallel computing capability requirement, Ck represents a neural network computing requirement, alpha, beta and gamma are preset proportionality coefficients, and q is a redundancy computing power.
Step S608: and distributing the calculation power resources by the sub-modules of the orchestrator according to the output result of the calculation power requirement analysis module, namely distributing the requirement of the business on the calculation power capability to each sub-module for orchestration.
Step S610: and the calculation force node completes the calculation.
In an exemplary embodiment of the present disclosure, as shown in fig. 7, an architecture for allocating network computing power resources includes: a plurality of edge node devices, e.g., a first edge node device 702, a second edge node device 704, a third edge node device 706, and so on. In addition, the network computing power resource allocation architecture further comprises: a plurality of resource pools, such as a first resource pool 708, a second resource pool 710, and a third resource pool 712, and the like.
First, a user traffic demand is received.
Secondly, a calculation power requirement analysis module in the calculation power network global service orchestrator analyzes the user service requirements.
And thirdly, performing resource allocation on the analysis result through the parallel calculation arrangement submodule, the logic operation arrangement submodule and the neural network calculation arrangement submodule.
In an exemplary embodiment of the present disclosure, the computation power requirements corresponding to the parallel computation orchestration submodule are allocated to the first edge node device 702 and the second edge node device 704 through the routing nodes, the computation power requirements corresponding to the logic computation orchestration submodule are allocated to the third edge node device 706 through the routing nodes, and the computation power requirements corresponding to the neural network computation orchestration submodule are allocated to the first resource pool 708, the second resource pool 710, the third resource pool 712, and the like through the routing nodes.
Corresponding to the method embodiment, the disclosure further provides a device for allocating network computing power resources, which can be used to execute the method embodiment.
Fig. 8 is a block diagram of an apparatus for allocating network computing resources in an exemplary embodiment of the present disclosure.
Referring to fig. 8, the apparatus 800 for allocating network computing resources may include:
the receiving module 802 is configured to receive a service requirement sent by a terminal.
The parsing module 804 is configured to parse the sub-requirements in the service requirements.
A partitioning module 806 configured to partition the network computing power resource into a plurality of computing power resource pools according to the sub-requirements.
A calculation module 808 configured to allocate the sub-requirements to corresponding computational resource pools for calculation.
In an exemplary embodiment of the disclosure, the dividing module 806 is further configured to: and if the sub-requirements are determined to be storage requirements and/or control requirements, dividing the network computing resources to obtain a logic computing resource pool.
In an exemplary embodiment of the disclosure, the dividing module 806 is further configured to: and if the sub-requirements are determined to be at least one of image processing requirements, computing requirements, password cracking requirements, numerical analysis requirements, data processing requirements and financial analysis requirements, dividing the network computing resources into parallel computing resource pools.
In an exemplary embodiment of the disclosure, the dividing module 806 is further configured to: and if the sub-requirements are determined to be neural network computing requirements and/or machine learning computing requirements, dividing the network computing power resources to obtain a neural network computing resource pool.
In an exemplary embodiment of the disclosure, the parsing module 804 is further configured to: and performing calculation analysis on the service requirement through calculation requirement, wherein an expression of the calculation analysis comprises: cbr ═ α x Σ Ai + β x Σ Bj + γ x Σ Ck + q, said Cbr characterizing said computation power requirements, said α, said β, and said γ being preset weights, said Ai characterizing logical computation requirements, said Bj characterizing parallel computation requirements, said Ck characterizing neural network computation requirements, and said q characterizing redundant computation power.
In an exemplary embodiment of the present disclosure, the business requirements include at least one of hyper-computational business requirements, AI inference training business requirements, target detection business requirements, and voice business requirements.
In an exemplary embodiment of the present disclosure, the computing resources in the computing resource pool include at least one of a CPU, a GPU, and an AI chip.
Since the functions of the network computing resource allocation apparatus 800 have been described in detail in the corresponding method embodiments, the disclosure is not repeated herein.
It should be noted that although in the above detailed description several modules or units of the device for action execution are mentioned, such a division is not mandatory. Indeed, the features and functionality of two or more modules or units described above may be embodied in one module or unit, according to embodiments of the present disclosure. Conversely, the features and functions of one module or unit described above may be further divided into embodiments by a plurality of modules or units.
In an exemplary embodiment of the present disclosure, an electronic device capable of implementing the above method is also provided.
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or program product. Thus, various aspects of the invention may be embodied in the form of: an entirely hardware embodiment, an entirely software embodiment (including firmware, microcode, etc.) or an embodiment combining hardware and software aspects that may all generally be referred to herein as a "circuit," module "or" system.
An electronic device 900 according to this embodiment of the invention is described below with reference to fig. 9. The electronic device 900 shown in fig. 9 is only an example and should not bring any limitations to the function and scope of use of the embodiments of the present invention.
As shown in fig. 9, the electronic device 900 is embodied in the form of a general purpose computing device. Components of electronic device 900 may include, but are not limited to: the at least one processing unit 910, the at least one memory unit 920, and a bus 930 that couples various system components including the memory unit 920 and the processing unit 910.
Wherein the storage unit stores program code that is executable by the processing unit 910 to cause the processing unit 910 to perform steps according to various exemplary embodiments of the present invention described in the above section "exemplary methods" of the present specification. For example, the processing unit 910 may perform a method as shown in the embodiments of the present disclosure.
The storage unit 920 may include a readable medium in the form of a volatile storage unit, such as a random access memory unit (RAM)9201 and/or a cache memory unit 9202, and may further include a read only memory unit (ROM) 9203.
Storage unit 920 may also include a program/utility 9204 having a set (at least one) of program modules 9205, such program modules 9205 including but not limited to: an operating system, one or more application programs, other program modules, and program data, each of which, or some combination thereof, may comprise an implementation of a network environment.
Bus 930 can be any of several types of bus structures including a memory unit bus or memory unit controller, a peripheral bus, an accelerated graphics port, a processing unit, or a local bus using any of a variety of bus architectures.
The electronic device 900 may also communicate with one or more external devices 940 (e.g., keyboard, pointing device, bluetooth device, etc.), one or more devices that enable a user to interact with the electronic device 900, and/or any device (e.g., router, modem, etc.) that enables the electronic device 900 to communicate with one or more other computing devices. Such communication may occur via input/output (I/O) interface 950. Also, the electronic device 900 may communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN) and/or a public network, such as the Internet) via the network adapter 960. As shown, the network adapter 960 communicates with the other modules of the electronic device 900 via the bus 930. It should be appreciated that although not shown, other hardware and/or software modules may be used in conjunction with the electronic device 900, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
Through the above description of the embodiments, those skilled in the art will readily understand that the exemplary embodiments described herein may be implemented by software, or by software in combination with necessary hardware. Therefore, the technical solution according to the embodiments of the present disclosure may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a usb disk, a removable hard disk, etc.) or on a network, and includes several instructions to enable a computing device (which may be a personal computer, a server, a terminal device, or a network device, etc.) to execute the method according to the embodiments of the present disclosure.
In an exemplary embodiment of the present disclosure, there is also provided a computer-readable storage medium having stored thereon a program product capable of implementing the above-described method of the present specification. In some possible embodiments, aspects of the invention may also be implemented in the form of a program product comprising program code means for causing a terminal device to carry out the steps according to various exemplary embodiments of the invention described in the above section "exemplary methods" of the present description, when said program product is run on the terminal device.
The program product for implementing the above method according to an embodiment of the present invention may employ a portable compact disc read only memory (CD-ROM) and include program codes, and may be run on a terminal device, such as a personal computer. However, the program product of the present invention is not limited in this regard and, in the present document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. The readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable disk, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
A computer readable signal medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable signal medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device and partly on a remote computing device, or entirely on the remote computing device or server. In the case of a remote computing device, the remote computing device may be connected to the user computing device through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computing device (e.g., through the internet using an internet service provider).
Furthermore, the above-described figures are merely schematic illustrations of processes involved in methods according to exemplary embodiments of the invention, and are not intended to be limiting. It will be readily understood that the processes shown in the above figures are not intended to indicate or limit the chronological order of the processes. In addition, it is also readily understood that these processes may be performed synchronously or asynchronously, e.g., in multiple modules.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.

Claims (10)

1.一种网络算力资源的分配方法,其特征在于,包括:1. A method for distributing network computing power resources, comprising: 接收终端发送的业务需求;Receive the service requirements sent by the terminal; 解析所述业务需求中的子需求;Parse the sub-requirements in the business requirement; 根据所述子需求将所述网络算力资源划分为多个算力资源池;dividing the network computing resources into a plurality of computing resource pools according to the sub-requirements; 将所述子需求分配至对应的算力资源池进行计算。Allocate the sub-requirements to the corresponding computing resource pool for calculation. 2.如权利要求1所述的网络算力资源的分配方法,其特征在于,根据所述子需求将所述网络算力资源划分为多个算力资源池包括:2. The method for allocating network computing resources according to claim 1, wherein dividing the network computing resources into multiple computing resource pools according to the sub-requirements comprises: 若确定所述子需求为存储需求和/或控制需求,则将所述网络算力资源划分得到逻辑运算资源池。If it is determined that the sub-requirements are storage requirements and/or control requirements, the network computing resources are divided to obtain a logical computing resource pool. 3.如权利要求1所述的网络算力资源的分配方法,其特征在于,根据所述子需求将所述网络算力资源划分为多个算力资源池还包括:3. The method for allocating network computing resources according to claim 1, wherein dividing the network computing resources into multiple computing resource pools according to the sub-requirements further comprises: 若确定所述子需求为图像处理需求、计算需求、密码破解需求、数值分析需求、数据处理需求和金融分析需求中的至少一种,则将所述网络算力资源划分得到并行计算资源池。If it is determined that the sub-requirements are at least one of image processing requirements, computing requirements, password cracking requirements, numerical analysis requirements, data processing requirements and financial analysis requirements, the network computing resources are divided to obtain a parallel computing resource pool. 4.如权利要求1所述的网络算力资源的分配方法,其特征在于,根据所述子需求将所述网络算力资源划分为多个算力资源池还包括:4. The method for allocating network computing resources according to claim 1, wherein dividing the network computing resources into multiple computing resource pools according to the sub-requirements further comprises: 若确定所述子需求为神经网络计算需求和/或机器学习计算需求,则将所述网络算力资源划分得到神经网络计算资源池。If it is determined that the sub-requirements are neural network computing requirements and/or machine learning computing requirements, the network computing resources are divided to obtain a neural network computing resource pool. 5.如权利要求1-4中任一项所述的网络算力资源的分配方法,其特征在于,解析所述业务需求中的子需求包括:5. The method for allocating network computing power resources according to any one of claims 1-4, wherein parsing the sub-requirements in the business requirements comprises: 通过算力需求对所述业务需求进行算力解析,所述算力解析的表达式包括:The computing power analysis is performed on the business requirements according to the computing power requirements, and the expressions of the computing power analysis include: Cbr=α×∑Ai+β×∑Bj+γ×∑Ck+q,Cbr=α×∑Ai+β×∑Bj+γ×∑Ck+q, 所述Cbr表征所述算力需求,所述α、所述β和所述γ为预设权重,所述Ai表征逻辑运算需求,所述Bj表征并行计算需求,所述Ck表征神经网络计算需求,所述q表征冗余算力。The Cbr represents the computing power requirement, the α, the β and the γ are preset weights, the Ai represents the logical operation requirement, the Bj represents the parallel computing requirement, and the Ck represents the neural network computing requirement , the q represents redundant computing power. 6.如权利要求1-4中任一项所述的网络算力资源的分配方法,其特征在于,所述业务需求包括超算类业务需求、AI推理训练业务需求、目标检测业务需求和语音业务需求中的至少一种。6. The method for allocating network computing resources according to any one of claims 1-4, wherein the business requirements include supercomputing business requirements, AI inference training business requirements, target detection business requirements and voice At least one of the business requirements. 7.如权利要求1-4中任一项所述的网络算力资源的分配方法,其特征在于,所述算力资源池中的算力资源包括CPU、GPU和AI芯片中的至少一种。7. The method for allocating network computing resources according to any one of claims 1-4, wherein the computing resources in the computing resource pool include at least one of a CPU, a GPU and an AI chip . 8.一种网络算力资源的分配装置,其特征在于,包括:8. A device for distributing network computing power resources, comprising: 接收模块,设置为接收终端发送的业务需求;The receiving module is set to receive the service requirements sent by the terminal; 解析模块,设置为解析所述业务需求中的子需求;A parsing module, configured to parse the sub-requirements in the business requirement; 划分模块,设置为根据所述子需求将所述网络算力资源划分为多个算力资源池;a dividing module, configured to divide the network computing resources into multiple computing resource pools according to the sub-requirements; 计算模块,设置为将所述子需求分配至对应的算力资源池进行计算。The computing module is configured to allocate the sub-requirements to the corresponding computing resource pool for computing. 9.一种电子设备,其特征在于,包括:9. An electronic device, characterized in that, comprising: 存储器;以及memory; and 耦合到所述存储器的处理器,所述处理器被配置为基于存储在所述存储器中的指令,执行如权利要求1-7中任一项所述的网络算力资源的分配方法。A processor coupled to the memory, the processor configured to perform the method for allocating network computing resources of any one of claims 1-7 based on instructions stored in the memory. 10.一种计算机可读存储介质,其上存储有程序,该程序被处理器执行时实现如权利要求1-7中任一项所述的网络算力资源的分配方法。10. A computer-readable storage medium on which a program is stored, and when the program is executed by a processor, implements the method for allocating network computing resources according to any one of claims 1-7.
CN202111396820.3A 2021-11-23 2021-11-23 Method, apparatus, electronic device and readable medium for allocating network computing resources Pending CN114090253A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111396820.3A CN114090253A (en) 2021-11-23 2021-11-23 Method, apparatus, electronic device and readable medium for allocating network computing resources

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111396820.3A CN114090253A (en) 2021-11-23 2021-11-23 Method, apparatus, electronic device and readable medium for allocating network computing resources

Publications (1)

Publication Number Publication Date
CN114090253A true CN114090253A (en) 2022-02-25

Family

ID=80303396

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111396820.3A Pending CN114090253A (en) 2021-11-23 2021-11-23 Method, apparatus, electronic device and readable medium for allocating network computing resources

Country Status (1)

Country Link
CN (1) CN114090253A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115550370A (en) * 2022-12-01 2022-12-30 浩鲸云计算科技股份有限公司 Computing power resource optimal scheduling allocation method based on multi-factor strategy
CN116402318A (en) * 2023-06-07 2023-07-07 北京智芯微电子科技有限公司 Multi-stage computing power resource distribution method and device for power distribution network and network architecture
CN116501501A (en) * 2023-06-21 2023-07-28 亚信科技(中国)有限公司 Method and device for managing and arranging computing power resources, electronic equipment and storage medium
CN118802400A (en) * 2023-12-26 2024-10-18 中移(杭州)信息技术有限公司 Multi-device command perception system, method and device, self-organizing network method and device

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110301967A1 (en) * 2010-06-08 2011-12-08 International Business Machines Corporation Probabilistic optimization of resource discovery, reservation and assignment
US20120303816A1 (en) * 2011-05-27 2012-11-29 International Business Machines Corporation Fair discount for network resource allocation
CN103858118A (en) * 2011-08-08 2014-06-11 国际商业机器公司 Dynamically acquiring computing resources in a networked computing environment
CN109358965A (en) * 2018-09-25 2019-02-19 杭州朗和科技有限公司 Cloud computing cluster resource dispatching method, medium, device and calculating equipment
CN111679905A (en) * 2020-05-11 2020-09-18 天津大学 Computational network fusion network model
CN111796937A (en) * 2020-06-29 2020-10-20 平安国际智慧城市科技股份有限公司 Resource allocation method based on memory, computer equipment and storage medium
CN112527509A (en) * 2020-12-21 2021-03-19 北京百度网讯科技有限公司 Resource allocation method and device, electronic equipment and storage medium
CN113037786A (en) * 2019-12-09 2021-06-25 中国电信股份有限公司 Intelligent computing power scheduling method, device and system
CN113315700A (en) * 2020-02-26 2021-08-27 中国电信股份有限公司 Computing resource scheduling method, device and storage medium
CN113419846A (en) * 2021-02-22 2021-09-21 阿里巴巴集团控股有限公司 Resource allocation method and device, electronic equipment and computer readable storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110301967A1 (en) * 2010-06-08 2011-12-08 International Business Machines Corporation Probabilistic optimization of resource discovery, reservation and assignment
US20120303816A1 (en) * 2011-05-27 2012-11-29 International Business Machines Corporation Fair discount for network resource allocation
CN103858118A (en) * 2011-08-08 2014-06-11 国际商业机器公司 Dynamically acquiring computing resources in a networked computing environment
CN109358965A (en) * 2018-09-25 2019-02-19 杭州朗和科技有限公司 Cloud computing cluster resource dispatching method, medium, device and calculating equipment
CN113037786A (en) * 2019-12-09 2021-06-25 中国电信股份有限公司 Intelligent computing power scheduling method, device and system
CN113315700A (en) * 2020-02-26 2021-08-27 中国电信股份有限公司 Computing resource scheduling method, device and storage medium
CN111679905A (en) * 2020-05-11 2020-09-18 天津大学 Computational network fusion network model
CN111796937A (en) * 2020-06-29 2020-10-20 平安国际智慧城市科技股份有限公司 Resource allocation method based on memory, computer equipment and storage medium
CN112527509A (en) * 2020-12-21 2021-03-19 北京百度网讯科技有限公司 Resource allocation method and device, electronic equipment and storage medium
CN113419846A (en) * 2021-02-22 2021-09-21 阿里巴巴集团控股有限公司 Resource allocation method and device, electronic equipment and computer readable storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KAZUKI MOCHIZUKI: ""Evaluation of Optimal Resource Allocation Method for Cloud Computing Environments with Limited Electric Power Capacity"", 《2011 14TH INTERNATIONAL CONFERENCE ON NETWORK-BASED INFORMATION SYSTEMS》, 13 November 2011 (2011-11-13) *
赵倩颖: ""一种基于域名解析机制的算力网络实现方案"", 《电信科学》, 20 October 2021 (2021-10-20) *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115550370A (en) * 2022-12-01 2022-12-30 浩鲸云计算科技股份有限公司 Computing power resource optimal scheduling allocation method based on multi-factor strategy
CN116402318A (en) * 2023-06-07 2023-07-07 北京智芯微电子科技有限公司 Multi-stage computing power resource distribution method and device for power distribution network and network architecture
CN116402318B (en) * 2023-06-07 2023-12-01 北京智芯微电子科技有限公司 Multi-stage computing power resource distribution method and device for power distribution network and network architecture
CN116501501A (en) * 2023-06-21 2023-07-28 亚信科技(中国)有限公司 Method and device for managing and arranging computing power resources, electronic equipment and storage medium
CN118802400A (en) * 2023-12-26 2024-10-18 中移(杭州)信息技术有限公司 Multi-device command perception system, method and device, self-organizing network method and device

Similar Documents

Publication Publication Date Title
CN114090253A (en) Method, apparatus, electronic device and readable medium for allocating network computing resources
US20200082316A1 (en) Cognitive handling of workload requests
US20210064666A1 (en) Natural language interaction with automated machine learning systems
CN112333623B (en) Spatial-based audio object generation using image information
US20210004672A1 (en) Unary relation extraction using distant supervision
US20170090992A1 (en) Dynamic transparent provisioning for application specific cloud services
US10305756B2 (en) Allocating operations of a streaming application to virtual machines based on monitored performance
US9762466B2 (en) Predictive cloud provisioning based on human behaviors and heuristics
US11294685B2 (en) Instruction fusion using dependence analysis
WO2022063614A1 (en) Accuracy of streaming rnn transducer
US11303712B1 (en) Service management in distributed system
US20200218989A1 (en) Conversational system management
US20200244745A1 (en) Gateway device allowing multiple infrastructural services to access multiple iot devices
US20230177385A1 (en) Federated machine learning based on partially secured spatio-temporal data
US20180139327A1 (en) Enhancing customer service processing using data analytics and cognitive computing
JP2024505334A (en) Structural self-awareness model for discourse analysis in multiparty dialogue
US10635492B2 (en) Leveraging shared work to enhance job performance across analytics platforms
US20150256483A1 (en) Allocating operators of a streaming application to virtual machines based on monitored performance
US10042867B2 (en) Generating storage plans in storage management systems
US9577841B2 (en) System and method for packet encapsulation wherein translation control entries (TCEs) may redirect DMA for software defined networks
US9736231B1 (en) Representational state transfer request routing
US11005920B2 (en) Synchronized release of resources used in deferential cloud services
US12373268B2 (en) Hybrid computing system management
US20210160375A1 (en) Call Transfer Support System
US20230342397A1 (en) Techniques for predicting a personalized url document to assist a conversation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20220225

RJ01 Rejection of invention patent application after publication