Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

GPU (Graphic Processing Unit) virtualization optimization method based on delayed submitting

An optimization method and virtualization technology, applied in the field of virtualization, can solve the problems of frequent and large data transmission in the GPU virtualization framework, and achieve the effect of reducing repeated data transmission.

Inactive Publication Date: 2014-07-30
PEKING UNIV
View PDF3 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0006] The present invention mainly aims at the problem of large and frequent data transmission in the GPU virtualization framework based on the remote API. Reduce duplicate or redundant data transfers as much as possible

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • GPU (Graphic Processing Unit) virtualization optimization method based on delayed submitting
  • GPU (Graphic Processing Unit) virtualization optimization method based on delayed submitting
  • GPU (Graphic Processing Unit) virtualization optimization method based on delayed submitting

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0081] Based on the open source fully virtualized virtual machine manager KVM, and the Linux guest operating system, plus the open source GPU virtualization framework Gvirtus, we have implemented a new optimized GPU virtualization framework ( Figure 4 ). Further description will be made below in conjunction with this figure. The dotted line part is the part added or modified by this technology. The application example takes CUDA application program and CUDA function library as an example (CUDA is a computing platform launched by the graphics card manufacturer Nvidia, and the CUDA program here generally refers to the program formed after programming on this computing platform. s application).

[0082] The Gvirtus open source virtualization framework is divided into two parts: the front end and the back end: the front end is in the client, responsible for rewriting the CUDA library, intercepting the call of the CUDA function by the program in the client, and sending the call i...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a GPU (Graphic Processing Unit) virtualization optimization method based on delayed submitting. The method comprises the following steps that 1, a front end of a GPU virtualization frame on a client reads a binary file of a to-be-executed CUDA (Compute Unified Device Architecture) application program, and searches and marks a circulation which can be delayed to submit; 2, when the circulation part which can be delayed to submit is executed and before the circulation is finished, the front end caches all CUDA function call information and dependency relationship, jumps over the execution of a function call, and sends the cached information to a rear end of the GPU virtualization frame on a host once after the circulation is finished; 3, the rear end reconstructs and executes the function call according to the cached information, and then packs all task execution results and sends the packed task execution results to the front end once, wherein the circulation part which can be delayed to submit does not execute the CUDA function call in circulation, and the circulation still can be executed correctly. The method can be used for reducing the communication frequency of the front and rear ends, thus optimizing the virtualization performance of the GPU.

Description

technical field [0001] The invention belongs to virtualization technology, relates to a GPU virtualization optimization method, in particular to a GPU virtualization optimization method based on deferred submission. Background technique [0002] GPU has the characteristics of strong processing capability, high memory bandwidth, and low cost. It can not only perform graphics processing well, but also has strong advantages in high-performance computing. Due to its powerful parallel computing capabilities, GPU has been widely used in scientific computing, biological information, climate forecasting, engineering simulation and other fields in recent years, and has played a huge role. [0003] Virtualization technology has the advantages of resource sharing, online migration, process isolation, etc., and plays a great role in cluster and cloud computing. It not only saves computing resources, but also improves the stability and reliability of the system. Because virtualization ...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(China)
IPC IPC(8): G06F9/455H04L29/08
CPCG06F9/45558G06F2009/45579G06T1/20
Inventor 汪小林王寒冰刘宇玺桑燕罗英伟
Owner PEKING UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products