Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Technique for inter-procedural memory address space optimization in GPU computing compiler

a technology of inter-procedural memory address space and compiler, which is applied in the direction of program control, architecture with multiple processing units, instruments, etc., can solve the problems of wasting gpu resources, relying on the tagging approach described above, and wasting resources, so as to accelerate the execution of program code and save resources. , the effect of accelerating the execution of the application

Inactive Publication Date: 2013-05-09
NVIDIA CORP
View PDF7 Cites 12 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Benefits of technology

The disclosed technique improves the performance of applications by reducing the need for a graphics processing unit to resolve memory access operations at run time. This saves resources and speeds up the execution of the application. Additionally, the graphics processing unit can optimize the program code by re-ordering memory access and analyzing aliases, which further improves execution speed.

Problems solved by technology

However, a given memory access operation targeting a given memory address may not specify any particular memory space.
Relying on the tagging approach described above is problematic for two reasons.
First, reading a tag for each memory access operation is a costly operation and wastes GPU resources.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Technique for inter-procedural memory address space optimization in GPU computing compiler
  • Technique for inter-procedural memory address space optimization in GPU computing compiler
  • Technique for inter-procedural memory address space optimization in GPU computing compiler

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0020]In the following description, numerous specific details are set forth to provide a more thorough understanding of the present invention. However, it will be apparent to one of skill in the art that the present invention may be practiced without one or more of these specific details.

System Overview

[0021]FIG. 1 is a block diagram illustrating a computer system 100 configured to implement one or more aspects of the present invention. Computer system 100 includes a central processing unit (CPU) 102 and a system memory 104 communicating via an interconnection path that may include a memory bridge 105. System memory 104 includes an image of an operating system 130, a driver 103, and a co-processor enabled application 134. Operating system 130 provides detailed instructions for managing and coordinating the operation of computer system 100. Driver 103 provides detailed instructions for managing and coordinating operation of parallel processing subsystem 112 and one or more parallel p...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

A device compiler and linker is configured to optimize program code of a co-processor enabled application by resolving generic memory access operations within that program code to target specific memory spaces. In situations where a generic memory access operation cannot be resolved and may target constant memory, constant variables associated with those generic memory access operations are transferred to reside in global memory.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS[0001]This application claims the benefit of U.S. provisional patent application titled “Method for Inter-Procedural Memory Space Optimization in GPU Computing Compiler” filed on Nov. 7, 2011 and having Ser. No. 61 / 556,782. The entire content of the foregoing application is hereby incorporated herein by reference.BACKGROUND OF THE INVENTION[0002]1. Field of the Invention[0003]The present invention generally relates to graphics processing unit (GPU) computing compilers, and, more specifically, to a technique for inter-procedural memory address space optimization in a GPU computing compiler.[0004]2. Description of the Related Art[0005]Graphics processing units (GPUs) have evolved over time to support a wide range of operations beyond graphics-oriented operations. In fact, a modern GPU may be capable of executing arbitrary program instructions. Such a GPU typically includes a compiler that compiles program instructions for execution on one or more...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Applications(United States)
IPC IPC(8): G06F15/80
CPCG06F8/443G06F8/41G06F8/433G06F8/4442G06F8/445G06F8/45G06F8/456G06F9/5066
Inventor KONG, XIANGYUNWANG, JIAN-ZHONGLIN, YUANGROVER, VINOD
Owner NVIDIA CORP
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products