Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
PARALLELIZATION METHOD AND SYSTEM
Document Type and Number:
WIPO Patent Application WO/2015/107378
Kind Code:
A1
Abstract:
A method in a system for handling compiled code is provided. The system comprises a Just-In-Time, JIT, compiler for compiling code, and at least one array processor unit comprising a plurality of processors for executing program code. The method comprises compiling input program code, whereby compiled program code is generated for the input program code. While compiling at least two parts of the compiled program code to be executed in parallel are identified. The identified at least two parts of compiled code are executed in parallel speculatively on at least two respective of the plurality of processors. Control if the at least two parts of in parallel executed code are in conflict with each other is performed, and if the parts are in conflict, the parts are executed again.

Inventors:
ISBERG ANDERS (SE)
GUSTAVSSON JONAS (SE)
RASMUSSON JIM (SE)
Application Number:
PCT/IB2014/000050
Publication Date:
July 23, 2015
Filing Date:
January 20, 2014
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
SONY CORP (JP)
International Classes:
G06F9/45
Foreign References:
US20100269102A12010-10-21
Other References:
JAN MARTINSEN ET AL: "Heuristics for Thread-Level Speculation in Web Applications", IEEE COMPUTER ARCHITECTURE LETTERS, 1 January 2014 (2014-01-01), pages 1 - 1, XP055127004, ISSN: 1556-6056, DOI: 10.1109/L-CA.2013.26
Attorney, Agent or Firm:
KAMEYA, Yoshiaki (Daiichi Tomizawa Building3-1-3, Yotsuya,Shinjuku-ku, Tokyo, JP)
Download PDF:
Claims:
CLAIMS

1. A method in a system (100) for handling compiled code comprising

a Just-In-Time, JIT, compiler (101 ) for compiling code, and

at least one array processor unit (102) comprising a plurality of processors (103) for executing program code, the method comprising:

- compiling (201 ) input program code, whereby compiled program code is

generated for the input program code,

- while compiling, identifying (202) at least two parts of the compiled program code to be executed in parallel,

- in parallel executing (203) the identified at least two parts of compiled code speculatively on at least two respective of the plurality of processors,

- controlling (204) if the at least two parts of in parallel executed code are in conflict with each other, and

if the parts are in conflict, repeat action (203) and (204).

2. Method according to claim 1 , further comprising that the identified at least two parts of compiled program code comprises a similar amount of instructions.

3. Method according to claim 1 or 2, further comprising that more than two parts of compiled program code are identified.

4. Method according to claim 1 , 2 or 3, further comprising that at least one of the

identified parts of compiled program code comprises two or more functions merged together into a single function.

5. Method according to claim 1 , 2 or 3, further comprising that all of the identified parts of compiled program code comprises two or more functions merged together to a single function.

6. Method according to claim 4 or 5, further comprising that more than two functions are merged together to a single function.

7. Method according to claim 4, 5 or 6 further comprising that the two or more functions comprises a similar amount of instructions.

8. Method according to any of claims 1-7, further comprising, in the compiling action, identifying parts of the program code being frequently executed, and, for such frequently executed parts, perform action (202), action (203) and action (204). 9. Method according to any of claims 1-8, further comprising, in the executing action, identifying parts of the program code being frequently executed, and, for such frequently executed parts, perform action (202), action (203) and action (204).

10. Method according to any of the claims 1-9, the at least one array processor unit (102) being one or more General Purpose Graphics Processing Unit, GPGPU, wherein the in parallel executing (203) is performed on a plurality of processors comprised in the GPGPU.

11. Method according to claim 10, wherein the GPGPUs is combined with a Central Processing Unit, CPU.

12. Method according to any of claims 1-1 1 , wherein statics of the code are gathered during the execution action (203). 13. Method according to any of claims 1-12, wherein the code is JavaScript code.

14. Method according to any of claims 1-13, wherein the identified parts of the compiled program code are functions. 15. Method according to any of claims 1-14, wherein more than two parts of the compiled program code are identified and executed in parallel.

16. Method according to any of claims 1-15, wherein the more than two parts of the

compiled program code are a plurality of functions.

17. A system (100) for handling compiled code comprising

a Just-In-Time, JIT, compiler (101 ) for compiling code, and

at least one array processor unit (102) comprising a plurality of processors (103) for executing program code,

wherein the JIT compiler (101 ) is adapted to compile input program code, whereby compiled program code is generated for the input program code, and is adapted to, while compiling, identify at least two parts of the compiled program code to be executed in parallel, and wherein the at least one array processor unit (102) is adapted to, in parallel, execute the identified at least two parts of compiled code speculatively on at least two respective of the plurality of processors,

the system being adapted to control if the at least two parts of in parallel executed code are in conflict with each other.

Description:
PARALLELIZATION METHOD AND SYSTEM

TECHNICAL FIELD

Embodiments herein relate to methods in a system for handling compiled code, and especially for in parallel executing parts of compiled code. Other embodiments herein relate to a system comprising a Just-In-Time, JIT, compiler for compiling code, and at least one array processor unit comprising a plurality of processors for executing program code.

BACKGROUND

Thread level speculation has been found to be a method that may be used to parallelize execution of applications without updating the application logic to utilize many cores. In particular, JavaScript based applications benefit significantly on application processors, Central Processing Units (CPU), with multiple cores.

When executing a JavaScript program, the code is typically compiled with a Just In Time compiler where compilation speed is important. The generated code is then made available as fast as possible for execution and more or less no optimizations are applied. By gathering statistics about how frequently different code parts are executed it is possible to spot important code segments "hot code areas" consuming a lot of processor time. The compiler focuses on those code areas and starts to optimize the code. Typically, this is done as a parallel activity, thereby enhancing the performance during execution time.

Thread level speculation is a different method to enhance performance. The concept is that once a new function is found, the function is started in a new thread in a speculative manner. This function may run in parallel with other functions that has their own thread. Once the execution of the function is ready, it is checked if there are any failures or conflicts, due to the speculation. If there is a failure or conflict, a rollback is executed and the function is re- executed in sequential manner. If the speculation fails, there is typically some mechanism that prohibits that the function is speculated on again or at least decreases the likelihood that the function is speculated on, in favor of other functions that is more likely to be successfully. In particular for web applications this methodology may be very useful.

SUMMARY

An object of embodiments herein is to provide an improved way of speculatively executing code in parallel.

According to a first aspect the object is achieved by a method in a system for handling compiled code. The system comprises a Just-In-Time (JIT) compiler for compiling code, and at least one array processor unit comprising a plurality of processors for executing program code. The JIT compiler compiles input program code, whereby compiled program code is generated for the input program code. While compiling, the JIT compiler identifies at least two parts of the compiled program code to be executed in parallel. The identified at least two parts of compiled code are speculatively executed in parallel on at least two respective of the plurality of processors. The system controls if the at least two parts of in parallel executed code are in conflict with each other, and if the parts are in conflict, execution is repeated.

The above mentioned object is achieved, in another aspect, by a system for handling compiled code. The system comprises a Just-In-Time (JIT) compiler for compiling code, and at least one array processor unit comprising a plurality of processors for executing program code. The JIT compiler is adapted to compile input program code, whereby compiled program code is generated for the input program code. The JIT compiler is adapted to, while compiling, identify at least two parts of the compiled program code to be executed in parallel. The at least one array processor unit is adapted to, in parallel, execute the identified at least two parts of compiled code speculatively on at least two respective of the plurality of processors. The system is adapted to control if the at least two parts of in parallel executed code are in conflict with each other.

BRIEF DESCRIPTION OF THE DRAWINGS

Examples of embodiments herein are described in more detail with reference to attached drawings in which:

Figure 1 shows a system in accordance with embodiments herein.

Figure 2 is a flow chart showing methods herein. DETAILED DESCRIPTION

Embodiments herein will be exemplified in the following non-limiting description.

Figure 1 shows a system 100 in accordance with embodiments herein. The system 100 for handling compiled code comprises a Just-In-Time, JIT, compiler 101 for compiling code, and at least one array processor unit 102. The array processing unit 102 comprises a plurality of processors 103 for executing program code. The array processor unit 102 may be one or more General Purpose Graphics Processing Unit, GPGPU. The potential of using a GPGPU is huge since there are many more processing units on a GPGPU compared to a CPU, traditionally used. Consequently, jobs that may be parallelized may run much faster on a GPGPU and typically consume less battery power. However, to utilize the power of a GPGPU, same code runs on the different processors with different input data. A typical use case may be imaging processing where 32, or more, units are started in parallel with different input data. As long as there is no data dependency, the execution will go 32 times faster.

" The in parallel executing is performed on the plurality of processors 103. The JIT compiler 101 is adapted to compile input program code, whereby compiled program code is generated for the input program code. The JIT compiler is adapted to, while compiling, identify at least two parts of the compiled program code to be executed in parallel. The array processor unit 102 is adapted to, in parallel, execute the identified at least two parts of compiled code speculatively on at least two respective of the plurality of processors. The system 100 is adapted to control if the at least two parts of in parallel executed code are in conflict with each other.

Thread level speculation may be utilized on a CPU in combination with a GPGPU or any other types of array processors. The program code may be JavaScript code. During JavaScript execution there is no guarantee that many similar jobs may be identified, that the jobs could be started in a synchronous manner and that there is no data dependency between in-data. This may be solved by combining speculation and gathering statics of the code during execution. It is to be noted that the methods described herein are not limited to JavaScript. The methods proposed may easily be extended to be used by any compiler or interpreter for a computer language.

A JIT compiler architectures may be extended to utilize also multicore processors such as GPGPUs (General Purpose Graphics Processing Unit) or any other types of array processors. An array processor, such as a GPGPU, may be used in combination with a CPU. Instead of generating code for the CPU only, as is the case traditionally, the code generator is extended to identify good candidates for the GPGPU and for selected parts of the code generate instructions for the GPGPU. Traditionally, instructions are instead generated for a CPU. Such candidates may for example be "for-loops" that under certain circumstances, e.g. when there are no data dependencies between loop turns, naturally could be parallelized. Once some good candidate is identified and in-data parameters are ready, execution of the parallel code block is started immediately in a speculative manner. Once the execution is completed, the result is checked for conflicts and/or failures. If there is at least one conflict, a rollback is executed. If the execution was successful, the result is saved for immediate or later usage. Further, it is proposed that the JIT compiler, if any frequently executed parts, so called hot code areas, are detected, tries to detect which of these hot code areas that are feasible for a GPGPU. With reference to Figure 2, a method in the system 100 for handling compiled code will now be described. Action 201

The JIT compiler 101 compiles input program code, whereby compiled program code is generated for the input program code. In the compiling action, parts of the program code being frequently executed may be identified, and, for such frequently executed parts, action 202, action 203 and action 204 may be performed.

Action 202

The JIT compiler 101 performs, while compiling, identifying at least two parts of the compiled program code to be executed in parallel. The identified parts of the compiled program code may be functions. The two or more functions may comprise a similar amount of instructions. More than two parts of the compiled program code may be identified and executed in parallel. The more than two parts of the compiled program code may be a plurality of functions. The identified two parts of compiled program code may comprise a similar amount of instructions. One of the identified two parts of compiled program code may comprise two or more functions merged together into a single function. Thus, merging of functions may be done into one being a super set of several functions to create one function that may run in parallel on different processors. Thereby it will be possible to run the single merged function in several cores in parallel with different indata. All of the identified parts of compiled program code comprising two or more functions may be merged together into a single function. All of the identified parts of compiled program code may comprise two or more functions merged together to a single function. More than two functions may be merged together to a single function. Thus, when a speculation candidate function is found, this function may be stored as suitable candidate, whilst new candidates are found. Once a feasible number of candidates are found, all the functions may be merged into a new function being a super-set of all candidate functions. The new super-set function may then be provided with an in-parameter which code portion shall be executed. Alternatively, instead of speculating directly statics about feasible candidates may be gathered during execution. Apply optimizations may then start once there is data supporting that a function is a good candidate for speculation.

Action 203

At least two respective of the plurality of processors 103 executes speculatively in parallel the identified at least two parts of compiled code. In the executing action, parts of the program code being frequently executed may be identified, and, for such frequently executed parts, action 202, action 203 and action 204 may be performed. Statics of the code may be gathered during the execution action 203. Action 204

The system controls 204 if the at least two parts of in parallel executed code are in conflict with each other, and if the parts are in conflict, action 203 and 204 are repeated.

Further, in the previous description specific details have been set forth, such as particular embodiments for purposes of explanation and not limitation. However, it will be appreciated by one skilled in the art that other embodiments may be employed apart from these specific details. In some instances, detailed descriptions of well-known methods, nodes, interfaces, circuits, and devices are omitted so as not obscure the description with unnecessary detail. Those skilled in the art will appreciate that the functions described may be implemented in one or more nodes, e.g. a wireless modem or a wireless device, using hardware circuitry, e.g., analogue and/or discrete logic gates interconnected to perform a specialized function, ASICs, PLAs, etc., and/or using software programs and data in conjunction with one or more digital microprocessors or general purpose computers.

Nodes that communicate using the air interface also have suitable radio communications circuitry. Moreover, the technology may additionally be considered to be embodied entirely within any form of computer-readable memory 604, such as solid-state memory, magnetic disk, or optical disk comprising an appropriate set of computer instructions that would cause a processor to carry out the techniques described herein.

Hardware implementation may include or encompass, without limitation, digital signal processor, DSP, hardware, a reduced instruction set processor, hardware, e.g., digital or analogue circuitry including but not limited to Application Specific Integrated Circuits, ASIC, and/or Field Programmable Gate Arrays, FPGAs, and where appropriate state machines capable of performing such functions.

In terms of computer implementation, a computer is generally understood to comprise one or more processors or one or more controllers, and the terms computer, processor, processing unit 601 and controller may be employed interchangeably. When provided by a computer, processor, or controller, the functions may be provided by a single dedicated computer or processor or controller, by a single shared computer or processor or controller, or by a plurality of individual computers or processors or controllers, some of which may be shared or distributed. Moreover, the term "processor" or "controller" also refers to other hardware capable of performing such functions and/or executing software, such as the example hardware recited above.

Although the description above comprises many specifics, they should not be construed as limiting but as merely providing illustrations of some presently preferred embodiments. The technology fully encompasses other embodiments which may become apparent to those skilled in the art. Reference to an element in the singular is not intended to mean "one and only one" unless explicitly so stated, but rather "one or more." All structural and functional equivalents to the elements of the above-described embodiments that are known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed hereby. Moreover, it is not necessary for a device or method to address each and every problem sought to be solved by the described technology for it to be encompassed hereby.

When using the word "comprise" or "comprising" it shall be interpreted as non-limiting, in the meaning of consist at least of.

When using the word action/actions it shall be interpreted broadly and not to imply that the actions have to be carried out in the order mentioned. Instead, the actions may be carried out in any suitable order other than the order mentioned. Further, some action/actions may be optional.

The embodiments herein are not limited to the above described examples. Various alternatives, modifications and equivalents may be used.