Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
HIGH-SPEED COMPUTER ACCELERATOR WITH PRE-PROGRAMMED FUNCTIONS
Document Type and Number:
WIPO Patent Application WO/2019/005443
Kind Code:
A1
Abstract:
A hardware accelerator for computers combines a stand-alone, high-speed, fixed program dataflow functional element with a stream processor, the latter of which may autonomously access memory in predefined access patterns after receiving simple stream instructions and provide them to the dataflow functional element, The result is a compact, high-speed processor that may exploit fixed program dataflow functional elements.

Inventors:
SANKARALINGAM KARTHIKEYAN (US)
NOWATZKI ANTHONY (US)
GANGADHAR VINAY (US)
Application Number:
PCT/US2018/036290
Publication Date:
January 03, 2019
Filing Date:
June 06, 2018
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
WISCONSIN ALUMNI RES FOUND (US)
International Classes:
G06F13/20; G06F13/16
Foreign References:
US20150261528A12015-09-17
US20170024167A12017-01-26
US20150268963A12015-09-24
US20060251092A12006-11-09
US20080244238A12008-10-02
Attorney, Agent or Firm:
BAXTER, Keith, M. (US)
Download PDF:
Claims:
CLAIMS

What we claim is:

1. A computer accelerator architecture comprising:

(1 ) a lower complexity general-purpose processor adapted to receive stream instructions from a higher complexity general -purpose processor, the stream instructions describing a logical stream of multiple data elements to be exchanged with a memory;

(2) a fixed program functional element to process successive input data elements of a logical stream, as initiated by availability of each input data element, io provide resultant output data elements of a logical stream; and

(3) a stream processor receiving the stream instructions from the simple general- purpose processor to autonomously read the input data elements of a logical stream from the memory according to the stream instructions and to autonomously write resultant output data elements of a logical stream to memory according to stream instructions; and wherein a pair consisting of only a single stream processor and only a single functional element, operates to intercommunicate data elements of a given logical stream.

2. The computer accelerator of claim 1 wherein the fixed program functional element does not include a program counter.

3. The computer accelerator of claim 2 wherein the fixed program functional element provides a map-reduce function,

4. The computer accelerator of claim 3 wherein the map-reduce function provides a mapping function selected from the group consisting of: multiplication and absolute difference and provides a reduction function selected from the group consisting of maximum, minimum, and addition.

5. "Hie computer accelerator of claim 2 wherein the fixed program functional element is a multiply accumulate function having at least, one output, that is a running total of a product of a pair of input arguments.

6. The computer accelerator of claim 2 wherein the fixed program functional element is a nonlinear transformation function having an output that is a base value summed to an interpolated value, where the base value is obtained from a first lookup table from most significant bits of an input argument and the interpolated value is a slope value obtained from a second lookup table from the most significant bits of the input argument times the least significant bits of the input argument.

7. The computer accelerator of claim 1 further including multiple functional elements and a switch assigning a single functional element at a time to the stream processor,

8. The computer accelerator of claim 7 further including multiple stream processors and multiple functional elements and a switch switchably assigning a single functional element to a single stream processor.

9. The computer accelerator of claim 1 wherein the functional elements provide for processing at least thirty-two bit arguments,

10. The computer accelerator of claim 9 wherein the functional elements can process in parallel separate data words having a length smaller than an argument length of the functional element by concatenating the separate data words together and processing them as an argument.

1 1. The computer accelerator of claim 1 wherein the stream processor simultaneously exchanges multiple streams with a given fixed program functional element.

12. The computer accelerator of claim 1 wherein stream instructions from the simple general-purpose processor to the stream processor may be received

asynchronously with respect to an operation of the functional element and provide for autonomous reading of multiple input values stored in memory or an autonomous writing of multiple output values from the computer accelerator according to different predefined memory access patterns.

13. The reconfigurable accelerator architecture of claim 1 1 wherein the simple general-purpose processor is further adapted to receive instructions and data from a primary processor to execute logical and arithmetic instructions in response to the instructions and data to return data to the primary processor without involvement of a stream processor and the at least one functional element.

14. The reconfigurable accelerator architecture of claim 1 wherein the stream processor provides pre-defined memory access patterns including a linear access pattern of contiguous addresses between two memory addresses and a strided access pattern of regularly spaced discontiguous addresses between two memory addresses.

15. A computer architecture comprising:

(1) a first general-purpose processor communicating with a memory for receiving instructions and reading and writing data;

(2) a second general-purpose processor simpler than the first general-purpose processor and adapted to receive stream instructions from the first general-purpose processor, the stream instructions describing a logical stream of multiple data elements to he exchanged with a memory;

(3) a plurality of fixed program functional elements each to sequentially process successive input data elements of a logical stream, as initiated by availability of each input data element, to provide resultant output data elements of a logical stream;

(4) a plurality of stream processors receiving the stream instructions from the simple general-purpose processor to autonomously read the input data elements of a logical stream from the memory according to the stream instructions and to autonomously write resultant output data elements of a logical stream to memory according to stream instructions; and

(5) an electronically controlled switch pairing a single stream processor with only a single functional element, the pair operating to intercommunicate data elements of a given logical stream.

16. The computer accelerator of claim 15 wherein the fixed program functional element does not include a program counter,

17. The computer accelerator of claim 16 wherein the fixed program functional element provides a multiply add-reduciion function having at least one output that is a first sum of a pair of second sums, each second sum being a sum of a pair of products, the products being products of a pair of input arguments,

18. The computer accelerator of claim 16 wherein the fixed program functional element is a multiply accumulate function having at least, one output that is a running total of a product of a pair of input arguments,

19. The computer accelerator of claim 16 wherein the fixed program functional element is a nonlinear transformation function having an output that is a base value summed to an interpolated value, where the base value is obtained from a first lookup table from the most significant bits of an input argument and the interpolated value is a slope value obtained from a second lookup table from the most significant bits of the input argument times the least significant bits of the input argument.

Description:
HIGH-SPEED COMPUTER ACCELERATOR WITH PREPROGRAMMED FUNCTIONS

STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR

DEVELOPMENT

CROSS REFERENCE TO RELATED APPLICATION

[0001] This application claims the benefit of U.S. Non-Provisionai application number 15/635,864 filed June 28, 2017, and hereby incorporated by reference.

BACKGROUND OF THE INVENTION

[0002] The present invention relates to electronic computer architectures and in particular to a high-speed computer accelerator having a limited functionality but improved applicability.

[0003] Computer accelerators provide special-purpose circuitry thai can be teamed with a general-purpose processor (GPU) to provide improved perfonnance in some computational applications.

[0004] Early computer accelerators expanded the hardware instruction set of the general-purpose processor with off-chip circuitry providing pre-programmed functions, that is, functions receiving data to execute a predetennined function on that data. These accelerators will henceforth be termed "fixed function'' accelerators referring to the fact that they execute one or more fixed functions. One example of such fixed-function accelerators was the so called "math coprocessor" providing specialized circuitry to execute predetermined floating-point and trigonometric function calculations. Such fixed-function accelerators are easily integrated into programming to the extent that their features can be invoked with a single program instruction requiring little more than the transfer of the necessary argument data and return of the calculated value.

[0005] As the power and speed of general-purpose processors increased, many fixed- function accelerators were rendered obsolete to the extent that their limited perfonnance gains were outweighed by the cost of the second integrated circuit in the computational burden of transferring data and control between the general-purpose processor and the accelerator. [0006] Possibly for this reason, current accelerator technology has largely focused on accelerators that operate as independently functioning special-purpose computers executing large blocks of instructions independent of the general processor. One example of a computer architecture accelerator is a graphic processor unit (GPU) which provides an array of special-purpose computer cores adapted to the specific time- consuming tasks associated with rendering graphics, Similar accelerators of this type of accelerator include application-specific accelerators such as the Microsoft Catapult FPGA accelerator, for use in data centers, and Google's Tensor processing unit for distributed machine learning. These accelerators are effectively special-purpose computers which, when combined with a general-purpose, high-performance

VonNeuman processor, for example, can greatly increase processing speed, for these specific applications.

[0007] The design of a computer architecture accelerators may not be practical for applications that are not associated with well-established problems or that do not have a sufficiently large demand to justify the time and expense of complex designs, changes to application programs, and changes to tool chains needed to use these accelerators,

SUMMARY OF THE INVENTION

[0008] The present invention provides a fixed-function accelerator that can substantially increase computational speeds for tasks that don't justify the design and manufacture of computer architecture accelerators. Speed advantages in executing the fixed functions is provided by using a high-speed dataflow architecture and by using a special -purpose stream processor to handle memory accesses, allowing long runs of data to be processed without involvement by the general-purpose processor. Proper selection of the individual functions implemented can give the accelerator broader applicability to a range of programs, implementing different fixed functions can reuse substantial portions of the architecture and the fixed functions are relatively simple to invoke from programs.

[0009] Specifically then, in one embodiment, the invention provides computer accelerator architecture having ( 1 ) a lower complexity processor adapted to receive stream instructions from a higher complexity general -purpose processor, the stream instructions describing a logical stream of multiple data elements to be exchanged with a memory; (2) a fixed program functional element to sequentially process successive input data elements of a logical stream, as initiated by availability of each input data element, to provide resultant output data elements of a logical stream; and (3) a stream processor receiving the stream instructions from the lower complexity general-purpose processor to autonomously read the input data elements of a logical stream from the memory according to the stream instructions and to autonomously write resultant output data elements of a logical stream to memory according to stream instructions. A pair consisting of only a single stream processor and only a single functional element operates to intercommunicate data elements of a given logical stream.

[0010] It is thus a feature of at least one embodiment of the invention to provide an accelerator that offers the versatility and wide applicability of a fixed- function accelerator, while still providing substantial speed advantages over execution of the same functions on sophisticated general-purpose processors.

[00111 The fixed program functional element may not include a program counter.

[0012] It is thus a feature of at. least one embodiment of the invention to employ dataflow architecture to provide substantial speed gains in function execution.

[0013] The fixed program functional element may provide a multiply add-reduction function having at least one output that is a first sum of a pair of second sums, each second sum being a sum of a pair of products, the products being products of a pair of input arguments.

[0014] It is thus a feature of at least one embodiment of the in vention to provide a fixed-function accelerator implementing the common map-reduce function.

[0015] Alternatively, or in addition, the fixed program functional element may be a multiply accumulate function having at least one output that is a running total of a product of a pair of input arguments,

[0016] It is thus a feature of at least one embodiment of the invention to provide a fixed-function accelerator implementing the common map-unit function.

[0017] Alternatively, or in addition, the fixed program functional element may be a nonlinear transformation function having an output that is a base value summed to an interpolated value, where the base value is obtained from a first lookup table from the most significant bits of an input argument and the interpolated value is a slope value obtained from a second lookup table from the most significant bits of the input argument times the least significant bits of the input argument.

[0018] The computer accelerator may include multiple functional elements and a switch assigning a single functional element at a time to the stream processor.

Alternatively, or in addition, the computer accelerator may include multiple stream processors and mukipie functional elements and a switch assigning a single functional element to a single stream processor.

[0019] It is thus a feature of at least one embodiment of the invention to increase the flexibility of the computer accelerator by allowing a selection among various functions that may be used for acceleration.

[0020] The functional elements provide for processing at least thirty-two bit arguments and may process in parallel separate data words having a length smaller than an argument length of the functionai element by concatenating the separate data words together and processing them as an argument.

[0021] It is thus a feature of at least one embodiment of the invention to permit, single instruction, multiple data type parallel processing.

[0022] The stream processor may simultaneously exchange multiple streams with a given fixed program functional element..

[0023] It is thus a feature of at least one embodiment of the invention provide multiple streams to the functional elements for maximum throughput and utilization of the memory bandwidth. T.

[0024] The stream instructions from the lower complexity general-purpose processor to the stream processor may be received asynchronously with respect to the operation of the functional element and provide for autonomous reading of multiple input values stored in memory or an autonomous writing of multiple output values from the accelerator according to different predefined memory access patterns.

[0025] It is thus a feature of at least one embodiment of the invention to permit parallel execution of memory access instructions and calculations for improved acceleration.

[0026] The lower complexity general -purpose processor may be adapted to receive instructions and data from the higher complexity, general -purpose processor to execute logical and arithmetic instructions and return the results to the higher complexity, general-purpose processor.

[0027] It is thus a feature of at least one embodiment, of the invention to permit the functional element to be incorporated into more complex accelerator functions implemented by the lower complexity, general-purpose processor or to allow the lower complexity, general-purpose processor to accept offloaded functions directly.

[0028] The stream processor may provide pre - defined memory access patterns including a linear access pattern of contiguous addresses betwee two memory addresses and a strided access pattern of regularly spaced discontiguous addresses between two memory addresses.

[0029] It is thus a feature of at least one embodiment of the invention to provide memory access patterns commonly used in multiple data instructions suitable for acceleration.

[0030] These particular objects and advantages may apply to only some embodiments falling within the claims and thus do not define the scope of the invention,

BRIEF DESCRIPTION OF THE DRAWINGS

[0031] Fig. 1 is a simplified block diagram of the accelerator of the present invention providing a bank of accelerators usable with a general-purpose processor;

[0032] Fig. 2 is a detailed block diagram of a single accelerator showing the combination of a lower complexity general-purpose processor, a fixed program functional element, scratchpad memory and a stream processor;

[0033] Fig. 3 is a detailed block diagram of the stream processor showing constituent stream dispatcher and stream engine elements;

[0034] Fig. 4 is a figure similar to that of Fig. 2 showing the provision of a pool of fixed program functional elements that may be individually paired with a given stream processor;

[0035] Fig. 5 is a figure similar to that of Fig. 4 showing a pooling of both stream processors and program functional elements for one-to-one pairing using a switch network;

[0036] Fig. 6 is a logical block diagram of a functional element of Fig. 2 showing associated vector buffers and dataflow coordination; [0037] Fig. 7 is a figure similar to that of Fig, 6 showing a map reduce functional element;

[0038] Fig. 8 is a figure similar to that of Figs, 6 and 7 showing a map unit functional element;

[0039] Fig. 9 is a figure similar to that of Figs. 6, 7, and 8 showing a nonlinear transformation functional element; and

[0040] Fig. 10 is a diagrammatic representation of computer memory showing definable memory access patterns for autonomous access by the stream processor.

DETAILED DESCRIPTION

[0041] Referring now to Fig. 1 , a computer architecture 10 constructed according to the present invention may provide for a higher complexity general-purpose processor 12. The higher complexity general-purpose processor 12. for example, may implement a sophisticated VoiiNeuman architecture including out-of-order execution and speculative execution using a rich instruction set of arithmetic and logical instructions sequenced with a program counter. Examples of such processors 12 include those manufactured by the Intel Corporation under the trade name of "i7" as well as similar devices

manufactured by Advanced Micro Devices, Inc., and ARM Ltd. As used herein," general-purpose" refers to a general -purpose instruction set and architecture sufficient to execute a wide range of programs, for example, such as could be executed at least by an Intel 8080 or subsequent processor in this family.

[0042] The higher complexity general -purpose processor 12 may communicate through an LI cache 14 with a memory system 16 using address and data lines 23. The memory system 16 provides a standard memory hierarchy including but not limited to additional levels of cache 18 coupled with one or more layers of increasingly larger scale memory 20, for example, composed of random access memory (RAM), disk memory and the [ike.

[0043] The memory system 16 may hold a program 22 for execution by the computer architecture 10 such as may benefit from hardware acceleration, for example, including image processing, machine learning, graph processing or the like.

[0044] The higher complexity general-purpose processor 12 may also communicate with a bank 21 of computer accelerators 24 by means of control lines 26 sending data that describe a pattern of memory access for obtaining argument data for the function of the computer accelerator 24 and a similar pattern for writing values back from the fixed- function accelerators 24. The control lines 26 may also send timing information initiating operation of the fixed-function accelerators 24. Control lines 26. as well as defining the memory access pattern and timing signals may provide for some limited configuration data, for example, selecting among different, fixed-function circuits available in the computer accelerator 24.

[0045] As will be discussed in greater detail below, during operation, the fixed- function accelerators 24, using the memory data pattern provided by the higher complexity general -purpose processor 12, may independently access the memory system 16 at the L2 cache using similar address and data lines 23 without further assistance of the higher complexity general -purpose processor 12, This access may be moderated by a load-balancing circuit 25 to eliminate deadlock by ensuring each computer accelerator 24 obtains sufficient access to memory system 16, for example, using conventional deadlock elimination techniques. During operation of the fixed-function accelerators 24, the higher complexity general-purpose processor 12 may shut down or be used for other tasks during that calculation.

[0046] Referring now to Fig. 2, each computer accelerator 24 includes a lower cornplexity general-purpose processor 30, a fixed-function element 32, a memory interface 34, scratchpad memory 36 and a stream processor 38, the latter including a stream dispatcher 40 and a stream engine 42.

Lightweight Processor

[0047] During operation, the lower complexity general-purpose processor 30 communicates with the higher complexity general-purpose processor 12 to receive instructions therefrom and issue stream instructions to the stream processor 38. The stream processor 38 in turn will control the memory interface 34 to obtain information necessary for calculation by the fixed-function element 32 (either directly or through the scratchpad memory 36) and to return information after that calculation to the memory interface 34 for storage (again either directly or through the scratchpad memory 36).

[0048] In this regard lower complexity, general -purpose processor 30 may be less complicated and/or slower than the higher complexity general -purpose processor 12 as is sufficient to provide coordination of the other components of the computer accelerator 24. For example, the lower complexity general -purpose processor 30 may be a

VonNeuman, single-issue, in-order core without speculative execution executing basic arithmetic and logical functions. In this regard, the lower complexity general -purpose processor 30 will require much less integrated circuit area than the higher complexity general-purpose processor 12 and will use much less power. In some embodiments, the lower complexity general-purpose processor 30 may be implemented with discrete circuitry and a fixed program and thus may not necessarily employ programmable computer architecture. The lower complexity general -purpose processor 30 and memory interface 34 may share the same memory access or, as depicted, may provide for separate memory access channels.

Fi ed- function unit

[0049] The fixed-function element 32 may include multiple input vector buffers 44, output vector buffers 46, and transfer vector buffers 48 for communicating arguments to and values from the fixed-function element 32 or in a loop around the fixed-function element 32. One or more indirect addressing vector buffers 51 may also be present of comparable size.

[0050] Referring now to Fig. 6, the fixed-function elements 32 provide a set of interconnected function primitives 33, tor example, but not limited to, those providing addition, multiplication, accumulation, and lookup tables. Each of these function primitives 33 may receive data from a corresponding input vector buffer 44 and provide results to corresponding output vector buffers 46 with possible movement of data through transfer vector buffer 48 in multiple cycles. Generally, the architecture of the fixed- function elements 32 will be that of a dataflow architecture meaning that they begin execution upon the availability of data in the input vector buffers 44 using, for example, a circuit identifying whether corresponding data is present at inputs of each of the function primitives 33.

[005 i ] Data flows through the function primitives 33 in a deterministic manner according to stages 66 eliminating race conditions which may be enforced either by a clocking mechanism or by providing similar delays (for example, enforced through no operation function primitives which perform no operation but delay) along each data path 68 between the function primitives 33. By operating without a program counter, extremely high-speed performance can be obtained providing calculations as fast as data is delivered to the input vector buffers 44. In addition, it will be understood, that the dataflow pathway between the function primitives 33 forms an effective pipeline so that early stages 66 of the fixed-function element 32 may be receiving new data from the input vector buffers 44 as data is being processed in the later stages 66.

[0052] Generally, function primitives 33 will provide for predication, limiting the need for unnecessary control steps and may operate in parallel on multiple data words concatenated together to form a single argument passing through the function primitives 33, Thus, for example, a function primitive 33 having an argument width of 32 bits may simultaneously process four 8-bit arguments concatenated together, for example, by performing saturation arithmetic in which carries and overflows are handled, for example, by suppressing the carry or overflow and setting the result to the highest permissible value, A similar technique can be used with under flows and borrows that setting the result to the lowest permissible value,

[0053] Referring now to Fig. 7, an important fixed-function element 32 may provide for a map reduce or multiply add-reduetion function. In such a function, arguments at pairs of input vector buffers 44 may be multiplied in a first stage 66a by multiply function primitives 33a providing the mapping. Output products from the multiply function primitives 33a are then added pairwise together at a second stage 66b providing a first part of the reduction by additional adder function primitives 33a. Outputs from this second set of adder function primitive 33a are then summed at a third stage 66c by an additional adder primitive 33a which provides its output to output vector buffer 46 completing the reduction. It will generally be understood that the map-reduce function may implement a variety of different mapping procedures including, for example, multiplication, absolute difference, and the like, and may provide a variety of different reduction functions including, for example, addition, maximum, and minimum functions, and the like,

[0054] Alternatively, or in addition, as shown in Fig. 8, fixed-function element 32 may provide for a map unit or multiply-and-accumuiate function. In such a function, arguments at pairs of input vector buffers 44 may be summed at a first stage 66a by adder function primitives 33a, Outputs of these adder function primitives 33 are provided to accumulator function primitives 33b at a second stage 66b which sums each current output to a running total of previous outputs for a predetermined number of arguments. The accumulator function primitives 33b then provide outputs to output vector buffers 46.

[0055] Alternatively, or in addition, as shown in Fig. 9, the fixed-function element 32 may provide for a nonlinear transformation or piecewise linear approximation. At a first stage 66a values from input vector buffers 44 are split into high-order bits 69 used to determine base values and low-order bits 71 used for interpolation. The high-order bits 69 are received by a base lookup table 72 which converts the high-order bits 69 to a base value 73 of the transformation according to a transformation function embedded in the lookup table 72, for example, at the time of manufacture or configuration. The high- order bits 69 are also sent to second lookup table 74 producing a slope value 75 for the transformation function. The slope value 75 is multiplied by low-order bits 71 at a next stage 66b and the resulting interpolated value 77 is then summed to the base value at a succeeding stage 66 to provide an output-to-output vector buffer 46.

[0056] Generally, the fixed-function element 32. does not provide a program counter and may or may not require control-flow instructions. For example, control-flow may be implicit in the interconnection 68 of the function primitives 33. Alternatively, or in addition, control-flow may be provided by the fixed-function elements 32, for example, internally implementing branch instructions and selecting among physical

interconnections. Calculations occur as soon as operands are available within the constraint of the regular sequencing through the fixed-function elements 32 which may- occur at high speed. The fixed-function elements 32 may receive data, for example, configuring the lookup table 74 but do not receive instructions providing an ordered set of execution steps according to instruction type.

Stream. Processor

[0057] Referring to Figs, 2 and 6, each of the input vector buffers 44 and output vector buffers 46 of the fixed-function element 32 provide first-in, first-out buffers that may be filled asynchronously or emptied asynchronously to the processing performed by the fixed-function element 32 under the control of the stream processor 38. The vector buffers thus provide for independent but parallel data acquisition and data processing. In one embodiment, each buffer may provide eight, sixty-four bit words, thus being 8x64 wide, and have an arbitrary depth, for example, greater than 10 and optionally 100, Additional similar buffers 51 of similar construction and independent of input vector buffers 44 and output vector buffers 46 may be used for storing streaming addresses for indirect loads and stores if required.

[0058] in essence, the stream processor 38 provides a state machine that can move data autonomously between the memory system 16 and another storage location once it receives program instructions from the lower complexity general-purpose processor 30, Generally, the stream processor 38 will move input data from the memory system 16 to either the scratchpad memory 36 or from the scratchpad memory 36 to the input vector buffers 44, or may move output data from the scratchpad memory 36 to the memory system 16, or from output vector buffers 46 to the scratchpad memory 36 or the memory system 16 according to a predefined pattern. In this regard, the stream processor 38 may provide for three separate circuits, one for memory, one for scratchpad, and one for controlling re-cycling of data from output port to input port and also the generation of constant values. These three circuits may operate independently (but for synchronization through the memory interface 34) for high-speed operation.

[0059] Referring now to Fig. 3, more specifically, the stream processor 38 may include a stream dispatcher 40 and stream engine 42. The stream dispatcher 40 receives a set of stream instructions at a stream queue 50 from the lower complexity general- purpose processor 30, these instructions defining a pattern of access of the memory system 1 6 to obtain data elements to be processed by the fixed-function element 32 (shown in Fig, 2). These queued stream instructions will ultimately be sent one of a set of stream engines 52. in the stream engine 42 for concurrent execution.

[0060] Each stream engine 42 may handle the necessary protocol for communicating (reading or writing data) with the memory system 1 6 and provides the ability to calculate a set of addresses to obtain a stream of such data according to the stream instructions for processing by the fixed- function element 32. To the extent that the accelerator 24 may communicate directly with an 1,2 cache 18 (as shown in Fig. 1 ) the stream engine 42 may handle memory coherence problems, for example, by sending a tag invalidation to the lower complexity general-purpose processor 30 whenever a stream crosses a page boundary. The LI cache 14 and L2 cache 18 may be assumed to be using virtual addresses when virtualization is employed. Necessary address translation using a translate lookaside buffer of an MMU is accomplished according to techniques understood in the art.

[0061] Stream instructions from the stream queue 50 will only be dispatched to the stream engines 52 by a dispatcher 55 when the necessary resources needed for the stream are available and according to the program order of program 22. The critical resources needed for a stream include availability of the input vector buffer 44 and output vector buffer 46 or of the scratchpad memory 36.

[0062] The dispatcher 55 determines availabilities of resources using a scoreboard 54 which provides a state of each stream resource as either "taken," ' ree," or "all requests in flight" which may be updated by the dispatcher 55. A critical resource moves from "free" to "taken" when the stream instructions are enrolled in the stream engines 52. The given stream of those instructions then logically owns the resource while in flight. When the stream is finished, the associated stream engine 52 in the stream engine 42 notifies the dispatcher 40 to update the scoreboard 54 to show the resource is in the tree state. The "all requests in flight state" indicates that all requests for the memory stream are completely sent to the memory system 16 but have not arrived. This state allows scheduling of another conflicting stream enabling two stream engines 52 to use the same critical resources in overlapping configuration for additional efficiency.

[0063] The stream processor 38 may also control the forwarding of streams to the stream engines 52 according to barrier instructions that may be issued by the lower complexity general-purpose processor 30. Barrier instructions prevent the issuance of new stream instructions to the stream processors until a given stream identified by the barrier instruction is complete. Thus, barrier instructions provide a method of ensuring proper execution order of the calculations performed by the fixed-function elements 32.

[0(364] Generally, then, stream instructions will include: stream instructions for providing a stream of data to or from the fixed-function element 32 without, involvement of the lower complexity general-purpose processor 30 or the higher complexity general- purpose processor 12; and barrier instructions used to enforce some degree of serialization of access of data by the stream processor 38 as will be discussed below. Ex amples of these stream instructions (shown in Table I) provided by the lower complexity general-purpose processor 30 to the stream processor 38 generally identify a source of data, destination data, and the data pattern as follows:

Table I-Example Stream Instructions

designated output port.

[0065] These instructions transfer data between storage locations autonomously using a designated pattern as will be discussed below.

[0066] Indirect addressing of data by the stream engine 52 is possible using stored data (for example, in a vector buffer 51) as an address value, In indirect addressing, data, for example, from the streaming pattern, is used as the address to obtain further data that is operated on by the fixed-function element 32. This indirect addressing effects pointers, useful, for example, when accessing the rows of a sparse matrix. The stream processor 38 may provide capability to facilitate indirect access by chaining two streams together, the first stream for accessing a contiguous or strided pattern of pointers, and subsequent streams to load those pointers' values from the memory system 16 and deliver them to the reconfigurabie fixed-function clement 32. Additional instructions are provided to generate constant values (rather than loading these from memory) and to discard unused output values (as opposed to loading them into nonfunctional memory areas).

[0067] Generally, each of these instructions may be optionally issued directly by the higher complexity general -purpose processor 12 as part of the instruction set architecture of the accelerator and die data in these instructions used with minimal processing by the lower complexity general-purpose processor 30 to control other components of the accelerator,

[0068] Referring now to Fig. 10, using data from the above instructions, the stream processor 38 provides a state machine for autonomous loading or writing data between storage locations using predefined patterns of memory access. The simplest pattern is a linear pattern 80 that reads (or writes) sequentially every memory address between a starting address 82 and ending address 84 designated by the instruction. This pattern is indicated by setting the stride length equal to zero. The starting address is designated in the above instructions as the source address for reads and the destination address for writes while the ending address is indicated either by providing an actual ending address or the necessary data to calculate and ending address for example the number of elements when the source is a port or a product of the access size times the stride length times the number of strides added to the starting address when the source is the memory or scratchpad,

[0069] Alternatively, the stream processor 38 may be programmed to use a strided pattern 85 by setting the stride length equal to a nonzero value which describes a gap or stride 86 in addresses between access portions 87 defined by the access size.

[0070] Similarly, an overlapped axis pattern 88 may be invoked by setting the access size to greater than the stride size which signals an overlapping pattern. A repeated pattern 89 is easily obtained by setting the stride length to zero with the repetition being provided by the number of strides.

[0071] As used herein, predefined memory access pattern means a limited number of patterns that may be defined by a discrete set of pattern instructions providing a pattern type and delimiter values where the pattern may be defined prior to the calculation for which the memory access is required to be performed as opposed to memory access patterns that are a function of calculations made on the data being accessed. Autonomous as is used herein means without necessary further guidance by the processors 12 or 30, [0072] As noted above, the lower complexity general-purpose processor 30 may also provide for barrier instructions to the stream processor 38, such instructions which block the issuance of new memory access instructions until certain previous instructions associated with a data storage resource are complete. For example, a barrier instruction (shown in Table II above) associated with a writing to the scratchpad memory 36 will block subsequent writing to the scratchpad memory 36 until all writings to the scratchpad memory 36 before the barrier instruction are completed, Barriers can also be used to signal completion of the calculation to the lower complexity general-purpose processor 30 to the extent that they indicate completion of a previous stream upon satisfaction.

Table II-Example Barrier Instructions

[0073] i will be appreciated that in the absence of barrier instructions all streams would be allowed to execute concurrently. Therefore, if two streams command read and write of the same scratchpad or memory address, the semantics of that operation would be undefined. Barrier instructions allow enforcement of the memory dependencies and can be implemented by the compiler and provided in the stream instructions, This is independent of resource contention.

[0074] The lower complexity general-purpose processor 30 may also expose hardware parameters of the computer accelerator 24 including a number and type of fixed-function elements 32 and a depth of buffers 50 and 50' for use by a compiler as is generally understood in the art. [0075] Referring now to Fig. 4, in this architecture, a fixed-function element 32 will be associated with a single stream processor 38 and vice versa; however, the invention contemplates that the accelerator 24 may be configured to select among different fixed- function elements 32 in a configuration step using a configuration multiplexer 56 and demultiplexer 58, for example, set at time of manufacture or in a configurations step, for example, by a system driver, to select one of the fixed- function elements 32 and leave the others idle or to be used by other stream processors. In other respects, this embodiment operates identically to that described with respect to Fig. 2.

[0076] This principle can he extended, as shown in Fig. 5, by pooling the fixed- function elements 32 in a separate pool 60 that may be connected, for example, so that any fixed-function element 32 in the pool 60 may be connected by a crossbar switch 62 to any of the different accelerators 24 which do not include internal fixed- function elements 32, In this way, improved utilization of a variety of different functional units may be had by allocating them freely among accelerators 24. Again, this configuration may be accomplished at the factory or during a configuration of the accelerators 24 prior to use and after the configuration operates essentially as the architecture of Fig. 2.

[0077] Certain terminology is used herein for purposes of reference only, and thus is not intended to be limiting. For example, terms such as "upper", "lower", "above", and "below" refer to directions in the drawings to which reference is made. Terms such as "front", "back", "rear", "bottom" and "side", describe the orientation of portions of the component within a consistent but arbitrary frame of reference which is made clear by reference to the text and the associated drawings describing the component under discussion, Such terminology may include the words specifically mentioned above, derivatives thereof, and words of similar import. Similarly, the terms "first", "second" and other such numerical terms referring to structures do not imply a sequence or order unless clearly indicated by the context.

[0078] The terms "lower complexity" and "higher complexity" refer only to relative complexity of the lower complexity and higher complexity processors and not absolute complexity. The term "fixed program functional element" refers to functional elements receiving numeric values to execute a function on those values to produce a numeric result where the function is not altered by the general-purpose processor associated with the accelerator.

[0079] When introducing elements or features of the present disclosure and the exemplary embodiments, the articles "a", "an", "the" and "said" are intended to mean that there are one or more of such elements or features. The terms "comprising", "including" and "having" are intended to be inclusive and mean that there may be additional elements or features other than those specifically noted. It is further to he understood that the method steps, processes, and operations described herein are not to be construed as necessarily requiring their performance in the particular order discussed or illustrated, unless specificall identified as an order of performance. It is also to be understood that additional or alternative steps may be employed,

[0080] References to microcontroller should be understood to include any circuit capable of executing the functions described herein including but not necessarily limited to VonNeuman architectures.

[0081 ] It. is specifically intended that the present invention not be limited to the embodiments and illustrations contained herein and the claims should be understood to include modified forms of those embodiments including portions of the embodiments and combinations of elements of different embodiments as come within the scope of the following claims. All of the publications described herein, including patents and nonpatent publications, are hereby incorporated herein by reference in their entireties.