<html><body style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><br><div><div>Am 21.05.2009 um 16:55 schrieb Paul Krueger:</div><br class="Apple-interchange-newline"><blockquote type="cite"><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; ">You can't emphasize these points enough. GPGPU technology has its place, but it's not perfect for everything. If you have an application where data can be partitioned up neatly and distributed to separate processing elements which tend to do the same limited things over and over (FFT's are a good example), then GPGPU's may be appropriate (as are FPGA's for similar reasons, although there are certainly other factors there). If you have an application where each processing thread may dynamically determine that it needs data from an arbitrary location within a very large block of memory or needs to do frequent updates within large data blocks in arbitrary ways, then GPGPU's are not appropriate because the communication and synchronization costs will typically kill you. That's especially true on any larger distributed memory architecture, but even on smaller systems you might overwhelm the memory subsystem. Many of the sorts of AI, graph, and intelligent applications that I am personally more interested in fall into the second category, so GPGPU's will likely not be of much help.</div></blockquote><div><br></div>The appeal of the GPU is that it has lots of computing elements (originally designed for 3d rendering tasks). The trend is that these computing elements are getting more flexible and numeric computations are getting more accurate. Now we are at a point where the first programming languages appear that will allow writing portable algorithms that are able to run on the GPU: CUDA, OpenCL, ...</div><div><br></div><div>The typical applications the average user will see are about manipulating large amounts of multimedia data. Think using FinalCut Pro to render visual effects, to convert video formats etc. Think iPhoto/Aperture plugins that manipulate large 'raw' images (dust filters, sharpeners, ...). At the same time scientists working with such kind of data will find its uses, too.</div><div><br></div><div>If you look back at the SIMD Connection Machine it was thought that there are graph applications that are able to run in parallel on such a machine. The graph was mapped to compute elements which could communicate efficiently with nearby elements. A somewhat typical application domain was also querying data collections. Spread those data collections across a huge number of compute elements and run MAP/REDUCE operations. An outcome was the WAIS protocol and a document query engine running on the connection machine ( <a href="http://en.wikipedia.org/wiki/Wide_area_information_server">http://en.wikipedia.org/wiki/Wide_area_information_server</a> )</div><div><br></div><div>For Lisp users it might be interesting to run numeric matrix operations in applications like Maxima on the GPU. Image understanding applications like Freedius ( <a href="http://www.ai.sri.com/project/FREEDIUS">http://www.ai.sri.com/project/FREEDIUS</a> ) could benefit from it. But there it would be probably written in the C/OpenCL side and used from Lisp via FFI. CAD applications, too. I could even imagine that a large Triple Store ( <a href="http://en.wikipedia.org/wiki/Triple_Store">http://en.wikipedia.org/wiki/Triple_Store</a> ) has algorithms in the query domain that could benefit from GPU / SIMD support. Also think about blackboards that have some dimensions as matrices mapped to the GPU (example of a blackboard system in Lisp: <a href="http://gbbopen.org/">http://gbbopen.org/</a> ).</div><div><br></div><div>As an example see this GBBOpen function: <span class="Apple-style-span" style="font-family: Times; font-size: 16px; font-weight: bold; -webkit-border-horizontal-spacing: 2px; -webkit-border-vertical-spacing: 2px; ">map-instances-on-space-instances</span></div><div><br></div><div> <a href="http://gbbopen.org/hyperdoc/ref-map-instances-on-space-instances.html">http://gbbopen.org/hyperdoc/ref-map-instances-on-space-instances.html</a></div><div><br></div><div> ' <span class="Apple-style-span" style="font-family: Times; font-size: 16px; ">Apply a function once to each unit instance on space instances, optionally selected by a retrieval pattern. '</span></div><div><font class="Apple-style-span" face="Times" size="4"><span class="Apple-style-span" style="font-size: 16px;"><br></span></font></div><div><font class="Apple-style-span" face="Times" size="4"><span class="Apple-style-span" style="font-size: 16px;">Then see FIND-INSTANCES: <span class="Apple-style-span" style="font-family: Helvetica; font-size: 14px; "><a href="http://gbbopen.org/hyperdoc/ref-find-instances.html">http://gbbopen.org/hyperdoc/ref-find-instances.html</a></span></span></font></div><div><br></div><div>I could imagine that SOME uses of these function could be speed up a lot by running in parallel on a GPU with, say, 256 compute elements.</div><div><br></div><div>But that is just speculation on my side. It really depends if users really have such application problems and they can be mapped to GPUs.</div><div><br></div><div>Regards,</div><div><br></div><div>Rainer Joswig</div><div><br></div><div><br><blockquote type="cite"><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div><br></div><div>Paul</div><div><br><div><div>On May 20, 2009, at 1:06 PM, Dan Weinreb wrote:</div><br class="Apple-interchange-newline"><blockquote type="cite"> <div text="#000000" bgcolor="#ffffff"> The instruction set is very restricted, and the communication<br> paths aren't there, as you suggested. GPGPU is especially<br> good for highly compute-intensive operations over not<br> all that much data. An FFT is an obvious example but<br> there are many, many good examples. (Not that I'm an<br> expert, but I do know that much.)<br> <br> There are CUDA-compatible devices that don't even<br> have a video connection, i.e. for GPGPU only.<br> The NVidia Tesla, called a "computing processor"<br> (weird name). 240 cores per board, and you can<br> chain together four of them.<br> <br> (My officemates are getting this info and telling to<br> me faster than I can type it in. Thanks, Andrew<br> and Scott.)<br> <br> -- Dan<br> <br> Jeremy Jones wrote: <blockquote type="cite" cite="mid:fcd1963f0905200635o23b60bb5k45866dbcb8b3adbb@mail.gmail.com"> <pre wrap="">On Wed, May 20, 2009 at 9:13 AM, Raffael Cavallaro
<a href="mailto:raffaelcavallaro@mac.com" class="moz-txt-link-rfc2396E"><raffaelcavallaro@mac.com></a> wrote:
</pre> <blockquote type="cite"> <pre wrap="">tomshardware.com ran this a couple of days ago:
<a href="http://www.tomshardware.com/reviews/nvidia-cuda-gpgpu,2299.html" class="moz-txt-link-rfc2396E"><http://www.tomshardware.com/reviews/nvidia-cuda-gpgpu,2299.html></a>
It's a summary of real-world results from apps using Nvidia's CUDA.
For certain things, like video encoding, they're seeing a 4x speedup
using the GPU over using the CPU. In addition, when they use the GPU,
it leaves the CPU free for other tasks.
</pre> </blockquote> <pre wrap=""><!---->
Why don't we just throw out the main CPU and fill our computers with
graphics cards? (Once CCL is ported to GPUs of course)
Seriously though, what does a CPU have that a GPU doesn't, besides a
different instruction set? More memory? Better i/o? Is the GPU
instruction set too specialized? I bet the answer is mainly software,
like OSes and device drivers. I remember in the old days it was
common to have a separate processor to handle i/o. Maybe that's what
the main CPU should be relegated to. OTOH, if the software is good
enough, it should just be distributed to whatever computing resources
are appropriate and available. Just thinking out loud.
_______________________________________________
Openmcl-devel mailing list
<a href="mailto:Openmcl-devel@clozure.com" class="moz-txt-link-abbreviated">Openmcl-devel@clozure.com</a>
<a href="http://clozure.com/mailman/listinfo/openmcl-devel" class="moz-txt-link-freetext">http://clozure.com/mailman/listinfo/openmcl-devel</a>
</pre> </blockquote> </div> _______________________________________________<br>Openmcl-devel mailing list<br><a href="mailto:Openmcl-devel@clozure.com">Openmcl-devel@clozure.com</a><br><a href="http://clozure.com/mailman/listinfo/openmcl-devel">http://clozure.com/mailman/listinfo/openmcl-devel</a><br></blockquote></div><br></div></div>_______________________________________________<br>Openmcl-devel mailing list<br><a href="mailto:Openmcl-devel@clozure.com">Openmcl-devel@clozure.com</a><br>http://clozure.com/mailman/listinfo/openmcl-devel<br></blockquote></div><br><div apple-content-edited="true"> <div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div style="word-wrap: break-word; -webkit-nbsp-mode: space; -webkit-line-break: after-white-space; "><div>Rainer Joswig, Hamburg, Germany</div><div><a href="http://lispm.dyndns.org/">http://lispm.dyndns.org/</a></div><div><a href="mailto:joswig@lisp.de">mailto:joswig@lisp.de</a></div><div><br></div></div></div></div></div></div></div></div></div></div></div><br class="Apple-interchange-newline"> </div><br></body></html>