mastodontech.de ist einer von vielen unabhängigen Mastodon-Servern, mit dem du dich im Fediverse beteiligen kannst.
Offen für alle (über 16) und bereitgestellt von Markus'Blog

Serverstatistik:

1,4 Tsd.
aktive Profile

#fp64

0 Beiträge0 Beteiligte0 Beiträge heute
Dr. Moritz Lehmann<p>Interesting take from an <a href="https://mast.hpc.social/tags/Nvidia" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Nvidia</span></a> engineer I met at <a href="https://mast.hpc.social/tags/ISC25" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ISC25</span></a>: "Do you need IEEE-754 compliant <a href="https://mast.hpc.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a>, or do you need digits? Digits we can get you through FP64 emulation."</p><p>Not sure what to make of that. Things were an absolute mess before the IEEE-754 standard, and I wouldn't want to ever go back to that. No standards means you cannot at all port software between hardware architectures even from within the same vendor. Having to re-architect software for each and every new chip is not gonna happen.</p>
Dr. Moritz Lehmann<p>With all the hot discussions about <a href="https://mast.hpc.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> right now, what do you say to Nvidia almost entirely axing FP64 on <a href="https://mast.hpc.social/tags/Blackwell" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Blackwell</span></a> Ultra B300?</p><p>B200 180GB FP64 performance: 37.2 TFlops/s<br>B300 288GB FP64 performance: 1.2 TFlops/s ☠️</p><p>They're going from the usual datacenter <a href="https://mast.hpc.social/tags/GPU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPU</span></a> FP64:FP32 ratio of 1:2 down to cheap 1:64, like on all their gaming/workstation GPUs.<br>Personally I think this is fantastic - it opens the doors wide for competitors to step into the <a href="https://mast.hpc.social/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a> market.</p><p><a href="https://www.nvidia.com/en-us/data-center/gb300-nvl72/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">nvidia.com/en-us/data-center/g</span><span class="invisible">b300-nvl72/</span></a></p>
Andrew Jones (hpcnotes)<p>Provocative question to debate at <a href="https://mast.hpc.social/tags/ISC25" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ISC25</span></a> - in talks, meetings, booths, over beers:</p><p>What is so special about <a href="https://mast.hpc.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> that science holds it as a gold standard?</p><p>If <a href="https://mast.hpc.social/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a> had never got as far as FP64, eg had got stuck at FP32, surely we would still have been able to do science?</p>
Benjamin Carr, Ph.D. 👨🏻‍💻🧬<p><a href="https://hachyderm.io/tags/JackDongarra" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>JackDongarra</span></a> Makes a Stand for Traditional <a href="https://hachyderm.io/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a>: "US still doesn’t have a clear, long-term plan for what comes next.... U.S. risks falling behind."</p><p>Challenges to high-performance computing threaten <a href="https://hachyderm.io/tags/US" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>US</span></a> <a href="https://hachyderm.io/tags/innovation" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>innovation</span></a></p><p>The <a href="https://hachyderm.io/tags/AI" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AI</span></a> boom has led chip makers to focus on <a href="https://hachyderm.io/tags/FP16" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP16</span></a> and <a href="https://hachyderm.io/tags/FP8" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP8</span></a>, not the <a href="https://hachyderm.io/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> used by scientific research. If chip companies stop making the parts that <a href="https://hachyderm.io/tags/scientists" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>scientists</span></a> need, then it could become harder to do important research. <br><a href="https://theconversation.com/challenges-to-high-performance-computing-threaten-us-innovation-255188" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">theconversation.com/challenges</span><span class="invisible">-to-high-performance-computing-threaten-us-innovation-255188</span></a></p>
Andrew Jones (hpcnotes)<p><a href="https://mast.hpc.social/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a> and other scientific computing folks should like this image ... it's a floating point 😀 </p><p><a href="https://mast.hpc.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> <a href="https://mast.hpc.social/tags/FLOPS" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FLOPS</span></a></p>
Benjamin Carr, Ph.D. 👨🏻‍💻🧬<p><a href="https://hachyderm.io/tags/ElCapitan" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>ElCapitan</span></a> Towers Above the <a href="https://hachyderm.io/tags/Top500" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Top500</span></a> in a Big <a href="https://hachyderm.io/tags/HPE" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPE</span></a> Win using <a href="https://hachyderm.io/tags/AMD" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>AMD</span></a> Instinct <a href="https://hachyderm.io/tags/MI300A" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>MI300A</span></a> system. With sustained <a href="https://hachyderm.io/tags/HPL" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPL</span></a> of 1.742EF and peak speed of 2.79EF of <a href="https://hachyderm.io/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a>, this is a big jump over previous generation systems. El Capitan has a unique architecture as it uses <a href="https://hachyderm.io/tags/APU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>APU</span></a> combining <a href="https://hachyderm.io/tags/CPU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>CPU</span></a> plus <a href="https://hachyderm.io/tags/GPU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPU</span></a> onto a package with high-bandwidth memory. Over 44,000 of these MI300A APUs are then packed into the HPE <a href="https://hachyderm.io/tags/Cray" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Cray</span></a> Shasta liquid cooled platform, and connected via the Slingshot interconnect. <br><a href="https://www.servethehome.com/el-capitan-towers-above-the-top500-in-a-big-hpe-and-amd-win/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">servethehome.com/el-capitan-to</span><span class="invisible">wers-above-the-top500-in-a-big-hpe-and-amd-win/</span></a></p>
Benjamin Carr, Ph.D. 👨🏻‍💻🧬<p><a href="https://hachyderm.io/tags/HPE" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPE</span></a> goes Cray for <a href="https://hachyderm.io/tags/Nvidia" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Nvidia</span></a>'s Blackwell <a href="https://hachyderm.io/tags/GPU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPU</span></a>, crams 224 into a single cabinet<br>Announced ahead of <a href="https://hachyderm.io/tags/SuperComputing" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>SuperComputing</span></a> conference in Atlanta, Cray's <a href="https://hachyderm.io/tags/EX154n" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>EX154n</span></a> platform will support up to 224 Nvidia Blackwell GPUs and 8,064 Grace CPU cores per cabinet. That works out to just over 10 petaFLOPS at <a href="https://hachyderm.io/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> for <a href="https://hachyderm.io/tags/HPC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPC</span></a> applications or over 4.4 exaFLOPS FP64 for sparse AI and machine learning workloads <br>HPE's new ProLiant servers offer choice of Gaudi, Hopper, or Instinct acceleration<br><a href="https://www.theregister.com/2024/11/13/hpe_cray_ex/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">theregister.com/2024/11/13/hpe</span><span class="invisible">_cray_ex/</span></a> <a href="https://hachyderm.io/tags/sc24" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>sc24</span></a></p>
Benjamin Carr, Ph.D. 👨🏻‍💻🧬<p>Happy <a href="https://hachyderm.io/tags/Exascale" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Exascale</span></a> day to all who celebrate and the companies that use it as marketing</p><p>Doug Eadline has a wonderful piece in HPCWire about how not all exa are equivalent, and the push to brand things with lower precision as <a href="https://hachyderm.io/tags/zetascale" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>zetascale</span></a> flies in the face of 30 years of agreeing that FLOPS are <a href="https://hachyderm.io/tags/fp64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>fp64</span></a>, so there is one and only one definition of <a href="https://hachyderm.io/tags/exaFLOPS" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>exaFLOPS</span></a>, and its bases on the <a href="https://hachyderm.io/tags/HPLinpack" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>HPLinpack</span></a> benchmark.</p><p>Go Read; In This Club, You Must 'Earn the Exa'<br><a href="https://www.hpcwire.com/2024/10/17/in-this-club-you-must-earn-the-exa/" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">hpcwire.com/2024/10/17/in-this</span><span class="invisible">-club-you-must-earn-the-exa/</span></a></p>
Andrew Jones (hpcnotes)<p>Experiment comparing reactions across social media platforms ...</p><p>One or more of <a href="https://mast.hpc.social/tags/Fortran" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Fortran</span></a> or <a href="https://mast.hpc.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> or <a href="https://mast.hpc.social/tags/onprem" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>onprem</span></a> <a href="https://mast.hpc.social/tags/hpc" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>hpc</span></a> will be obsolete by 2030.</p><p>Ignoring this message counts as agreeing :-)</p>
Rechenkraft.net<p><a href="https://mstdn.social/tags/PYSCFbeta" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>PYSCFbeta</span></a>: Bei <a href="https://mstdn.social/tags/GPUGRID" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPUGRID</span></a> gibt es eine neue <a href="https://mstdn.social/tags/Linux" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Linux</span></a> (beta) App für <a href="https://mstdn.social/tags/NVIDIA" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>NVIDIA</span></a> <a href="https://mstdn.social/tags/GPU" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>GPU</span></a>, die aufgrund von "quantum chemistry" Berechnungen <a href="https://mstdn.social/tags/FP64" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>FP64</span></a> (double precision) nutzt:<br><a href="https://www.gpugrid.net/forum_thread.php?id=5429#60963" rel="nofollow noopener" translate="no" target="_blank"><span class="invisible">https://www.</span><span class="ellipsis">gpugrid.net/forum_thread.php?i</span><span class="invisible">d=5429#60963</span></a><br><a href="https://mstdn.social/tags/BOINC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>BOINC</span></a> <a href="https://mstdn.social/tags/DC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DC</span></a> <a href="https://mstdn.social/tags/VC" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>VC</span></a> <a href="https://mstdn.social/tags/VolunteerComputing" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>VolunteerComputing</span></a> <a href="https://mstdn.social/tags/DistributedComputing" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>DistributedComputing</span></a> <a href="https://mstdn.social/tags/CitSci" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>CitSci</span></a> <a href="https://mstdn.social/tags/CitizenScience" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>CitizenScience</span></a> <a href="https://mstdn.social/tags/Rechenkraft" class="mention hashtag" rel="nofollow noopener" target="_blank">#<span>Rechenkraft</span></a></p>