toad.social is one of the many independent Mastodon servers you can use to participate in the fediverse.
Mastodon server operated by David Troy, a tech pioneer and investigative journalist addressing threats to democracy. Thoughtful participation and discussion welcome.

Administered by:

Server stats:

279
active users

#rocm

0 posts0 participants0 posts today
Replied in thread

@eugenialoli same with photo #raw processing, non of the #foss apps use the color profiles of the cameras, no 16bit raw and many other issues. I gave up retouch due to the still bad gimp ux and the bad implementation of non destructive editing.

The sad reality is the #linuxdesktop is not ready for professional #mediaproduction and this is such a bad thing in times like this.

#davinciresolve also barely runs on everything except nvidia on #linux and still has no #flatpak

Even blender is a pain with amd #rocm

Any #Linux #kernel ,
#graphics or #GPU people out there?

I'm trying to understand the relationship between the #amgdpu driver shipped with the kernel; and the "andgpu-dkms" driver that comes with #ROCm .

Specifically, with a recent enough kernel, do we really need to install the ROCm version of the driver? Does the ROCm version contain stuff the general driver does not? Or is the ROCm stack (esp. libhsa) tightly tied to a very specific version of the driver?

#AMD splits #ROCm toolkit into two parts – ROCm #AMDGPU drivers get their own branch under Instinct #datacenter #GPU moniker
The new #datacenter Instinct driver is a renamed version of the #Linux AMDGPU driver packages that are already distributed and documented with ROCm. Previously, everything related to ROCm (including the amdgpu driver) existed as part of the ROCm software stack.
tomshardware.com/pc-components

Tom's Hardware · AMD splits ROCm toolkit into two parts – ROCm AMDGPU drivers get their own branch under Instinct datacenter GPU monikerBy Aaron Klotz
Replied in thread

Even now, Thrust as a dependency is one of the main reason why we have a #CUDA backend, a #HIP / #ROCm backend and a pure #CPU backend in #GPUSPH, but not a #SYCL or #OneAPI backend (which would allow us to extend hardware support to #Intel GPUs). <doi.org/10.1002/cpe.8313>

This is also one of the reason why we implemented our own #BLAS routines when we introduced the semi-implicit integrator. A side-effect of this choice is that it allowed us to develop the improved #BiCGSTAB that I've had the opportunity to mention before <doi.org/10.1016/j.jcp.2022.111>. Sometimes I do wonder if it would be appropriate to “excorporate” it into its own library for general use, since it's something that would benefit others. OTOH, this one was developed specifically for GPUSPH and it's tightly integrated with the rest of it (including its support for multi-GPU), and refactoring to turn it into a library like cuBLAS is

a. too much effort
b. probably not worth it.

Again, following @eniko's original thread, it's really not that hard to roll your own, and probably less time consuming than trying to wrangle your way through an API that may or may not fit your needs.

6/

AI rabbit hole ... I've been playing with Ollama and some stability diffusion tools on my MacBook Pro M2 Max and my Linux desktop ... the desktop is way faster and only has an RX6800 in it, so of course I'm now thinking about an Rx7900XTX ... (I don't do Nvidia cards) ...

Anyone have experience with this upgrade? Is going from 16gb of VRAM to 24gb going to make a massive difference?

Using radeontop I can see it's using all 16gb at some points, but not consistently ... and I'm not sure if that's an issue or a feature. I believe #rocm still has some issues.

Just how deep is #Nvidia's #CUDA moat really?
Not as impenetrable as you might think, but still more than Intel or AMD would like
It's not enough just to build a competitive part: you also have to have #software that can harness all those #FLOPS — something Nvidia has spent the better part of two decades building with its CUDA runtime, while competing frameworks for low-level #GPU #programming are far less mature like AMD's #ROCm or Intel's #OneAPI.
theregister.com/2024/12/17/nvi #developers

The Register · Just how deep is Nvidia's CUDA moat really?By Tobias Mann
Continued thread

Even better, in the afternoon I managed to find a workaround for my #GPGPU software building but hanging when trying to run it, which seems to be related to an issue with some versions of the #AMD software stack and many integrated GPUs, not just the #SteamDeck specifically. So exporting the HSA_ENABLE_SDMA=0 environment vriable was sufficient to get my software running again. I'm dropping the information here in case others find it useful.

#ROCm #GPU #APU #HIP

2/2

Replied in thread

One of the nice things of the refactoring that I had to do to introduce CPU support is that it also allowed me to trivially had support for #AMD #HIP / #ROCm.
That, and the fact that AMD engineers have written a drop-in replacement for the Thrust library that we depend on in a couple of places. (This is also one of the things that is holding back a full #SYCL port for #GPUSPH, BTW.)

#ZLUDA Takes On Third Life: #OpenSource Multi-GPU #CUDA Implementation Focused On AI
ZLUDA is being rebuilt to focus on multi-GPU vendor support and will take a particular emphasis on #machinelearning / #AI type workloads. Previously ZLUDA was more focused on enabling professional creator workloads while now it will be more focused on CUDA-based AI/#ML software. The new ZLUDA code will be focused on #RDNA1 and newer support along with #ROCm 6.1+ compute stack support.
phoronix.com/news/ZLUDA-Third-

#AMD asks #developer to take down #ZLUDA, dev vows to rebuild
Earlier this year, AMD quietly stopped funding ZLUDA, #opensource #CUDA #translationlayer project allowed programscompiled for #Nvidia #CUDA #GPU to run on on Radeon GPU with the #ROCm software stack.
"The code that was previously here has been taken down at AMD's request," the developer wrote. "The code was released with AMD's approval through an email. AMD's legal department now says it's not legally binding"
tomshardware.com/pc-components

#AMD Releases #ROCm 6.2 With New Components
ROCm 6.2 is a big update with several new software components, improving the existing #PyTorch and #TensorFlow support, and a variety of other enhancements as AMD works to better compete with #NVIDIA's #CUDA.
ROCm 6.2 also introduces an offline installer to help those running ROCm on systems without an active Internet connection. ROCm 6.2 is also the first release officially supporting #Ubuntu 24.04 LTS.
phoronix.com/news/AMD-ROCm-6.2