It all depends on the size of the model you are running, if it cannot fit in GPU memory, then it has to go back and forth with the host (cpu memory or even disk) and the GPU. This is extremely slow. This is why some people are running LLMs on macs, as they can have a large amount of memory shared between the GPU and CPU, making it viable to fit some larger models in memory.
- 1 Post
- 5 Comments
Joined 6 months ago
Cake day: February 17th, 2025
You are not logged in. If you use a Fediverse account that is able to follow users, you can follow this user.
Thanks for sharing, I find it hard to discover new “lemmy spaces”??? On here
mlflexer@lemm.eeto Programming@programming.dev•Ubuntu 25.10 Looks To Make Use Of Rust Coreutils & Other Rust System ComponentsEnglish10·5 months agoThey don’t seem to have 100% pass of the tests, but I might be missing something?
Would love to take the jump, but I think I’ll wait until they pass all tests
You should only need to have Java and then download the server and open the port if they want to play vanilla mc. If they want modded then idk.
You might also want to check this out, haven’t used it myself but it looks cool if you don’t like wasting server resources: https://github.com/timvisee/lazymc
Oh, I thought you could get 128gb ram or more, but I can see it does not make sense with the <24gb… sorry for spreading misinformation, I guess, in this case a GPU of the same GB ram would probably be better