Google was working on a feature that would do just that, but I can’t recall the name of it.
They backed down for now due to public outcry, but I expect they’re just biding their time.
Google was working on a feature that would do just that, but I can’t recall the name of it.
They backed down for now due to public outcry, but I expect they’re just biding their time.
Not with this announcement, but it was.
I’m also going to push forward Tilda, which has been my preferred one for a while due to how minimal the UI is.
We all mess up! I hope that helps - let me know if you see improvements!
I think there was a special process to get Nvidia working in WSL. Let me check… (I’m running natively on Linux, so my experience doing it with WSL is limited.)
https://docs.nvidia.com/cuda/wsl-user-guide/index.html - I’m sure you’ve followed this already, but according to this, it looks like you don’t want to install the Nvidia drivers, and only want to install the cuda-toolkit metapackage. I’d follow the instructions from that link closely.
You may also run into performance issues within WSL due to the virtual machine overhead.
Good luck! I’m definitely willing to spend a few minutes offering advice/double checking some configuration settings if things go awry again. Let me know how things go. :-)
It should be split between VRAM and regular RAM, at least if it’s a GGUF model. Maybe it’s not, and that’s what’s wrong?
Ok, so using my “older” 2070 Super, I was able to get a response from a 70B parameter model in 9-12 minutes. (Llama 3 in this case.)
I’m fairly certain that you’re using your CPU or having another issue. Would you like to try and debug your configuration together?
Unfortunately, I don’t expect it to remain free forever.
No offense intended, but are you sure it’s using your GPU? Twenty minutes is about how long my CPU-locked instance takes to run some 70B parameter models.
On my RTX 3060, I generally get responses in seconds.
It’s a W3C managed standard, but there are tons of behavior not spelled out in the specification that platforms can choose to impose.
The standard doesn’t impose a 500 character limit, but there’s nothing that says there can’t be a limit.
I mean, sysvinit was just a bunch of root-executed bash scripts. I’m not sure if systemd is really much worse.
Systemd was created to allow parallel initialization, which other init systems lacked. If you want proof that one processor core is slower than one + n, you don’t need to compare init systems to do that.
I bought a used 2018 model over a new current model because of the lack of physical function keys.
Also, Dell, bring back Fn + Left for Home and Fn + Right for End!
Who looked at a great keyboard layout and decided, “I know! I’ll make this Developer Edition hardware more difficult to develop on!”
PNG support lossless compression through deflation, but there are encoders that can apply a lossy filter to the image to make the compression more effective.
PNG doesn’t support lossy compression natively, to be clear.
Rust specializes in making parallel processing secure and approachable, so it’s going get used in problems where parallel processing and efficiency matter.
Rust is also now allowed to be used in the Linux kernel for the same reasons, which is exciting!
Sounds like a job for Ansible. ;-)
Of course! Here’s the documentation for the docker-compose module: https://docs.ansible.com/ansible/latest/collections/community/docker/docker_compose_module.html
It does rely on the recently-made-legacy docker-compose client, so for now it’s still required to install that. If you need some advice or pointers, let me know.
That depends on what you mean by container. I use it to orchestrate Docker containers for my infrastructure and then some.
Thank you! I was struggling to remember the proposal name.