Getting Keycloak and Headscale working together.
But I did it after three weeks.
I captured my efforts in a set of interdependent Ansible roles so I never have to do it again.
Getting Keycloak and Headscale working together.
But I did it after three weeks.
I captured my efforts in a set of interdependent Ansible roles so I never have to do it again.
I still have it on my Pixel 8 Pro. It requires a double tap to occur in less than 300 milliseconds.
This bug has been the bane of my existence for almost four years now: https://issuetracker.google.com/issues/204650736
The Android version of the app still has the zoom/cursor offset bug when using a software keyboard from when they sunset RDP 8. That has been a severe usability bug for over three years now.
It’ll probably be there, but at least it can be disabled in the settings now. It won’t go away on its own.
This was when I stopped using it for a while. I sent multiple feedback messages as it really irritated me.
I consider it a big deal. I’m clicking “Not Now” buttons all day when I just want to use a piece of software for its main purpose. And then because it says “Not Now” I get asked again and again and again.
Thank you! I was struggling to remember the proposal name.
Google was working on a feature that would do just that, but I can’t recall the name of it.
They backed down for now due to public outcry, but I expect they’re just biding their time.
Not with this announcement, but it was.
I’m also going to push forward Tilda, which has been my preferred one for a while due to how minimal the UI is.
Pixel Experience is unfortunately dead now. 🙁
We all mess up! I hope that helps - let me know if you see improvements!
I think there was a special process to get Nvidia working in WSL. Let me check… (I’m running natively on Linux, so my experience doing it with WSL is limited.)
https://docs.nvidia.com/cuda/wsl-user-guide/index.html - I’m sure you’ve followed this already, but according to this, it looks like you don’t want to install the Nvidia drivers, and only want to install the cuda-toolkit metapackage. I’d follow the instructions from that link closely.
You may also run into performance issues within WSL due to the virtual machine overhead.
Good luck! I’m definitely willing to spend a few minutes offering advice/double checking some configuration settings if things go awry again. Let me know how things go. :-)
It should be split between VRAM and regular RAM, at least if it’s a GGUF model. Maybe it’s not, and that’s what’s wrong?
Ok, so using my “older” 2070 Super, I was able to get a response from a 70B parameter model in 9-12 minutes. (Llama 3 in this case.)
I’m fairly certain that you’re using your CPU or having another issue. Would you like to try and debug your configuration together?
Unfortunately, I don’t expect it to remain free forever.
No offense intended, but are you sure it’s using your GPU? Twenty minutes is about how long my CPU-locked instance takes to run some 70B parameter models.
On my RTX 3060, I generally get responses in seconds.
I’ve had this issue on several Pixel devices - I most recently had this exact issue on my Pixel 6 Pro, but I think recent-ish Firefox updates resolved some of the issues.
In my case, Firefox would lockup, but I could easily switch to another app without an issue. I would have to kill the Firefox app to get it responsive again.