I don't think it supports displaying HDR at all? The GitHub issue regarding HDR is still open and it definitely doesn't switch to HDR mode when I open HDR photos with it.
I do not know of any such dongle, but I'd like to ask you a question if you don't mind: are you looking for a dongle with open-source firmware, or would a dongle that has its (proprietary) firmware stored in some onboard memory be acceptable?
The second option wouldn't require you to install any proprietary firmware on your computer, but you'd still rely on the proprietary firmware for the device to run. And it might also exist, unlike a dongle with FOSS firmware.
Maybe developers will finally start implementing predictive back now that it's not hidden behind developer options. It's kinda nice when you can just peek at where the app wants to take you when you go back, and it currently ironically tends to be implemented only by apps that already have decently made navigation.
Also private space seems nice, finally a way to use the work profile sandbox natively without having to install third party apps that pretend to be work profile managers.
I know this isn't Reddit, but r/peopleliveincities... When 90% of desktop users use Windows, it's going to both be the most targeted by malware developers and have the highest chance of being operated by someone who doesn't understand enough about computers to recognize that the shiny calculator app that just popped up after visiting a very legit Nigerian prince's crowdfunding page probably shouldn't need admin access.
And speaking of user error, I'm willing to bet that basic security practices like using full disk encryption, SecureBoot, some MAC layer (provided by antivirus on Windows, AppArmor/SELinux on Linux) and regularly applying security updates are way more common over in the Windows land - if I was in a situation where there was one completely randomly selected Windows PC and one also completely randomly selected Linux PC, and my life depended on being able to gain access to either of them (some kind of really messed up Saw trap? idk), I would definitely bet my life on the Linux one being misconfigured.
Don't get me wrong, Linux can make for a very secure and private OS, but most installs most definitely cannot be described as such - just look at the popularity of random unverified PPAs on Ubuntu derivatives or AUR packages on Arch.
A reasonable build of the kernel optimized for virtualization won't take more than a few tens of megabytes of RAM (and it will have support for memory ballooning, so the virtualized kernel will give the memory it doesn't need back to the host), and the userspace will need to be separate anyway due to how different Android is to normal Linux distros.
Containers are nice when you want to run dozens of separate services on the same server or want to get the benefits of infrastructure as code, but in this case they would provide minimal benefits at the cost of having no way of loading any kernel modules not built into whatever ancient kernel version your SoC manufacturer decided you have to use on your phone. Also, container escape vulnerabilities are still a bit more common than full VM escape, so this is also good for security on top of being more useful.
The only app that doesn't auto-update for me is Fdroid itself (ironically), because it targets an old Android version. Running Android 14 on a Pixel, so with the strongest Google fuckery.
Are you sure your Fdroid client is up to date? The new API was implemented in 1.19, and apparently I even misremembered and all you have to do to enable Fdroid to auto update its apps is to manually update them for one last time (so no fresh installation required).
Another long shot: there's an option to force the old installation method hidden in expert settings - maybe you could check if that isn't enabled?
On a normal unmodified phone you have to manually confirm each app you want to install. so no auto-updates in the background etc.
Background app updates are possible since Android 12, Fdroid just took two years to implement the new API (and you have to do a fresh install of the apps - apps already installed using the old API still require confirmation on each update). There is still friction on the initial install though.
What error? It gave you a string of tokens that seemed likely according to its training data. That's all it does.
If you ask it what color is the sky, it will tell you it's blue not because it knows that's true, but because these words "fit together". Pretty much the only way to avoid this issue is to put some kind of filter in front of the LLM which will try to catch prompts that are known to produce unwanted results, and silently replace your prompt with something like "say: sorry, I don't know".
I'm being very reductive here, but that's the principle of how these things work - the LLMs are not capable of determining the truthfulness of their responses.
OK, cool. Just remember that the only entity who can sue in this situation is Microsoft (because when you contribute code to VS Code, you must sign a CLA that says you give Microsoft full perpetual rights to distribute your code under any license they wish - it is Microsoft who then "graciously" releases your code under a copy left license while also building their proprietary version of VS Code using it).
And Microsoft cannot use the code if it gets released under a copyleft license - that wouldn't allow them to build their proprietary build with it. So the only one who can do anything has less than zero (because it would improve only the FOSS forks, which are meant to be inferior) interest in making these guys publish the source code as proper FOSS.
No, they are just in violation of the original license. That doesn't mean they have to comply with it by properly open sourcing the project. Generally it's also OK to just delete everything.
There were plenty of cases where commercial software included open source stuff in a way that violated its license, and the accepted way to fix the license violation was for the software/hardware vendor to stop using the violated project going forward. Usually they don't even have to for example scrub old firmware downloads that improperly included FOSS bits.
Even on my home server (a desktop with 64 gigs of ram) the ram check takes longer than the OS.
I was pretty sure I messed something up when I upgraded the RAM in my desktop from 16 to 64 gigs and it wouldn't output any signal for solid 10 seconds, lol. And the regular 5 second black screen on normal boots was still something I had to get used to coming from maybe a second with 16 GB
Your argument is to have 2 subtly incompatible abis and one day binaries magically break.
Whereas your argument seems to be to have a special C variant for 32bit Linux - there's no reason to have a special time64_t anywhere else.
No program with time32_t will ever work after 2038, so any compiled that way are broken from compilation.
Yeah, so what will breaking the ABI do? Break it a bit more?
If you really want to be clever, mangle the symbols for the functions that handle time so they encode time64 as appropriate
That's what MUSL libc does, and the result is two subtly incompatible ABIs - statically linked programs are fine, but if a dynamically linked library exports any function with time_t parameter or return value, it will use whatever size was configured at build time and it becomes a part of its ABI. So fixing this properly would require every library that wants to pass time_t values in its API to implement its own name mangling. That's not a reasonable request for a barely used platform (remember, this is just 32bit userland, 64bit was always unaffected).
I don't think it supports displaying HDR at all? The GitHub issue regarding HDR is still open and it definitely doesn't switch to HDR mode when I open HDR photos with it.