Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)MO
Posts
18
Comments
433
Joined
2 yr. ago

  • You're right, my bad. Dynamic linking and dynamic compilation are different thinks.

    The library inter operation is a part of the translation layers that, like fex-emu which is becoming more and more supported by Fedora.

    https://github.com/FEX-Emu/FEX/blob/main/ThunkLibs/README.md

    manually vetted libraries where you can clean up the ABI

    Yes, but usually games are ran with wine which does have a standard set of libraries it uses.

  • maybe it’s not as big of a thing as I imagine it being.

    Yes, see my other comments in this thread for an explanation of this. The trick is that not all the calls are translated, as wine is able to use the arm version of the libraries rather than the x86 version.

  • Should be awful for gaming. It’s possible to run x86 things with emulation, sure, but performance (especially single-thread)

    Most modern software (games excluded), is dynamically compiled. This means that it’s not all one “bundle” that runs, but rather a binary that calls reusable pieces of code, “libraries” from the binary itself. Wine is dynamically compiled.

    What makes modern x86 to arm translators special, is that the x86 binary, like an x86 version of wine, can call upon the arm versions of the libraries it uses ­— like graphic drivers. It’s because of this that the people on r/emulationonandroid managed to play GTA 5 with 30 fps via the computer version. There definitely is overhead, but it’s not that much, and a beefy machine like this could absolutely handle it.

    https://moonpiedumplings.github.io/blog/scale-22/#exhibition-hall

    The Facebook/Meta table had a booth where they had an ARM macbook that was running steam and they were installing games on it.

  • ut I honestly doubt ARM can with the overhead of emulation

    Most modern software (games excluded), is dynamically compiled. This means that it's not all one "bundle" that runs, but rather a binary that calls reusable pieces of code, "libraries" from the binary itself. Wine is dynamically compiled.

    What makes modern x86 to arm translators special, is that the x86 binary, like an x86 version of wine, can call upon the arm versions of the libraries it uses ­— like graphic drivers. It's because of this that the people on r/emulationonandroid managed to play GTA 5 with 30 fps via the computer version. There definitely is overhead, but it's not that much, and a beefy machine like this could absolutely handle it.

    https://moonpiedumplings.github.io/blog/scale-22/#exhibition-hall

    The Facebook/Meta table had a booth where they had an ARM macbook that was running steam and they were installing games on it.

  • I never got uefi images booting properly on those grub multi boot utility drives. Granted the last time I bothered with it was like 10 years ago now

    I haven't had any issues with Ventoy, everything I've attempted to boot works. Doesn't matter how it does it if it works.

  • Right, but you could have just made one yourself

    And then there would be a bus factor of one. It's not just about making a helm chart for myself, it's about having something that can be shared with the community, that doesn't depend on any single person to be maintained and updated.

    It's about having an organization that provides "packages" for Kubernetes, for people/orgs that don't have the time, expertise, and energy to maintain them.

    I greatly respect Ananace, who is in the comments of this post, and mentioned their Helm charts. The work is excellent. But looking through the commits, it's just one person, doing something that primarily consists of bumping version numbers. Contrast this to the Matrix ESS helm chart, where the commits consist of many more contributors, and also include feature additions to the helm chart.

  • Hello Ananace! :)

    I actually have seen your helm charts many, many times before when searching for matrix, synapse, or lemmy on Artifacthub.

    An official helm chart isn't really a hard requirement to me, even if I were to use one and it were to stop getting maintained, I could continue on my own. But an official helm chart has big community benefits that are very important to me. Like, there becomes the option of paid support, which is a must have for many entities. Also, an official organization may support a wider variety of usecases than someone making helm charts for personal use.

    I also ended up chatting with one of the core devs of Synapse about ways to improve regular Python Synapse for use with Kubernetes back in the ending of January, so hopefully it’ll improve in that direction when time allows

    Do you know anything about the claims that they have rewritten synapse in rust?

  • Yes and no. There are many things that are much easier with Kubernetes, once you figure Kubernetes out.

    High availability is the most notable example — yes, it's doable in docker, via something like swarm, but it's more difficult. In comparison, the ideas of clustering and working with more than one server are central to the architecture of Kubernetes.

    Another thing is that long term deployments with Kubernetes can be more maintainable, since everything is just yaml files and version is just a number. If you store your config in code, then it's easier to replicate it to another server, either internally, or if you share it for other people to use (Helm is somewhat like this).

  • This helm chart is not just matrix/synapse, but also element (web ui), and "matrix authentication service", which adds SSO/OIDC support to a normal synapse instance, which is pretty neat. I haven't seen any helm charts that include the full matrix stack, just separate synapse or element helm charts. And helm definitely makes deploying services to Kubernetes easier than other ways of deploying applications.

    The other reason why I like an official helm chart, is because I have seen unofficial one's be stopped being maintained by the community member(s) maintaining them. With an official one, it will (probably) be maintained indefinitely.

  • Licenses like SSPLv1

    The SSPL requires that all software used to deploy SSPL software is open sourced. If I deploy my software on Windows, do I have to provide the source code for Windows? What about the proprietary hardware drivers, or Intel Management Engine?

    The SSPL is not the next generation of licenses, it is effectively unusable. And both Redis and Mongo, dual licensed their software as the SSPL, and a proprietary license — effectively making their entire software proprietary.

    make money without risk of a cloud vendor snapping up their software (think Redis, MongoDB, etc) need a place at the table.

    Except Redis, and Mongo were making money. They had well valued, well earning SAAS offerings — it's just that the offerings integrated into existing cloud vendors would be more popular (because vendor lock in). They just wanted more money, and were hoping that by going proprietary, they could force customers away from the cloud offers to themselves, and massively increase their revenue.. They did not get that.

    Another thing is that it's not "stealing" Mongo/Redis' when cloud vendors offer SAAS's of Mongo/Redis. Mongo/Redis, and their SAAS offerings, are only possible because the same cloud vendors put more money than Mongo/Redis make yearly into Linux and other software that powers the SAAS offerings of Mongo/Redis, like Kubernetes. Without that software, Mongo/Redis wouldn't have a SAAS offering at all.

    I definitely think that it's bad when a piece of software doesn't get any funding it needs to develop, especially when it powers much more modern software, like XZ. But Mongo/Redis weren't suffering from a lack of funding at all. They're just mad they had to share their toys, and tried to take them away. But it didn't even matter in the end.

  • Which means my distro-morphing idea should work in theory with OpenStack

    I also don't recommend doing a manual install though, as it's extremely complex compared to automated deployment solutions like kolla-ansible (openstack in docker containers), openstack-ansible (host os/lxc containers), or openstack-helm/genestack/atmosphere (openstack on kubernetes). They make the install much more simpler and less time consuming, while still being intensely configurable.