Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)YL
Posts
480
Comments
46
Joined
2 yr. ago

Technology @lemmy.world

This smiling robot face made of living skin is absolute nightmare fuel

Gaming @lemmy.ml

Dr Disrespect Admits To 'Inappropriate' Messages With Minor: 'I'm No Fucking Predator Or Pedophile'

Gaming @lemmy.ml

Nearly 6 months later, Palworld devs confirm Nintendo never drew so much as an inch of its legal sword over bootleg Pokémon allegations

Gaming @lemmy.ml

Dr Disrespect fired by the game studio he co-founded: 'It is our duty to act with dignity on behalf of all individuals involved'

Gaming @lemmy.ml

Xbox executives just cannot give a straight answer to questions about Tango Gameworks

Gaming @lemmy.ml

How Embracer's cuts killed a potential Red Faction sequel and gutted a promising studio

Gaming @lemmy.ml

CD Projekt passes the 'Best DLC' crown to FromSoftware with Witcher-themed Elden Ring art

Gaming @lemmy.ml

Paradox Lays Off Entire Studio Before Its Game Was Even Released

Gaming @lemmy.ml

Play Your Way – enjoy your GOG games with Luna cloud streaming

Selfhosted @lemmy.world

A guide to passing GPUs through to Proxmox, XCP-ng VMs

Gaming @lemmy.ml

Sim game Life By You canceled after 3 early access delays

United States | News & Politics @lemmy.ml

New 'Washington Post' chiefs can’t shake their past in London

  • Three side remarks about China, which can be a peculiar example to compare to for Russia, maybe even any other country:

    • They actually banned consoles for a quite significant 15 years (2000–2015), which strongly tilted their market towards PC.
    • Their companies actively make PC-type gaming handhelds, and many of them are even well-established in the business ahead the current “Steam Deck” wave/bandwagon: GPD (once called GamePad Digital, first release in 2016), OneXPlayer (2020), Ayaneo (2021).
    • Chinese gaming companies are quite at the whim of the censorship, and occasional “crackdowns” out of the blue, and many have therefore reoriented themselves for an international audience to de-risk their business.
  • Linux @lemmy.ml

    Linux's New DRM Panic "Blue Screen of Death" In Action

    Technology @lemmy.world

    Russia continues work on homegrown game console despite technology and scale issues

    Gaming @lemmy.ml

    Russia continues work on homegrown game console despite technology and scale issues

    Gaming @lemmy.ml

    In 1989, a Nintendo bigwig licensed SimCity on the spot by sliding a million dollar check across the table

    Technology @lemmy.world

    The Increasing Impatience Of The Speed Of The PCI-Express Roadmap

    Gaming @lemmy.ml

    Starfield’s Paid Mods Ignite A Review Bombing On Steam

    Gaming @lemmy.ml

    A follow-up to the legendary Disco Elysium might have been ready to play within the next year⁠—ZA/UM's devs loved it, management canceled it and laid off the team

    Gaming @lemmy.ml

    Microsoft backs up Geoff Keighley after he ignited console warrior outrage over the Gears of War: E-Day trailer

  • Likely due to being a prototype. Production laptops from Tuxedo tend to have the “TUX” penguin in a circle logo on the Super key by default. They also have been offering custom engraved keyboard (even with the entire keyboard engraved from scratch to the customer’s specifications) as added service, so probably there will be suppliers or production facility to change the Super key.

    By the way, there was one YouTube channel that ended up ordering a laptop with Windings engraving from them: https://youtu.be/nidnvlt6lzw?t=186

  • How does this analogy work at all? LoRA is chosen by the modifier to be low ranked to accommodate some desktop/workstation memory constraint, not because the other weights are “very hard” to modify if you happens to have the necessary compute and I/O. The development in LoRA is also largely directed by storage reduction (hence not too many layers modified) and preservation of the generalizability (since training generalizable models is hard). The Kronecker product versions, in particular, has been first developed in the context of federated learning, and not for desktop/workstation fine-tuning (also LoRA is fully capable of modifying all weights, it is rather a technique to do it in a correlated fashion to reduce the size of the gradient update). And much development of LoRA happened in the context of otherwise fully open datasets (e.g. LAION), that are just not manageable in desktop/workstation settings.

    This narrow perspective of “source” is taking away the actual usefulness of compute/training here. Datasets from e.g. LAION to Common Crawl have been available for some time, along with training code (sometimes independently reproduced) for the Imagen diffusion model or GPT. It is only when e.g. GPT-J came along that somebody invested into the compute (including how to scale it to their specific cluster) that the result became useful.

  • This is a very shallow analogy. Fine-tuning is rather the standard technical approach to reduce compute, even if you have access to the code and all training data. Hence there has always been a rich and established ecosystem for fine-tuning, regardless of “source.” Patching closed-source binaries is not the standard approach, since compilation is far less computational intensive than today’s large scale training.

    Java byte codes are a far fetched example. JVM does assume a specific architecture that is particular to the CPU-dominant world when it was developed, and Java byte codes cannot be trivially executed (efficiently) on a GPU or FPGA, for instance.

    And by the way, the issue of weight portability is far more relevant than the forced comparison to (simple) code can accomplish. Usually today’s large scale training code is very unique to a particular cluster (or TPU, WSE), as opposed to the resulting weight. Even if you got hold of somebody’s training code, you often have to reinvent the wheel to scale it to your own particular compute hardware, interconnect, I/O pipeline, etc.. This is not commodity open source on your home PC or workstation.

  • The situation is somewhat different and nuanced. With weights there are tools for fine-tuning, LoRA/LoHa, PEFT, etc., which presents a different situation as with binaries for programs. You can see that despite e.g. LLaMA being “compiled”, others can significantly use it to make models that surpass the previous iteration (see e.g. recently WizardLM 2 in relation to LLaMA 2). Weights are also to a much larger degree architecturally independent than binaries (you can usually cross train/inference on GPU, Google TPU, Cerebras WSE, etc. with the same weights).

  • There is even a sentence in README.md that makes it explicit:

    The source files in this repo are for historical reference and will be kept static, so please don’t send Pull Requests suggesting any modifications to the source files […]

  • GIMP is a special case. GIMP is being getting outdeveloped by Krita these days. E.g.:

    https://gitlab.gnome.org/GNOME/gimp/-/issues/9284

    Or compare with:

    https://www.phoronix.com/news/Krita-2024-GPUs-AI

    GIMP had its share of self inflicted wounds starting with a toxic mailing list that drove away people from professional VFX and surrounding FilmGimp/CinePaint. When the GIMP people subsequently took over the GEGL development from Rhythm & Hues, it took literally 15 years until it barely worked.

    Now we are past the era of simple GPU processing into diffusion models/“generative AI” and GIMP is barely keeping up with simple GPU processing (like resizing, see above).

  • From my own statistics how many I feel worthy posting/linking on Lemmy, the most direct alternative to Kotaku is Eurogamer. PCGamer, PCGamesN and Rock Paper Shotgun are occasionally OK, but you have to cut through a lot of spam and clickbait (i.e. exactly this “50 guides per week” type of corporate guidance). Not sure if this is also the state that Kotaku will end up in. The Verge sometimes also have good articles, but the flood of gadget consumerism articles there is obnoxious.

  • The PS Vita side of Sony customer has gotten a deep taste of Sony’s issues of catering everything to a singular console. And same with PSVR2: Of course it must be PS5 exclusive, because everything are adornments towards their shiny console — and went on to not sell a lot of PS5.

  • Well, if you have a constructive suggestion which site to link instead regarding kernel developments, I am all ears:

    • Not sure that raw commits are readable or have sufficient context for non kernel development readers here
    • LWN, particularly timely/kernel development news there, has gone mostly paywall, and there will be (legitimate) complaint if I link articles needing a LWN subscription
  • Not sure what called for this blatant personal attack. My post history speaks for itself, quite in comparison to yours. And Phoronix is well-known Linux website, and its test suite is in fact even referenced in various regression tests/patches in LKML (also not sure what/if any kind of kernel development you have done).