Skip Navigation

Posts
1
Comments
224
Joined
2 yr. ago

  • If anything, my take home message is that the reach of copyright law is too long and needs to be taken down a peg.

    Exactly! Copyright law is terrible. We need to hold AI companies to the same standard that everyone else is held. Then we might actually get big corporations lobbying to improve copyright law for once. Giving them a free pass right now would be a terrible waste of an opportunity in addition to being an injustice.

  • First time I've heard of Mojeek. Why should I trust it more than any other company? Is there anything particular about its economic model or governance that makes it less likely to decide to be unethical?

  • AI companies will probably get a free pass to ignore robots.txt even if it were enforced by law. That's what they're trying to do with copyright and it looks likely that they'll get away with it.

  • The general public doesn't have to understand anything about how it works as long as they get a clear "verified by ..." statement in the UI.

  • lsblk is just lacking a lot of information and creating a false impression of what is happening. I did a bind mount to try it out.

     bash
        
    sudo mount -o ro --bind /var/log /mnt
    
      

    This mounts /var/log to /mnt without making any other changes. My root partition is still mounted at / and fully functional. However, all that lsblk shows under MOUNTPOINTS is /mnt. There is no indication that it's just /var/log that is mounted and not the entire root partition. There is also no mention at all of /. findmnt shows this correctly. Omitting all irrelevant info, I get:

     
        
    TARGET                                                SOURCE                 [...]
    /                                                     /dev/dm-0              [...]
    [...]
    └─/mnt                                                /dev/dm-0[/var/log]    [...]
    
      

    Here you can see that the same device is used for both mountpoints and that it's just /var/log that is mounted at /mnt.

    Snap is probably doing something similar. It is mounting a specific directory into the directory of the firefox snap. It is not using your entire root partition and it's not doing something that would break the / mountpoint. This by itself should cause no issues at all. You can see in the issue you linked as well that the fix to their boot issue was something completely irrelevant.

  • Essentially ULWGL will allow you to run your non-steam games using Proton, Proton-GE, or other Proton forks using the same pressure vessel containerization and runtime that Valve use to run games with Proton

    This is the crucial piece of information. In less technical terms: Proton is designed to run in a very specific environment and it might be incompatible with your system. Steam runs Proton inside a bubble so that it interacts less with your system and so the incompatibilities don't become a problem. ULWGL aims to create the same bubble so it's the correct way to run proton.

  • If this isn't violating the DMA then the DMA is stupid. Legislation should limit the company's control, not force it into a specific action while allowing it to maintain as much control as possible.

    In other words the DMA should effectively say "you don't get to choose how your platform is used", not "you get to make the rules, but just don't be the only one who can develop for your platform".

  • Grub can load booster images, the issue is about incorrect grub.cfg generation.

    What they're saying in the issue is that grub-mkconfig will not create a correct "Arch Linux" menu entry for booster, but if you go to "Advanced options" and choose the "booster" menu entry it works. I can confirm this. It happened on the system I'm currently using.

    Specifically, the problem is that grub-mkconfig does not add the booster image to the initrd of the default menu entry. You can add it manually. For example I had to change this

     
        
    initrd  /intel-ucode.img
    
      

    to this

     
        
    initrd  /intel-ucode.img /booster-linux-zen.img
    
      

    If I recall correctly this issue was not present last time I set up a system with booster. It might be a regression or maybe it only happens in specific system configurations.

  • I use booster and it's cool. I don't see any noticeable difference in boot times but the image generation is much faster. mkinitcpio would take several seconds while booster takes about one.

    First time I tried it it didn't boot because of something missing in the generated image. I tried a universal booster image (set universal: True in /etc/booster.yaml) and it worked. Technically this builds a larger image than necessary but it's still only 34MB and takes a second to build, so I never bothered to troubleshoot what was missing. The universal image even handles luks encrypted root partitions without additional configuration of booster (you still have to configure kernel parameters).

    Another issue I noticed is that if you use grub-mkconfig and your only initramfs is booster, it will generate an incorrect main boot entry. It will add booster as an option in "advanced options" so your system is still bootable if this happens to you. The quick fix is to manually add the initrd entry under the main menuentry in grub.cfg.

  • Report that to PopOs. I doubt they would want this ad to appear to their users. They will probably remove it.

  • Aha I see what you're saying. It's possible that dr CD considered the second part to be crucial, but it doesn't seem that people who listened to his message felt the same way, myself included. I probably speak for a lot of people when I say we hadn't realized just how much these platforms are "subsidized" and how much damage that does to the entire market. So that part ended up being associated in our minds with the term enshittification.

  • "Enshitification" does not mean "I don't like it". It is specifically about platforms that start out looking too good to be true and turn to shit when the user base is locked in. The term is generally used for cases where the decline in quality was pre-planned and not due to external factors. Using the same term each time is, in my opinion, an appropriate way to point out just how common this pattern is.

  • From the naming it's clear that GE wants this to be the new standard, but it's not really a new standard. This is porting Steam's launcher, which already exists, to non-Steam clients.

  • This is great. Proton is getting a lot of testing just based on Steam's userbase and it is backed by Valve. We also have a lot of data on proton's performance and potential game-specific fixes in the form of protondb. Making sure that non-Steam launchers can use all that work and information is crucial to guaranteeing the long-term health of linux gaming. Otherwise it is easy to imagine a future where proton is doing great but the other launchers are keep running into problems and are eventually abandoned.

    One thing that I am curious is how this handles the AppId. If this AppId is used to figure out which game-specific fixes are needed, then it will have to be known. Do we have a tool/database that figures out the AppId from the game you are launching outside of Steam?

  • If you have a large enough bank roll and continuously double your bet after a loss, you can never lose without a table limit.

    Unless your bank roll is infinite, you always lose in the average case. My math was just an example to show the point with concrete numbers.

    In truth it is trivial to prove that there is no winning strategy in roulette. If a strategy is just a series of bets, then the expected value is the sum of the expected value of the bets. Every bet in roulette has a negative expected value. Therefore, every strategy has a negative expected value as well. I'm not saying anything ground-breaking, you can read a better write-up of this idea in the wikipedia article.

    If you don't think that's true, you are welcome to show your math which proves a positive expected value. Otherwise, saying I'm "completely wrong" means nothing.

  • So help me out here, what am I missing?

    You're forgetting that not all outcomes are equal. You're just comparing the probability of winning vs the probability of losing. But when you lose you lose much bigger. If you calculate the expected outcome you will find that it is negative by design. Intuitively, that means that if you do this strategy, the one time you will lose will cost you more than the money you made all the other times where you won.

    I'll give you a short example so that we can calculate the probabilities relatively easily. We make the following assumptions:

    • You have $13, which means you can only make 3 bets: $1, $3, $9
    • The roulette has a single 0. This is the best case scenario. So there are 37 numbers and only 18 of them are red This gives red a 18/37 to win. The zero is why the math always works out in the casino's favor
    • You will play until you win once or until you lose all your money.

    So how do we calculate the expected outcome? These outcomes are mutually exclusive, so if we can define the (expected gain * probability) of each one, we can sum them together. So let's see what the outcomes are:

    • You win on the first bet. Gain: $1. Probability: 18/37.
    • You win on the second bet. Gain: $2. Probability: 19/37 * 18/37 (lose once, then win once).
    • You win on the third bet. Gain: $4. Probability: (19/37) ^ 2 * 18/37 (lose twice, then win once).
    • You lose all three bets. Gain: -$13. Probability: (19/37) ^ 3 (lose three times).

    So the expected outcome for you is:

    $1 * (18/37) + 2 * (19/37 * 18/37) + ... = -$0.1328...

    So you lose a bit more than $0.13 on average. Notice how the probabilities of winning $1 or $2 are much higher than the probability of losing $13, but the amount you lose is much bigger.

    Others have mentioned betting limits as a reason you can't do this. That's wrong. There is no winning strategy. The casino always wins given enough bets. Betting limits just keep the short-term losses under control, making the business more predictable.

  • Im not 100% comfortable with AI gfs and the direction society could potentially be heading. I don’t like that some people have given up on human interaction and the struggle for companionship, and feel the need to resort to a poor artificial substitute for genuine connection.

    That's not even the scary part. What we really shouldn't be uncomfortable with is this very closed technology having so much power over people. There's going to be a handful of gargantuan immoral companies controlling a service that the most emotionally vulnerable people will become addicted to.

  • Well, not really, because television broadcast standards do not specify integer framerates. Eg North America uses ~59.94fps. It will take insanely high refresh rates to be able to play all common video formats including TV broadcasts. Variable refresh rate can fix this only for a single fullscreen app.

  • Exactly this. I can't believe how many comments I've read accusing the AI critics of holding back progress with regressive copyright ideas. No, the regressive ideas are already there, codified as law, holding the rest of us back. Holding AI companies accountable for their copyright violations will force them to either push to reform the copyright system completely, or to change their practices for the better (free software, free datasets, non-commercial uses, real non-profit orgs for the advancement of the technology). Either way we have a lot to gain by forcing them to improve the situation. Giving AI companies a free pass on the copyright system will waste what is probably the best opportunity we have ever had to improve the copyright system.

  • LLMs can do far more

    What does this mean? I don't care what you (claim) your model "could" do, or what LLMs in general could do. What we've got are services trained on images that make images, services trained on code that write code etc. If AI companies want me to judge the AI as if that is the product, then let them give us all equal and unrestricted access to it. Then maybe I would entertain the "transformative use" argument. But what we actually get are very narrow services, where the AI just happens to be a tool used in the backend and not part of the end product the user receives.

    Can it write stories in the style of GRRM?

    Talking about "style" is misleading because "style" cannot be copyrighted. It's probably impractical to even define "style" in a legal context. But an LLM doesn't copy styles, it copies patterns, whatever they happen to be. Some patterns are copyrightable, eg a character name and description. And it's not obvious what is ok to copy and what isn't. Is a character's action copyrightable? It depends, is the action opening a door or is it throwing a magical ring into a volcano? If you tell a human to do something in the style of GRRM, they would try to match the medieval fantasy setting and the mood, but they would know to make their own characters and story arcs. The LLM will parrot anything with no distinction.

    Any writer claiming to be so unique that they aren’t borrowing from other writers is full of shit.

    This is a false equivalence between how an LLM works and how a person works. The core ideas expressed here is that we should treat products and humans equivalently, and that how an LLM functions is basically how humans think. Both of these are objectively wrong.

    For one, humans are living beings with feelings. The entire point of our legal system is to protect our rights. When we restrict human behavior it is justified because it protects others; at least that's the formal reasoning. We (mostly) judge people based on what they've done and not what we know they could do. This is not how we treat products and that makes sense. We regulate weapons because they could kill someone, but we only punish a person after they have committed a crime. Similarly a technology designed to copy can be regulated, whereas a person copying someone else's works could be (and often is) punished for it after it is proven that they did it. Even if you think that products and humans should be treated equally, it is a fact that our justice system doesn't work that way.

    People also have many more functions and goals than an LLM. At this point it is important to remember that an LLM does literally one thing: for every word it writes it chooses the one that would "most likely" appear next based on its training data. I put "most likely" in quotes because it sounds like a form of prediction, but actually it is based on the occurrences of words in the training data only. It has nothing else to incorporate to its output, and it has no other need. It doesn't have ideas or a need to express them. An LLM can't build upon or meaningfully transform the works it copies, it's only trick is mixing together enough data to make it hard for you to determine the sources. That can make it sometimes look original but the math is clear, it is always trying to maximize the similarity to the training data, if you consider choosing the "most likely" word at every step to be a metric of similarity. Humans are generally not trying to maximize their works' similarity to other peoples' works. So when a creator is inspired by another creator's work, we don't automatically treat that as an infringement.

    But even though comparing human behavior to LLM behavior is wrong, I'll give you an example to consider. Imagine that you write a story "in the style of GRRM". GRRM reads this and thinks that some of the similarities are a violation of his copyright so he sues you. So far it hasn't been determined that you've done something wrong. But you go to court and say the following:

    • You pirated the entirety of GRRM's works.
    • You studied them only to gain the ability to replicate patterns in your own work. You have no other user for them, not even personal satisfaction gained from reading them.
    • You clarify that replicating the patterns is achieved by literally choosing your every word to be the one that you determined GRRM would most likely use next.
    • And just to be clear you don't who GRRM is or what he talks like. Your understanding of what word he would most likely use is based solely on the pirated works.
    • You had no original input of your own.

    How do you think the courts would view any similarities between your works? You basically confessed that anything that looks like a copy is definitely a copy. Are these characters with similar names and descriptions to GRRM's characters just a coincidence? Of course not, you just explained that you chose those names specifically because they appear in GRRM's works.