Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)VR
Posts
0
Comments
1,052
Joined
2 yr. ago

  • raytracing is insanely expensive. If you saw what current cards can render in real time, you would see a very very noisy, incomplete image that looks like shit. Without ai denoising and a lot of temporal shit (which only looks good in screenshots). It is very very very far from being able to render an actual frame with decent performance.

  • no, it isn't. A lot of systems still run power through the battery even after it's charged. The battery ends up in a state of constant trickle charging.

    In the deck, once the battery is charged, and the power cable is connected, the battery is bypassed and effectively disconnected.

    And the deck only charges to 100% if it was below 90 to begin with.

  • because of the "perceptual" part.

    A normal hash has the property that it produces wildly different hashes for even the tiniest of changes in the file.

    Perceptual hashing flips that requirement on its head, and therefore makes finding a suitable hash function much harder.

  • This is not error correction issue though. Error correction means taking known data and adding redundancy to it so that damaed pieces can be repaired. This makes the message longer.

    An llm's output does not contain error correction. It's just the output. And it doesn't contain any errors, mathematically speaking. The hallucination is the correct output. It is what the statistics it gathered from its training set determined is most likely. A "correct" llm output is indistinguishable from a "hallucination", mathematically, and always will be. A hallucination is simply "some output that some human, somewhere, doesn't like", and that's uncomputable. And outputs that people subjectively consider as "hallucinations" cannot be eliminated, because an llm is, fundamentally, a probabilistic algorithm. If you added error correction to an llm's output all you'd be able to recover is the llm's original output, "hallucinations" and all.

    Tldr: "hallucinations" are a subjective thing. A Hallucination" is not an error that can be corrected after-the-fact, because it is not an error in the first place.