Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)RE
Posts
0
Comments
143
Joined
2 yr. ago

  • A claim that we have a computing model that shares a design with the operation of a biological brain is philosophical and conjecture

    Mathematical actually. See the 1943 McCulloch and Pitts paper for why Neural networks are called such.

    We use logic and math to approximate neurons

  • Facts.

    Even if you could design your own model... How do you acquire a dataset even a fraction of the size those pretrained models from the corps.

    Then how do you train the model in a reasonable time. Other than relying on cloud computing which leads to the same problem of only corps can play this game properly right now.

    I designed and collected/labeled the data for a relatively small deep CNN for my masters thesis and training it on 60000 images was taking over a dozen hours (this was 5 years ago at this point so that part may be misremembered) on a 1080ti.

  • Hey did you know the specific heat of water is 4.184 j/g°c because i learned that once in a physics class more than a decade ago and i still remember.

    I am not a physicist and the memorization of that fact does not help me. Thanks brain.

  • I have... feelings about LLMs being the big thing in AI/ml right now.... because its really not much new. Maybe the transformer model kind of but ultimately LLMs are massive supervised learning neural nets trained on obscene amounts of data. And then other models use that pretrained "foundational model" to work and just tune their parameters. Which is why prompt engineer is becoming a thing.

    Corpos are playing by the book here and trying to extinguish any competition before it begins by having people rely on their "foundation" models instead of innovating their own solutions

    How many tutorials can you find for implementing LLM NLP tasks that dont include "import this model from X company" id wager its only maybe 33%

  • Ive never understood how the same crowd that spouts not your keys not your crypto would ever trust any password manager they havent personally read the source code for/compiled/self hosted.

    Not your server not your safe/secure password