Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)JL
Posts
0
Comments
23
Joined
4 mo. ago

  • At best, that's reductive and at worst it's completely wrong. I get that this was probably a joke answer, but I feel like this misconception is unfortunate since it misrepresents ancient Egyptian culture and also undermines the impact of the unique evil of chattel slavery that was practiced in the US.

  • If you're building a website, you'll probably want to stick to Javascript over Rust.

    This MDN article does a pretty good job at introducing the concept of making network requests in Javascript: https://developer.mozilla.org/en-US/docs/Learn_web_development/Core/Scripting/Network_requests. It focuses on the "fetch" API as the tool for making requests, which is the standard way to make network requests in Javascript. There are other tools like Axios that may make things easier, but "fetch" should be fine for your use case.

    Another concept that will be relevant here is asynchronous programming: https://developer.mozilla.org/en-US/docs/Learn_web_development/Extensions/Async_JS/Introducing. Basically, there will be some delay between when you make the request and when you get a response. So you'll need to write your logic in a way that does the "waiting" part correctly.

    One important detail is that most APIs use some form of authentication. So when you're "grabbing the data" from an external site, the site knows who you are and that you are allowed to access that data. Getting authentication right might be a little tricky, but here is an entry point: https://developer.mozilla.org/en-US/docs/Web/HTTP/Guides/Authentication. Basically, you'll need to figure out what authentication strategy your headless CMS is using, and then make sure to safely pass those credentials when making your network requests. If the API(s) you are using are public, you won't need to worry about this.

    If your goal is primarily to get data from an external source, this should be a good starting point. You don't necessarily need to get too deep into the backend or even the technical details of things like HTTP or REST. However, if you're interesting in getting a deeper understanding of Web APIs, the other comment talking about building a skeleton API would be a good exercise.

  • I 100% agree with the idea that rap is certainly not the only genre to glorify violence, drugs, etc. But the specific song choice is not really a great example in my opinion considering the last verse is

    C'mon you gotta listen unto me

    lay off that whiskey, and let that cocaine be.

    This song always struck me as a cautionary tale. Nothing about the song really seems to glorify the behavior.

  • For this specific HN post, speedrunning is a bit of a misnomer. He used similar tooling to effectively add support for a physical keyboard and additionally a bunch of keyboard shortcuts that are capable of inputting custom text, songs, and fabric patterns.

    There's a YouTube video where the author showcases this. It's pretty short and a really interesting watch: https://youtu.be/Yw8Alf_lolA

  • Permanently Deleted

    Jump
  • I do have enough time, but I don't have the self control. If I could hold myself to an hour a day, that would be fantastic, but I inevitably get myself too addicted and end up spending closer to 4 hours a day. At that point, all my other chores aren't getting done. As a result, I haven't played video games in several years.

  • Why can't the world have more of this and less of... everything else that's going on right now? 😕

    I don't really have a point, it's just sad that humans have the capacity to do such cool, fun, creative things, and instead we're burning the world down so computers can churn out garbage and blowing each other up because we're different from each other.

  • But with the rise of AI, the dynamic is changing: We are observing a significant increase in request volume, with most of this traffic being driven by scraping bots collecting training data for large language models (LLMs) and other use cases. Automated requests for our content have grown exponentially, alongside the broader technology economy, via mechanisms including scraping, APIs, and bulk downloads. This expansion happened largely without sufficient attribution, which is key to drive new users to participate in the movement, and is causing a significant load on the underlying infrastructure that keeps our sites available for everyone.

    https://diff.wikimedia.org/2025/04/01/how-crawlers-impact-the-operations-of-the-wikimedia-projects/

  • This is part of the larger problem that AI tools are trained on (and profit off of) content that is produced and hosted by others who are now seeing their traffic change from humans to bots. For content sources that pay for hosting with ads, this means a loss in revenue to pay for hosting. For content sources like Wikipedia, they are seeing their hosting costs increase significantly due to the increase in bot traffic. Even if you want every website that depends on ad revenue to fail (which I don't entirety agree with), AI is still damaging the open web in other ways. Websites like Wikipedia for example may soon be forced to lock content behind logins or leverage aggressive captchas just to fight the bot traffic, which makes things worse for those of us that still prefer to use actual websites over AI summaries.