Maybe see if there are some esports you like if more traditional ones are unavailable at this time. There are also things like chess tournaments if you aren't really into video games.
As a solo developer, some things are out of scope like writing translations or ensuring full compliance with accessibility standards. What's important is to have some knowledge of what things block progress in these areas. For example, not treating all strings like ASCII, or preferring native widgets/html elements as those better support accessiblity tools.
Accessibility and internationalization first. A lot of projects start without it and tack it on later. It's so much better to have good roots and promote diversity and inclusivity from the start.
If you have a high end GPU, or lots of RAM you can run some good quality LLMs offline. I recommend watching Matthew Berman for tutorials (there are some showing paid hosting aswell).
Knowledge level: Enthusiastic spectator, I don't make or finetune llms, but I do watch AI news, try out local llms, and use things like Github copilot and chat gpt.
Question: Is it better to use code llama 34b or llama2 13b for a non coding related task?
Context: I'm able to run either model locally, but I can't run the larger 70b model. So I was wondering if running the 34b code llama would be better since it is larger. I heard that models with better coding abilities are better for other types of tasks too and that they are better with logic (I don't know if this is true I just head l heard it somewhere).
Interesting that it only works on windows, a lot of AI projects I've seen have been the other way around.