• 0 Posts
  • 53 Comments
Joined 1 year ago
cake
Cake day: November 23rd, 2024

help-circle
  • As far as I know that’s generally what is often done, but it’s a surprisingly hard problem to solve ‘completely’ for two reasons:

    1. The more obvious one - how do you define quality? When you’re working with the amount of data LLMs require as input and need to be checked for on output you’re going to have to automate these quality checks, and in one way or another it comes back around to some system having to define and judge against this score.

      There’s many different benchmarks out there nowadays, but it’s still virtually impossible to just have ‘a’ quality score for such a complex task.

    2. Perhaps the less obvious one - you generally don’t want to ‘overfit’ your model to whatever quality scoring system you set up. If you get too close to it, your model typically won’t be generally useful anymore, rather just always outputting things which exactly satisfy the scoring principle, nothing else.

      If it reaches a theoretical perfect score, it would just end up being a replication of the quality score itself.


  • Luanti and Minecraft are two distinct, if similar-looking things.

    Luanti is an open-source voxel game engine implementation which allows running a wide variety of different ‘games’ on it (including two which mimic Minecraft very closely, like the above-mentioned Mineclonia).

    Minecraft is the closed-source game owned by Mojang.

    The two don’t interact and servers for the one are completely unrelated to the other as well.

    So, to answer the question - yes, they still need a Minecraft license if they want to play Minecraft. But this is disconnected from having a Luanti server, for which you don’t need any licenses but which will in turn also only allow you to play Luanti stuff, not Minecraft.





  • I think you really nailed the crux of the matter.

    With the ‘autocomplete-like’ nature of current LLMs the issue is precisely that you can never be sure of any answer’s validity. Some approaches try by giving ‘sources’ next to it, but that doesn’t mean those sources’ findings actually match the text output and it’s not a given that the sources themselves are reputable - thus you’re back to perusing those to make sure anyway.

    If there was a meter of certainty next to the answers this would be much more meaningful for serious use-cases, but of course by design such a thing seems impossible to implement with the current approaches.

    I will say that in my personal (hobby) projects I have found a few good use cases of letting the models spit out some guesses, e.g. for the causes of a programming bug or proposing directions to research in, but I am just not sold that the heaviness of all the costs (cognitive, social, and of course environmental) is worth it for that alone.



  • I’ve been exclusively reading my fiction books (all epubs) on Readest and absolutely love it. Recently I also started using it for my nonfiction books and articles (mostly pdf) as an experiment, and it’s workable but a little more rough around the edges still.

    You can highlight and annotate, and export all annotations for a book once you are done, for which I have set up a small pipeline to directly import them into my reference management software.

    It works pretty well with local storage (though I don’t believe it does ‘auto-imports’ of new files by default) and I’ve additionally been using their free hosted offering to sync my book progress. It’s neat and free up to 500mb of books, but you’re right that I would also prefer a byo storage solution, perhaps in the future.

    The paid upgrades are mostly for AI stuff and translations which I don’t really concern myself with.



  • Open source/selfhost projects 100% keep track of how many people star a repo, what MRs are submitted, and even usage/install data.

    I feel it is important to make a distinction here, though:

    GitHub, the for-profit, non-FOSS, Microsoft-owned platform keeps track of the ‘stars of a repo’, not the open-source self-host projects themselves. Somebody hosts their repo forge on Codeberg, sr.ht, their own infrastructure or even GitLab? There’s generally very little to no algorithmic number-crunching involved. Same for MR/PRs.

    Additionally - from my knowledge - very few fully FOSS programs have extensive usage/install telemetry, and even fewer opt-out versions. Tracking which couldn’t be disabled I’ve essentially never heard of in that space, because every time someone does go in that direction the public reaction is usually very strong (see e.g. Audacity).


  • Interesting, so Metal3 is basically kubernetes-managed baremetal nodes?

    Over the last years I’ve cobbled together a nice Ansible-driven IaC setup, which provisions Incus and Docker on various machines. It’s always the ‘first mile’ that gets me struggling with completely reproducible bare-metal machines. How do I first provision them without too much manual interference?

    Ansible gets me there partly, but I would still like to have e.g. the root file system running on btrfs which I’ve found hard to accomplish with just these tools when first provisioning a new machine.


  • Sioyek really is amazing, especially for academic-style reading with a lot of jumping back and forth, and very customizable. I also heartily recommend it, but do be aware that there are some rough edges remaining.

    If you ever get stuck, there are a lot of additional tricks and workarounds for some of the quirks hidden in the project’s github issues. And if there’s a feature you feel sorely missing check out the main branch version instead of the latest official point release which is a couple years behind now (e.g. still missing integrated dual-page view which the development version has for close to 2 years now)


  • When I was stumbling on some of his output it unfortunately felt very click-baity, always playing on your FOMO if you didn’t set up/download/buy the next best thing until the other next best thing in the video after.

    In other words, I think he’s cool to check out to get to know of a thing, but to get a deeper level of understanding how a thing works I would recommend written materials. There are good caddy/nginx tutorials out there, but a linux networking book will get your understanding further yet.

    If it has to be video, I would at least recommend a little more slowed down, long-form content like Learn Linux TV.


  • I’ve been using NetBird for quite a while now. It has grown from an experiment in connecting to the server without exposing it to quite a stable setup that I make use of every day, and even got my partner and some of my family to use. That is the hosted offering, however, not me self hosting my own server component.

    For a couple of months now, I’ve been eyeing pangolin though. It just seems like such an upgrade concerning identity and SSO - but equally a complete overhaul of my infrastructure and a steep learning curve.

    I am itching to get it running but would probably have to approach it step-by-step, and roll it out pretty slowly, while transferring the existing services.




  • That’s a little annoying with all the others not working. Haven’t seriously tried most of them so I’m afraid I can’t really help you there - though if you ever try Q4OS that others have suggested let me know if it works well cause I may give that a whirl too on the little eee.

    If you decide to stick with antix, I could maybe see if I find some of my old notes. I vaguely remember the wifi giving me some trouble and the homebrewed settings panels of the distro can be… a little funky :-)

    Good luck!


  • I was running AntiX out of your list on my old atom eee-pc pretty successfully the last 2-3 years. Was using it as a workbench pc with an old vga screen and keyboard connected, and it worked well enough for simple pdf /datasheet reading and terminal sessions.

    For specs, I think it was the same cpu but only 1gb of ram. Honestly with 2gb of ram your options are much broader, the one part you’ll run into trouble with is the browser with multiple tabs anyway. I thought to remember there was also a community-maintained 32bit Archlinux variant?

    Edit: https://www.archlinux32.org/ that’s the one I believe. It has a more restricted package repo but otherwise is just Arch.



  • I liked the game well enough when it came out, had a good friend at the time whom I always traded little game design insights and fun facts about the AssCreed games with.

    But the one thing nowadays I always remember about this one is that the ‘opening’ part is looooo(…)oong - until you really swing you sword and hidden blade about it takes hours of grand opening, shipping to America, learning the controls, doing little ‘preview’ missions in a restricted zone, then

    Spoiler

    finally switching to the actual main character only to have to do a new tutorial intro all over for a couple of hours.

    It felt somewhat compelling the first time round but on subsequent playthroughs it really stretched your patience - imo, of course.