• 0 Posts
  • 10 Comments
Joined 1 year ago
cake
Cake day: October 20th, 2023

help-circle
  • Part of it is that Ubuntu/Canonical so aggressively pushed Snaps which became a huge culture war. So you have people who hate the idea of those style of packages because they hate Snap AND people who hate flatpak because they are Team Ubuntu for some reason.

    And the other aspect is that it is incredibly space inefficient (by the very nature of bundling in dependencies) and is prone to “weirdness” when it comes to file system permissions and the like. And many software projects kind of went all in on them because it provides a single(-ish) target to build for rather than having a debian and an arch and a redhad and a…


  • That’s nice.

    That doesn’t work if you are on an airplane (unless you want to spend the entire flight downloading one episode). Or if you just don’t want to deal with hotel wifi. Or if you just don’t want to expose your internal home network at all.

    Which is the point and why this is one of those big features of plex that there are so many tickets and requests to get into jellyfin et al. Because yes, you can just copy files from your NAS to your phone’s internal storage (assuming you don’t care about transcoding and the like)… at which point there isn’t much use to a metadata oriented media server/service.

    Or you can just set up Plex to always download the next 10 episodes of whatever show you are watching when it has network access. I mean… that probably won’t work (see: 40%) but when it does, it is awesome. Which is the “it just works” functionality.

    Which gets back to the issue where, because it is FOSS, it is the greatest thing ever and anyone asking for anything else is wrong and stupid. Which is a shame because if the Jellyfin devs could actually get the “download the next N episodes” functionality to reliably work (even at 80-90%) it would be a killer app. And, for what it is worth, I have liked the devs a lot when I interacted with them in the past. But the users and evangelists are just… what we can see in this thread.



  • Yeah.

    Jellyfin is spectacular for LAN usage on two computers. Once you start using devices (because, you know, that is what people tend to plug into their TVs…) or going on travel, it rapidly becomes apparent that it just isn’t a competitor.

    Hell, a quick google suggests jellyfin STILL doesn’t have caching of media for offline viewing. Plex’s works maybe 40% of the time but… 40% is still higher than 0%.

    I have a lifetime pass for Plex and encourage anyone who even kind of cares to get one next time it is on sale (or shortly before the scheduled price hike). I have tried Jellyfin a few times over the years and… it is basically exactly what I hate with FOSS “alternatives”. It isn’t an alternative in the slightest but people insist on talking it up because they want it to be and that just makes people less willing to try genuinely good alternatives.


    To put it bluntly, Plex is an “offline netflix” as it were. Jellyfin is a much better version of smbstation and all the other stuff we used to stream porn to our playstations back in the day.



  • I tried bookywrm a year or two back. It was nice that I could import my goodreads history. But something like 70% of my books just weren’t in the database.

    And I am not just talking kindle exclusive authors. It was (not necessarily one of the books but at the same level) seminal works like Lois Lowry’s The Giver.

    These days I am checking out Storygraph. Import was very smooth. Not super huge on the excessive amounts of AI but AI generated premises that (theoretically) cater to my own interests seems like a good use of the devil’s silicon.


  • Even those examples are the kinds of things that “fall apart” if you actually think things through.

    Art? Actual human artists tend to use a ridiculous amount of “AI” these days and have been for well over a decade (probably closer to two, depending on how you define “AI”). Stuff like magic erasers/brushes are inherently looking at the picture around it (training data) and then extrapolating/magicking what it would look like if you didn’t have that logo on your shirt and so forth. Same with a lot of weathering techniques/algorithms and so forth.

    Same with coding. People more or less understand that anyone who is working on something more complex than a coding exercise is going to be googling a lot (even if it is just that you will never ever remember how to do file i/o in python off the top of your head). So a tool that does exactly that is… bad?

    Which gets back to the reality of things. Much like with writing a business email or organizing a calendar: If a computer program can do your entire job for you… maybe shut the fuck up about that program? Chatgpt et al aren’t meant to replace the senior or principle software engineer who is in lots of design meetings or optimizing the critical path of your corporate secret sauce.

    It is replacing junior engineers and interns (which is gonna REALLY hurt in ten years but…). Chatgpt hallucinated a nonsense function? That is what CI testing and code review is for. Same as if that intern forgot to commit a file or that rockstar from facebook never ran the test suite.

    Of course, the problem there is that the internet is chock full of “rock star coders” who just insist the world would be a better place if they never had to talk to anyone and were always given perfectly formed tickets so they could just put their headphones on and work and ignore Sophie’s birthday and never be bothered by someone asking them for help (because, trust me, you ALWAYS want to talk to That Guy about… anything). And they don’t realize that they were never actually hot shit and were mostly always doing entry level work.

    Personally? I only trust AI to directly write my code for me if it is in an airgapped environment because I will never trust black box code I pulled off the internet to touch corporate data. But I will 100% use it in place of google to get an example of how to do something that I can use for a utility function or adapt to solving my real problem. And, regardless, I will review and test that just as thoroughly as the code Fred in accounting’s son wrote because I am the one staying late if we break production.


    And just to add on, here is what I told a friend’s kid who is an undergrad comp sci:

    LLMs are awesome tools. But if the only thing you bring to the table is that you can translate the tickets I assigned to you to a query to chatgpt? Why am I paying you? Why am I not expensing a prompt engineering course on udemy and doing it myself?

    Right now? Finding a job is hard but there are a lot of people like me who understand we still need to hire entry level coders to make sure we have staff ready to replace attrition over the next decade (or even five years). But I can only hire so many people and we aren’t a charity: If you can’t do your job we will drop you the moment we get told to trim our budget.

    So use LLMs because they are an incredibly useful tool. But also get involved in design and planning as quickly as possible. You don’t want to be the person writing the prompts. You want to be the person figuring out what prompts we need to write.


  • Again: What is the percent “accurate” of an SEO infested blog about why ivermectin will cure all your problems? What is the percent “accurate” of some kid on gamefaqs insisting that you totally can see Lara’s tatas if you do this 90 button command? Or even the people who insist that Jimi was talking about wanting to kiss some dude in Purple Haze.

    Everyone is hellbent on insisting that AI hallucinates and… it does. You know who else hallucinates? Dumbfucks. And the internet is chock full of them. And guess what LLMs are training on? Its the same reason I always laugh when people talk about how AI can’t do feet or hands and ignore the existence of Rob Liefeld or WHY so many cartoon characters only have four fingers.

    Like I said: I don’t like the AI Assistants that won’t tell me where they got information from and it is why I pay for Kagi (they are also AI infested but they put that at higher tiers so I get a better search experience at the tier I pay for). But I 100% use stuff like chatgpt to sift through the ninety bazillion blogs to find me a snippet of a helm chart that I can then deep dive on whether a given function even exists.

    But the reality is that people are still benchmarking LLMs against a reality that has never existed. The question shouldn’t be “we need this to be 100% accurate and never hallucinate” and instead be “What web pages or resources were used to create this answer” and then doing what we should always be doing: Checking the sources to see if they at least seem trustworthy.


  • People love to make these claims.

    Nothing is “100% accurate” to begin with. Humans spew constant FUD and outright malicious misinformation. Just do some googling for anything medical, for example.

    So either we acknowledge that everything is already “sewage” and this changes nothing or we acknowledge that people already can find value from searching for answers to questions and they just need to apply critical thought toward whether I_Fucked_your_mom_416 on gamefaqs is a valid source or not.

    Which gets to my big issue with most of the “AI Assistant” features. They don’t source their information. I am all for not needing to remember the magic incantations to restrict my searches to a single site or use boolean operators when I can instead “ask jeeves” as it were. But I still want the citation of where information was pulled from so I can at least skim it.