• 0 Posts
  • 20 Comments
Joined 2 years ago
cake
Cake day: June 13th, 2023

help-circle
  • So maybe we’re kinda staring at two sides of the same coin. Because yeah, you’re not misrepresentin my point.

    But wait there’s a deeper point I’ve been trying to make.

    You’re right that I am also saying it’s all bullshit - even when it’s “right”. And the fact we’d consider artificially generated, completely made up text libellous indicates to me that we (as a larger society) have failed to understand how these tools work. If anyone takes what they say to be factual they are mistaken.

    If our feelings are hurt because a “make shit up machine” makes shit up… well we’re holding the phone wrong.

    My point is that we’ve been led to believe they are something more concrete, more exact, more stable, much more factual than they are — and that is worth challenging and holding these companies to account for. i hope cases like these are a forcing function for that.

    That’s it. Hopefully my PoV is clearer (not saying it’s right).


  • Ok hear me out: the output is all made up. In that context everything is acceptable as it’s just a reflection of the whole of the inputs.

    Again, I think this stems from a misunderstanding of these systems. They’re not like a search engine (though, again, the companies would like you to believe that).

    We can find the output offensive, off putting, gross , etc. but there is no real right and wrong with LLMs the way they are now. There is only statistical probability that a) we’ll understand the output and b) it approximates some currently held truth.

    Put another way; LLMs convincingly imitate language - and therefore also convincing imitate facts. But it’s all facsimile.



  • Surely you jest because it’s so clearly not if you understand how LLMs work (at the core it’s a statistic model - and therefore all approximation to a varying degree).

    But great can come out of this case if it gets far enough.

    Imagine the ilk of OpenAI, Google, Anthropic, XAI, etc. being forced to admit that an LLM can’t actually do anything but generate approximations of language. That these models (again LLMs in particular) produce approximations of language that are so good they’re often indistinguishable from the versions our brains approximate.

    But at the core they cannot produce facts because the way they are made includes artificially injected randomness layered on-top of mathematically encoded values that merely get expressed as tiny pieces of language (tokens) - ones that happen to be close to each other in a massively multidimensional vector space.

    TLDR - they’d be forced to admit the emperor has no clothes and that’s a win for everyone (except maybe this one guy).

    Also it’s worth noting I use LLMs for work almost daily and have studied them quite a bit. I’m not a hater on the tech. Only the capitalists trying to force it down everyone’s throat in such a way that we blindly adopt it for everything.









  • I see Jellyfin suggested as an alternative to Plex here. I hope it is one day.

    At the moment it’s nowhere close.

    I’ve been running Jellyfin side-by-side Plex for two years and it’s still not a viable replacement for anyone but me. Parents, my partner, none of the possible solutions for them come anywhere near close to the usability of Plex and its ecosystem of apps for various devices.

    That will likely change because plex is getting worse every day and folks can contribute their own solutions to the playback issues. With plex it’s more noise, more useless features. So one gets better (Jellyfin) and one gets worse (Plex).

    But at the moment it really isn’t close for most folks who are familiar with the slickness of commercial apps.

    Even from the administrative side, Jellyfin takes massively more system resources and it doesn’t reliably work with all my files.

    Again, Jellyfin will get there it’s just not a drop in replacement for most folks yet.

    And for context I started my DIY streaming / hosting life with a first gen Apple TV (pretty much a Mac mini with component video outs) that eventually got XBMC and then Boxee installed on it. I even have the forksaken Boxee box.



  • We use NGINX’s 444 on every LLM crawler we see.

    Caddy has a similar “close connection” option called “abort” as part of the static response.

    HAProxy has the “silent-drop” option which also closes the TCP connection silently.

    I’ve found crawling attempts end more quickly using this option - especially attacks - but my sample size is relatively small.

    Edit: we do this because too often we’ve seen them ignore robots.txt. They believe all data is theirs. I do not.





  • I’m with you on all those points. Loblaws is theft of a different kind and to your most important point: these stores have made it impossible for smaller stores to exist - so I also get that you might not have a choice.

    And thanks for responding so kindly. Appreciate you.

    My perspective here, to share, is that I don’t understand the hard time others are having. Not that I can’t understand - but rather that I am not in their shoes. I’ve been trying to cut all workers slack because they’re also trapped in a system and we’re powerful if we stick together - so forgive me if I also forgot to empathize with you.


  • I have to ask - how did you think a cashier can help with your request?

    I’m not trying to be a jerk here, but cashiers are pretty low on the corporate ladder at Loblaws. If you knew it wasn’t their job, what were you looking for and what did you want in response?

    I don’t shop at Loblaws. We’ve been shopping locally since the boycott last May.

    But one of the things worth considering is what we expect of each other. Was one person’s rudeness reflective of everyone else who works somewhere?

    There are so many way we can work together to make our communities and country stronger. Not shopping at Loblaws can be one of them. But for this reason? I’m sorry, it just seems strained and unusual.

    Edited to remove racist euphemism.


  • Recently switched from VsCodium to neovim - but still use Codium for some specific tasks.

    My setup customization focuses around Telescope, Treesitter, Trouble & Blink.

    But the advice I got was to start with vim keybindings in VSCode. I used those for six weeks until I got the hang of the basics and it had gone from frustrating to somewhat second nature.

    Then I made the move.

    I still use Codium for Terraform work (I have struggled to get the Terraform LS working well in neovim and I don’t use it often enough to warrant the effort) and as a GUI git client - I like the ability to add a single line from multiple files and I haven’t looked up how to do it any other way - I’ve got other stuff to do and it’s not slowing me down.

    But I grew to hate Codium / VS code tabs in larger codebases. I was spending so much time looking for open tabs ( I realise this is a me problem). While neovim has tabs, it’s much more controlled and I typically use them very differently and very sparingly.

    If I need to look up a data structure I just call it up temporarily with Telescope via a find files call or a live grep call (both setup to only use my project directory by default), take a peak, and move on.

    The thing is - security risks are going to exist anywhere you install plugins you haven’t audited the code for. Unless you work in an IDE where there’s a company guaranteeing all plugins - there are always going to be risks.

    I’d argue that VSCode, while a bigger target, has both a large user base and Microsoft’s security team going for it. I don’t see the theme being compromised as much as problem because it got solved and also prompted some serious security review of many marketplace plugins. Not ideal, but not terrible.