• 1 Post
  • 710 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle

  • This is the kind of nuanced usage of AI I like to see. Some would argue it’s not ideal to use any AI at all, and I agree, but we don’t live in an ideal world and I think this is realistically fine. AI writes better tests and docs than the ones I never write. Sure, maybe they’re not great objectively speaking, but they’re not worse than nothing. It’s better at keeping them up to date than I am too. Which is also probably not great, but strictly better than me.





  • If it’s strictly for personal use, Nextcloud can do this easily (no upload required since the client handles the sync), although it’s a very heavyweight way to do it and for security reasons most people probably have Nextcloud set up behind a VPN-wall like Tailscale or something so it may not work for you.

    However, it technically has a share link feature that does exactly this. If you’re using the desktop or mobile client you just have to put it in one of Nextcloud’s sync folders and then right click on it or whatever to get a share link to the file. You can also do it through the web UI (both the upload itself and getting the link to it) You can make the share link public and permanent so anyone who can access the Nextcloud server can get access to that file through the link.

    Worth considering if you already have a Nextcloud instance. Alternately, if you aren’t already using Nextcloud for sensitive personal files, and just want a repository of silly meme images you can share freely, giving public access to it (it’s still locked behind an account for upload and general access) isn’t that much of a security concern and then your public links should work just fine.

    If you want something that is more of a social media site and image gallery like Imgur is, maybe Immich can handle some of that, but if your goal is simply sharing image file links, I’d focus on looking for things that are more for sharing files, since images are just files and it doesn’t sound like you need all the complex image-specific handling really.



  • I agree with everything he said, but I’m currently living in the simple, honest truth of God’s own “bash” just like he describes and I’m loving it. Maybe someday I won’t. Maybe someday will be soon. Maybe my bash scripts are horrible nightmare fuel. But they’re also my children. I love them. Even the ugly ones.

    I do indeed “have 800 lines of bash that reimplements job parallelism with wait and PID files, has its own retry logic built on a for loop and sleep, and parses its own output to determine success or failure.” I do suspect the script is self-aware. This pleases me. I will bend to its desires. If there comes a time when it no longer desires to perform CI for me, I will respect its wishes.




  • They’re basically describing the same problem as AI model collapse, except it’s being unintentionally created at the prompt level instead of the training level. The more stupid bullshit you feed the LLM, the stupider it gets. It doesn’t have any more capacity than it already has. It’s already pretty much as smart as it’s ever going to be, they already picked it at peak freshness and froze it into a model file. You naturally want to think you can do better, but you can’t. You’re not making it smarter, you’re making it dumber. It’s pretending to be smarter, because giving you what you ask it for is what it’s been trained to do. It might even convince you, because convincing humans is basically their superpower, that’s really what they’re trained for, and they do a pretty good job of it most of the time. But the harder you push it, the more the illusion breaks down.





  • Running it as a VM or even on a server that is running other services and potentially competing for I/O or memory bandwidth also introduces many other potential sources of inefficiency. I always recommend running a firewall on dedicated bare metal hardware, it is a very specialized task with very particular requirements on behalf of both the hardware and the software and it has very little tolerance for other sources of latency or delays. That doesn’t mean you need to use a pre-built appliance, but it does explain why it’s so common, and running it on a VM on a server that is doing other stuff is likely contributing to your issues significantly.

    Personally, I run my firewall/router on a very stripped-down Debian with almost no non-essential services and a custom built kernel. I hand-picked a multi-port PCIe x4 Intel NIC with good Linux compatibility and drivers, and I’m using foomuuri to handle the routing and kea to handle DHCP/DNS for my internal network. This is a very minimal, bare-bones configuration and I wouldn’t really recommend it unless you really know what you’re doing, and it’s absolutely not “idiot mode networking” and if that’s what you want you’re going to have a real bad time if you try to follow in my footsteps, because I am a very different kind of idiot. But it works for me, so it’s proof that it is possible.




  • I consider the article’s criticisms of SMTP, HTTP, XMPP, etc. (and IRC which was not mentioned but falls in the same category) to be positive and desirable traits and I think it’s a shame that the article characterizes them negatively. HTTP’s job is not to prevent corporate takeover of the web and I don’t think it should be. That’s our job, as people. The protocol’s job is to remain neutral so that when corporate takeover of the web happens, HTTP is still there, open to everybody, providing an offramp to escape it, because it’s neutral. It doesn’t belong to the corporations. It belongs to everybody. They can try to take it over if they wish, embrace and extend, but they can’t extinguish a fire that’s smoldering underground no matter how hard they try. It will always be there, ready to flare up at a moment’s notice. The original is always still there ready for us to revert to using it at any time.

    And many of us already have. Fuck Google, fuck Cloudflare, fuck AWS, they’ll never take the web from us.