• 0 Posts
  • 47 Comments
Joined 24 days ago
cake
Cake day: June 9th, 2024

help-circle
  • I don’t agree with the whole list, but the CLA requirement and corpo projects pinky-promising they’d never do a bad thing and then going to do a bad thing as soon as their investors demand returns is certainly a major risk and harm. I’ve started self-hosting everything for my personal use, and if it’s not AGPL, then I assume at some point I’m going to get fucked and shouldn’t rely on it.

    Also, the endless stupidity around everyone using Discord as their primary means of communication, discussion, issue reporting and whatnot. Politely, fuck Discord, and fuck anyone who thinks Discord is the right way to make anything accessible to the public.

    There’s lots of other alternatives, including ye olde IRC and forums and even simple mailing lists - and no, I don’t mean ‘sign up for our newsletter!’ nonsense, but an actual real mailing list. And, if you want something a little more modern, there’s always Matrix which is probably feature-complete enough to compete with whatever you’d want to use Discord for anyways.



  • Yeah, exactly: if you know how it works, then you know how to fix it. I don’t think you need a comprehensive knowledge about how everything you run works, but you should at least have good enough notes somewhere to explain HOW you deployed it the first time, if you had to make any changes as well as anything you ran into that required you to go figure out what the blocking issue was.

    And then you should make sure that documentation is visible in a form that doesn’t require ANYTHING to actually be working, which is why I just put pages of notes in the compose file: docker doesn’t care, and darn near any computer on earth made in the last 40 years can read a plan text file.

    I don’t really think there’s any better/worse reverse proxy for simple configurations, but I’m most familiar with nginx, which means I’ve spent too long fixing busted shit on it so it’s the choice primarily because, well, when I break it, I already probably know how to fix what’s wrong.


  • I’m a grumpy linux greybeard type, so I went with… plain text files.

    Everything is deployed via docker, so I’ve got a docker-compose.yml for each stack, and any notes or configuration things specific to that app is a comment in the compose file. Those are all backed up in a couple of places, since all I need to do is drop them on a filesystem, and bam, complete restoration.

    Reverse proxy is nginx, because it’s reliable, tested, proven, works, and while it might not have all those fancy auto-config options other things have, it also doesn’t automatically configure itself into a way that I’d prefer it didn’t, either.

    I don’t use any tools like portainer or dockge or nginx proxy manager at this point, because dealing with what’s just a couple of config files on the filesystem is faster (for me) and less complicated (again, for me) than adding another layer of software on top (and it keeps your attack surface small).

    My one concession to gui shit for the docker is an install of dozzle because it certainly makes dealing with docker logs simple, and it simplifies managing the ~40 stacks and ~85 containers that I’ve got setup at the moment.


  • That’s a fair assessment. I’ll admit to having a severe case of doomerism when it comes to tech lately, and the levels of shit tech bros will go to to monetize shit has me skeptical there’s any sort of protocol or technology that could be made bro-resistant for more than a short period of time.

    EEE is pretty prevalent and has been a very standard practice with these tech companies for a long time. See: Meta and Threads for a recent example.







  • I think it’s less that it’s “impossible” but rather that it’s expensive.

    Honestly we’ve in general shoved too much shit into the browser that’s not strictly related to just browsing web sites.

    And you “have to” support all the layers and layers and layers of added stuff, or you can’t “compete”.

    But, at the same time, the goals of making a good-enough browser that mostly works and isn’t completely enshittified and captured by corpo big tech interests is a very worthy project and 100% support what they’re doing.





  • Nope, that curl command says ‘connect to the public ip of the server, and ask for this specific site by name, and ignore SSL errors’.

    So it’ll make a request to the public IP for any site configured with that server name even if the DNS resolution for that name isn’t a public IP, and ignore the SSL error that happens when you try to do that.

    If there’s a private site configured with that name on nginx and it’s configured without any ACLs, nginx will happily return the content of whatever is at the server name requested.

    Like I said, it’s certainly an edge case that requires you to have knowledge of your target, but at the same time, how many people will just name their, as an example, vaultwarden install as vaultwarden.private.domain.com?

    You could write a script that’ll recon through various permuatations of high-value targets and have it make a couple hundred curl attempts to come up with a nice clean list of reconned and possibly vulnerable targets.



  • That’s the gotcha that can bite you: if you’re sharing internal and external sites via a split horizon nginx config, and it’s accessible over the public internet, then the actual IP defined in DNS doesn’t actually matter.

    If the attacker can determine that secret.local.mydomain.com is a valid server name, they can request it from nginx even if it’s got internal-only dns by including the header of that domain in their request, as an example, in curl like thus:

    curl --header 'Host: secret.local.mydomain.com' https://your.public.ip.here -k

    Admittedly this requires some recon which means 99.999% of attackers are never even going to get remotely close to doing this, but it’s an edge case that’s easy to work against by ACLs, and you probably should when doing split horizon configurations.