• 0 Posts
  • 32 Comments
Joined 1 year ago
cake
Cake day: July 7th, 2023

help-circle
  • The reverse is also true. Any dev wanting to contribute to Linux in rust which linus himself allowed (despite his silence on this matter) are just going to have to deal with constant headache trying to maintain compatibility with the C interfaces which the devs keep breaking. Either they should’ve never allowed rust in the kernel or they should force devs to at least act in good faith and collaborate (and any that refuse to, well they should be ousted because they can’t behave responsibly). This entire situation is so toxic and I see that as a failure in leadership. That zfs comment is also a little toxic but I don’t think it’s a direct quote. It also doesn’t seem like a fair comparison because from what I can tell zfs isn’t even part of the kernel code base and due to legal reasons cannot be. While it would be great for the kernel not to break it, it is, for all intents and purposes an external project. This rust debacle is different because it’s rust kernel devs and c kernel devs both operating in the same project and trying to find some kind of alignment. To me it seems like there’s enough of an acknowledgment of the value of memory safety that rust support was considered but there’s no authority figure actually supporting it or defending the devs that were invited to actually contribute in it. What a mess.


  • This specific talk was about defining shared common interfaces so these different groups could work together and the guy who actually talked him into stepping down essentially said “I’m gonna keep writing C and if that breaks your rust stuff that’s not my problem”. This isn’t about convincing the c devs to write rust it’s about convincing them to work together when some of them seem to have made up their mind to sabotage rust support (either through indifference or willful interface regressions). Personally I’m more ashamed what this points to for someone new wanting to come in contribute to Linux.


  • Man, despite loving Foss this whole debacle is so disillusioning for anyone that ever wanted to pivot to working on it full time. You don’t have to agree with people wanting to try new things, but the bare minimum is not to spew vitriol to keep them quiet or claim you’ll break their stuff and that’s their problem because they aren’t doing things the same way as you but still depend on a shared ecosystem. All we have to do is be bloody polite to each other and build cool sh*t, why is that so hard. All the best to the Linux rust rewrite for these folks, but to me it just feels like both projects are losing here. Linux losing the passion and drive for adopting more modern stuff and all the folks with that drive opting to restart from scratch because too many people refuse to get along.










  • She.

    And how are they keeping anything together. Market share isn’t substantially better than before and rather than focusing on the product mozilla was created for they keep pivoting to weird BS like this AI grab. I actually think market shares gone up recently… cause google pushed through manifestv3. That would’ve happened even if mozilla did nothing. I think mozillq is still the better browser but that sure as hell doesn’t seem to be because of whose in charge.







  • In general yes. You can think of each container in a docker network as a host and docker makes these hosts discoverable to each other. Docker also supports some other network types that may not follow this concept if you configure them as such (for example if you force all containers to use the same networking stack as one container (I do this with gluetun so I can run everything in a vpn) all services will be reachable only from the gluetun host instead of individual service hosts).

    Furthermore services in a container are not exposed outside of it by default. You must explicitly state when a port in a container is reachable by your host (the ports: option).

    But getting back to the question at hand, what you’re looking for is a reverse proxy. It’s a program that accepts requests from multiple requested and forwards them somewhere else. So you connect to the proxy and it can tell based on how you connect (the url) whether to send the request to sonarr or radarr. http://sonarr.localhost and http://radarr.localhost will both route to your proxy and the proxy will pass them to the respective services based on how you configure it. For this you can use nginx, but I’d recommend caddy as it’s what I’m using and it makes setting up things like this such a breeze.