I prefer a hybrid approach. A document explaining some common things to do and generally the idea behind why the API is structured that way (shows me you actually thought about it, and makes it more logical to find different parts of it without necessarily looking it up), and then an API spec showing all the parameters.
Not to rub it in, but in my forties could be read as almost the entirety of the modern web was developed during my adulthood.
Absolutely nothing... This article literally just says that somebody on an internet forum pointed out that what might happen is that if your account has been around longer than the average lifespan then they'll investigate and maybe terminate it after determining it's no longer owned by the original account owner. Valve today doesn't have the support capacity to perform this kind of investigation. Valve in 50-60 years will be an entirely different beast. This speculation means nothing.
From the stories I've heard from corporate software employees, this does sound like exactly the kind of thing you gotta do to show some manager the guy is buddy-buddy with that they're actually not doing their job. And even then they didn't listen.
We have to work under the assumption that most development is done by inexperienced or, to put it bluntly, bad programmers. I would MUCH rather have bad JS code than bad assembly. One may crash a single tab in my browser, the other may crash my entire computer.
Eh, it's pretty useless without the subscription anyway
Assuming you put everything important in home, that is...
Reminds me of my git commit messages!
Heh nice try but we don't write textbooks or exams for the outliers.
It's the quickest way to prove to yourself that you know what you're doing... Most of the time, anyway...
And the of is an optional field for if they have an onlyfans
Computer monitors are significantly more expensive for the same size and are overkill for the applications TVs are generally used for.
I'm imagining a scenario where you're working on a feature that changes the DB state (e.x. introduces a new DB migration that changes some columns) and the bug is on an unrelated part of the code from your feature. In this hypothetical, going back to the state of the upstream branch would make your local environment non functional, and the bug is on an unrelated part of the code. Fairly specific scenario but hey, you can worktree for that. It's not particularly thorough, though.
I think they meant "all of this is what might happen next"
But yeah even then social media will still continue to somehow have an ever increasing number of "users"
I sympathize with this. People on the spectrum already have a hard time just living, then they have the extra hurdle of having a hard time communicating their hard time. It feels very human to want to hide the struggle, it makes you feel more like you belong in the world and that you are just one of the other billions out there being "normal" and doing "normal" things. That's a longing that I've felt, but I'm fortunate enough to not have felt it to the degree your father seems to have. I hope you and your brother find a way to get through to him.
I personally go for "wizzywig" but to each their own.
Yeah, what do people expect? It's not like they own massive, self-sufficient data centers on several continents.
I have an Ubuntu server with two network interfaces - an ethernet and a WiFi network let's call eth0 and wlan0. So far I've been able to set it up as a router by enabling packet forwarding and then doing some iptables trickery. These are my iptable commands:
iptables -t nat -A POSTROUTING -o eth0 -j MASQUERADE iptables -A FORWARD -i eth0 -o eth0 -j ACCEPT iptables -A FORWARD -i eth0 -o eth0 -m state --state RELATED,ESTABLISHED -j ACCEPT
If I'm understanding correctly, the first command says "if you receive packets from a device, do NAT and then forward them with your IP", the second one says to forward packets from eth0 to eth0, and the last line says "if you get packets back, only accept them if a connection has already been previously established". This Ubuntu server is connected to a router which is connected to a modem that actually has internet access. I've set it up so that my router uses my Ubuntu server as the default gateway during DHCP requests. This works fine, I'm able to use devices to connect to the internet, and if I do a trace route, it first goes to the Ubuntu server, then to the router, then out into the great beyond.
Now, I've run:
iptables -D FORWARD -i eth0 -o eth0 -j ACCEPT iptables -A FORWARD -i eth0 -o wlan0 -j ACCEPT iptables -A FORWARD -i wlan0 -o eth0 -m state --state RELATED,ESTABLISHED -j ACCEPT
Which, if I'm understanding correctly, should forward packets through to the WiFi interface instead, but it isn't working. I'm still able to access other devices on the network but not the open internet. I also tried doing iptables -t nat -A POSTROUTING -o wlan0 -j MASQUERADE
which as far as I can tell is unnecessary, but that didn't do anything. When I do trace route this time, it is able to get to the Ubuntu server but no further. I've also tried doing iptables -L -v
but neither the wlan0 -> eth0 rule or the reverse have any packet count. I also tried doing iptables -A FORWARD -i lan0 -o wlan0 -j LOG --log-prefix "FORWARD: "
to just log it first, but nothing shows up in /var/log/syslog even if I try to connect to the internet from a device.
I'm at a loss here so any help even debugging or if I'm going about this wrong would be greatly appreciated. My ultimate goal is to set up a failover so that if the LAN interface doesn't have a connection, it'll start sending packets through the WiFi interface which will be connected to a different internet connection.
I have a fairly old router that doesn’t support gigabit. I also have a network switch that does support gigabit. If I connect two devices directly to the switch, then connect the switch up to the router, will the connection between the two devices support gigabit? If I’m understanding correctly the router would just act as DHCP server and give the two devices a local IP address, but the actual connection between them wouldn’t go through the router at all.
I'm planning to migrate my email to a different provider, but they don't give much storage, so I was wondering what people would recommend for this kind of setup: basically I'd like to use the new provider as something like a relay. I'd want them to only store an email or two at a time and have some kind of self hosted solution that just grabs the emails from the provider and stores them after deleting them off the provider so it's never storing my entire email history, and also keeps my sent emails somewhere so that I have a copy of it. Ideally I'd wanna be able to set this up with a mail client like NextCloud's.
EDIT: Thanks for the info guys! Very excited to get this all set up
At the moment I have a bunch of self-hosting services hosted in the cloud. I plan to get rid of my cloud resources entirely and run stuff on some server hardware I acquired recently but my ISP doesn't give me a static IP and I'm behind a NAT or whatever it's called (the thing that makes multiple people's home connections be behind a single public IP) so I don't think I can even expose directly to the internet. So my plan is to have a very small and cheap server at a data center and proxy my actual server behind that.
My question is, is there a way that I can set things up so that the same domain can connect directly to the server when I'm at home, and to the proxy when I'm not? The difference would be what connection I'm connected to (my home WiFi vs 5G/others' WiFi). I'm thinking I could maybe run DNS on the server and configure my router to use that as a DNS server, but wouldn't my phone/laptop cache DNS entries? So it'd still try to connect to the local IP even when I'm out.
Hey guys, as I'm sure many of you are already aware there's a couple of bugs that are plaguing the home page. I've made a hotfix for these bugs:
- New posts popping up and pushing all the other posts downwards - this is more prevalent on the larger lemmy instances, you'll just be scrolling and suddenly everything's pushed down because new posts are being added to the top of the page as they're being created
- Default "All" not working - this is more something admins would be aware of but in the site settings you can set the default "Listing Type" to "All" instead of "Local", but if you do this the home page doesn't properly load "All", it'll show you the "Local" feed with "All" selected in the tabs. Seems this feature wasn't implemented correctly in time for 0.17.4
Since 0.18 is still a little while out, and I'm sure the devs are both really busy on it (they also said they won't be making another 0.17 release), I went ahead and made a hotfix for these. I have it up on jcgurango/lemmy-ui:0.17.4-hotfix so if you're using docker you can just upgrade to that image. I'm not really sure how ansible works, so I can't help with that.
Here's the repo I have these changes on for those who want to check or build it themselves: https://github.com/jcgurango/lemmy-ui - I've based it on the v0.17.4 tag on the upstream repo.