It’s not just DNS. I have this rule in my firewall:
udp dport 15600 counter drop comment "Block Samsung TV shenanigans"
So far, it has blocked 20575 packets (constituting 1304695 bytes) in 6 days and 20 hours.
It’s not just DNS. I have this rule in my firewall:
udp dport 15600 counter drop comment "Block Samsung TV shenanigans"
So far, it has blocked 20575 packets (constituting 1304695 bytes) in 6 days and 20 hours.
Anticheats can be very invasive, they can theoretically scan all the files inside your computer (whether it is practically done, I don’t know but it surely feels like it’s been done), take screenshots regularly, send your hardware information, etc. So yeah, if you are someone who takes security seriously…
Ah you got my comment wrong! I didn’t mean to suggest Gecko is closed source. I just wanted another web engine that is also open source.
Servo was an experimental ground for Mozilla in some ways (like testing out a new CSS engine and porting it back to Gecko if it works). So it’s quite normal for people to be unaware of it, it was not meant for the public.
But later on it was abandoned by Mozilla and stuck in a limbo, until it got picked up by the Linux Foundation. Now it’s a standalone project and I wish them well. We really need a new FOSS web engine.
I recall reading somewhere the earlier compilers had a hard limit on the length of function names, due to memory constraints.
The modern electronic devices are far more railroaded than it was back in the day tho.
Want to download an application? There’s the App Store. No need to download random .exes from sketchy websites (and learn what a “computer virus” is the hard way)
Downloaded a picture? It’s instantly inside your gallery. Back then we needed to find a folder called “Download” or “My Documents” using something called the Explorer!
iPhone and Android made a lot of things dumber and easier to take in, but I feel like it had a detrimental effect on digital literacy.
A question is a question… Just answer the damn question without judging it.
Better let someone know what the thing does instead of forcing them to experiment with it and (possibly) break it.
Indeed, the Ryzen laptops are very nice! I have one (the 4800H) and it lasts ~8 hours on battery, far more than what I expected from laptops of this performance level. My last laptop barely achieved 4 hours of battery life.
I had stability issues in the first year but after one of the BIOS updates it has been smooth as butter.
If proper SATA ever goes away, I’d wager that there will still be SATA-to-USB adapters on sale. Heck, people still find ways to connect floppy drives to their modern PCs.
I use IPv6 exclusively for my homelab. The pros:
No more holepunching kludge with solutions like ZeroTier or Tailscale, just open a port and you are pretty much good to go.
The CGNAT gateway of my ISP tends to be overloaded during the holiday seasons, so using IPv6 eliminates an unstability factor for my lab.
You have a metric sh*t ton of addressing space. I have assigned my SSH server its own IPv6 address, my web server another, my Plex server yet another, … You get the idea. The nice thing here is that even if someone knows about the address to my SSH server, they can’t discover my other servers through port scanning, as was typical in IPv4 days.
Also, because of the sheer size of the addressing space, people simply can’t scan your network.
This is why I try my damnedest not to write in weakly typed languages.
string
+ object
makes no logical sense, but the language will be like “'no biggie, you probably meant string + string so let’s convert the object to string”! And so all hell breaks loose when the language’s assumption is wrong.
Private addresses don’t necessitate NAT. IPv6 also allows private addresses in the form of fd00::/8
, like fd00:face:b00b:1::1
.
.local
is already used by mDNS
I have a 64-bit computer, it can address up to 18.4 exabytes, but my computer only has 32GB, so I will never use the vast majority that address space. Am I “wasting” it?
You are using the addressing bits in the form of virtual memory. Right now. Unless you run a unikernel system, then in that case you could be right, but I doubt it.
Anyway, this is apples and oranges. IP addresses are hierarchical by design (so you have subnets of subnets of subnets of …), memory addresses are flat for the most part, minus some x86 shenanigans.
Yes they are all “used” but you don’t need them. We are not using 2^128 ip addresses in the world.
But we do need them! The last 64 bits of your IPv6 addresses are randomized for privacy purposes, it’s either that or your MAC address is used for them. We may not be using those addresses simultaneously but they certainly are used.
Despite that, there still are plenty of empty spaces in IPv6, that’s true. But they will still be used in the future should the opportunity arise. Any “wastage” is artificial, not a built-in deficiency of the protocol. Whereas if we restricted the space to 40 bits, there will be 24 bits wasted forever no matter how.
You’re not “wasting” them if you just don’t need the extra bits
We are talking about addresses, not counters. An inherently hierarchical one at that (i.e. it goes from top to bottom using up all bits). If you don’t use the bits you are actually wasting them.
you can gradually make the other bits available in the form of more octets
So why didn’t we make other bits available for IPv4 gradually? Yeah, same issue as that: Forwards compatibility. If you meant that this “IPv5” standard should specify compulsory 64-bit support from the very beginning, then why are you arbitrarily restricting the use of some bits in the first place?
If you’re worried about wasting registers it makes even less sense to switch from a 32-bit addressing space to a 128-bit one in one go
All the 128 bits are used in IPv6. ;)
Every time there’s a “just add an extra octet” argument, I feel some people are completely clueless about how hardware works.
Most hardware comes with 32-bit or 64-bit registers. (Recall that IPv6 came out just a year before the Nintendo 64.) By adding only an extra octet, thus having 40 bits for addressing, you are wasting 24 bits of a 64-bit register. Or wasting 24 bits of a 32-bit register pair. Either way, this is inefficient.
And there’s also the fact that the modern internet is actually reaching the upper limits of a hypothetical 64-bit IPv5: https://lemmy.world/comment/10727792. Do we want to spend yet another two decades just to transition to a newer protocol?
Tell that to your ISP which has fucked their IPv6 deployment up. In my experience IPv6 is actually faster since it bypasses the IPv4 CGNAT.
On busy days my IPv4 connection can get as slow as 15KB/s, now that’s trash.
And we are facing the effects of it as we’re speaking. CGNAT and protocols like TURN were not invented without a reason.
Our network architecture has the tendency to waste IP addresses. A subnet may have 10 devices but have 256 IPs (e.g. a /24 network like 192.168.0.0 to 192.168.0.255) - that’s 246 wasted addresses. This wastage is kinda unavoidable since we’d need to keep our routing tables from being too fragmented.
With that in mind it is entirely possible for 64-bit addressing space to not be enough, unless we revert to methods like NAT which come with their own disadvantages.
We have already used up about one /11 block of the IPv6 internet. That’s 128-11=117 bits. If we replace the standardized /64 subnets of IPv6 with old /24 subnets typical in IPv4 networks, you get 61 bits. That’s dangerously close to the upper limit of a hypothetical 64-bit IPv5 internet.
For SSDs this has historically not been the case, there’s no way in hell you could buy a 1TB SSD within $200 a decade ago.