fast.com gives 500 Mbps
fast.com gives 500 Mbps
going to librespeed.org got me 482 down
that makes sense, and I’m looking now. However, the only thing that has anything other than zero in the ‘Real-time rate’ on the router is the computer i’m typing this on, which is at ~30KB/s up and down
I’ve got a coax cable (not fiber) coming into the house, in the USA. My understanding is that there is some amount of shared network with the neighbors.
That is the correct question, and mostly no, I don’t have any specific problem.
The biggest motivator for me looking at it is probably just hobby/interest/how-does-this-work.
That said, my partner and I both work from home ~50% and are often pulling files/data that are a couple GB from the work network, and having those go faster would be nice. Probably the limiting factor in those, though, is the upload from the work network and so faster download for us likely wouldn’t matter, but I’d like to be able to say “I looked into it, honey.”
“DO NOT EVER TURN THIS SERVER OFF - CALL RON” is very good
If you arent an actual journalist who is being personally, specifically hunted then you probably don’t need to take the same precautions as one.
And yea, the guide boils down to “none of these things are 100% safe but they are realistic things you can do that can offer more protection than not doing them.”
Your skimming of the article missed how they do indeed talk about the shortcomings of every suggestion they have. For example, the article also does indeed talk about how you can turn off gps but your phone will still ping towers revealing your location, and goes on to say that you can put your phone in a faraday bag but that isnt practical for most people but is indeed an option if you want to do it.
I use the parental controls on the router to put the roomba in grounded-child mode.
That said, I’m not actually positive it works… it is able to connect to home assistant, so it definitely has local network connectivity, but I haven’t proved to myself that it is actually unable to connect to its remote servers since it isn’t really that big of a deal to me.
it is indeed infrequent, but the modern world has trained me to expect convenience and instant-ness. Last time i wanted a 12-year-old email I was in the car with friends and and to pull it up. it wasn’t anything important at all, to be clear, but i’m hoping to search my 12-year-old emails with the same convenience as last month’s.
I think that that is right that I fundamentally want an archive, not what a normal mail server provides. Part of my thought on looking at mail servers is that those would integrate directly with whatever other front-end/client that I’d normally use, whereas an archive maybe would not.
And regarding archive-specific stuff, I am seeing some things on a search, but I guess i’m wondering if folks here have any recommendations. When I look at , for example, nothing comes up for email archive, just for email servers. That, plus what I see when searching, makes me think that the archive-specific stuff is either oriented to business or oriented to a CLI (like NotMuch, which was mentioned in the discussion here and does look cool).
This looks like a good backend for sure, but the web frontends look a little lacking and I’m not seeing anything about a mobile frontend (other than if a web one was up, which would be fine). Have you tried any of the web frontends?
This article isnt about how emails associated with logins got released in a breach, but that documents that are uploaded to the archive are stamped with the email address of the account that uploaded it and that can be viewed by anyone who downloads the document.
So in standard, everyday use of the site, email addresses are being revealed and are associated with the actions of that person. Like if I upload a copy of the manual for my washing machine or something, which is a more benign example, my email is linked to that document now.
Then combine this with (1) the internet archive says in multiple spots that they dont reveal this info anywhere, and (2) the issue has been raised to the organization, and it becomes more of a specific negligence from them.
This article isnt about how emails associated with logins got released in a breach, but that documents that are uploaded to the archive are stamped with the email address of the account that uploaded it and that can be viewed by anyone who downloads the document.
So in standard, everyday use of the site, email addresses are being revealed and are associated with the actions of that person. Like if I upload a copy of the manual for my washing machine or something, which is a more benign example, my email is linked to that document now.
Then combine this with (1) the internet archive says in multiple spots that they dont reveal this info anywhere, and (2) the issue has been raised to the organization, and it becomes more of a specific negligence from them.
The nsa wants to watch people who are watching the pornhub video of someone else watching porn. The third level there is more difficult to find
The folks who found it are presenting at Defcon this weekend, according to the article.
I imagine some of the industry press (i.e. Wired) are just looking through the Defcon agenda to figure out what to write. I saw two or three other articles about hacks or exploits and things like that that also mentioned it was bring presented at Defcon.
Unrelated to your actual post (plan to read later), but is your RSS busted? The rss link on the webpage gives a 404 and my RSS reader is erroring on it as well…
Ive got this working with Caddy and Adguard
I use Caddy as my reverse proxy. It is running on the machine in the basement with all the different docker-container-services on different ports. My registrar is set up so that *.my-domain.com goes to my IP.
Caddy is then configured for ‘service-a.my-domain.com’ to port 1234, and the others going to their ports. This is just completely standard reverse proxy.
For some subdomains (i.e. different services) ive whitelisted only the local network. There is some config for that.
Im pretty sure that I also have to have adguard do a dns rewrite on the local network as well. That is, adguard has a rewrite for ‘*.my-domain.com’ to go to 192.168.0.22 (the local machine with caddy). I think i had to do this to ensure that when the request gets to caddy it is coming from the local whitelisted network rather than my public IP (which changes every couple months, but could be more).
Everyone who downvoted me didnt read the article, or didnt read what i said, or didnt read op, or something, i dont remember what they didnt read but they cannot be real because the only way to disagree with me is to not have read something or other (or did read it, cant remember which)
I read the fun blogpost that is not an academic paper and ive downvoted you. Does that mean i dont actually exist or that u dont actually exist???
I turned off QoS and immediately am getting 930 on speedtest.net from the desktop browser!
Also, very helpful to know Issue 1 here. I assumed that the router would be the best spot to test since it is farthest upstream (other than the modem). I didn’t know it could pass traffic faster than it can decode, but that makes sense that people would have tried to make that the case. The router is still getting ~500 Mbps while the browser is much closer to the full 1000.