I stopped using Google late last year and it’s pretty eye opening how much freer I feel now. Previously, any searches I made would follow me around. Make a one time search for something I’d see that being advertised later on. As a result I started searching more using private browsing. I’d often forget though and end up being tracked.
Ultimately switching to Firefox and DuckDuckGo I no longer have to do private searches. No more being followed around the internet.
Also I’m not convinced private browsing works. For example I still use it for YouTube but I noticed despite YouTube not knowing who I am, the videos on the home page include some that are very related to my usual videos. I guess they are using IP’s to still deliver relatable videos.
Private browsing keeps your computer from remembering things about what you did. It cannot keep other people’s computers from remembering everything about interacting with you.
It sucks because it’s sometimes (but not very often) useful but it’s not like they are under any obligation to support it or are getting any money from doing it.
I dunno, but I suspect that they aren’t using Google’s cache if that’s the case.
My guess is that the site uses its own scrapper that acts like a search engine and because websites want to be seen to search engines they allow them to see everything. This is just my guess, so it might very well be completely wrong.
At least some of these tools change their “user agent” to be whatever google’s crawler is.
When you browse in, say, Firefox, one of the headers that firefox sends to the website is “I am using Firefox” which might affect how the website should display to you or let the admin knkw they need firefox compatibility (or be used to fingerprint you…).
You can just lie on that, though. Some privacy tools will change it to Chrome, since that’s the most common.
Or, you say “i am the google web crawler”, which they let past the paywall so it can be added to google.
Or, you say “i am the google web crawler”, which they let past the paywall so it can be added to google.
If I’m not wrong, Google has a set range of IP addresses for their crawlers, so not all sites will let you through just because your UA claims to be Googlebot
Depends. Not every site, or its pages, will be crawled by the Internet Archive. Many pages are available only because someone has submitted it to be archived. Whereas Google search will typically cache after indexed.
Well that really sucks because it was often the only way to actually find the content on the page that the Google results “promised”. For numerous reasons - sometimes the content simply changes, gets deleted or is made inaccessible because of geo-fencing or the site is straight up broken and so on.
Yes, there’s archive.org but believe it or not, not everything is there.
Or locked behind 100 pages of unnecessarily paginated content. Seriously, one of the best features that a webpage has over a physical printed page is the ability to search it for what you were looking for… smh:-(.
In a shocking turn of events, google decided once again to make their namesake service worse for everyone.
Legitimately baffling, keeping this feature doesn’t really seem like it would impact anyone except those that use it, while removing it not only impacts those people that already use it, but those who would potentially have reason to in the future.
Cannot think of a single benefit to removing a feature like this.
It is only baffling if you still think that Google’s aim is to help people. At one point they were trying to gain market share and so that was true. It is not anymore.
It was meant for helping people access pages when way back, you often couldn’t depend on a page loading,” Google’s Danny Sullivan wrote. “These days, things have greatly improved. So, it was decided to retire it."
They still go down, Danny. And fairly frequently at that. Y’all are fuckin’ stupid.
I’d say things are much worse than they used to be. Sure, in the past sites would disappear or completely fail more often. But, because most sites were static, those were the only ways they could fail. These days the cache feature is useful for websites that have javascript bugs preventing them from displaying properly, or where the content-management-system still pretends the link works but where it silently just loads different content.
i maintain redirects for old URLs for which the content still exists at another address. i've been doing that since i started working on web sites 20-some years ago. not many take the time to do that, but i do. so there's at least a few web sites out there that if you have a 20 year old bookmark to, chances are it still works.
Fuck. I sometimes use the text-only version to access sites with too many moving elements or when the site is geoblocked or doesn’t respect cookies choices and denies access. So far, it has been the most reliable one for me.
By they way, I just found out that they removed the button, but typing cache:www.example.com into Google still redirects you to the cached version (if it exists). But who knows for how long. And there’s the question whether they’ll continue to cache new pages.
I hope they only kill the announced feature but keep the cache part.
Just today I had to use it because some random rss aggregator website had the search result I wanted but redirected me somewhere completely different…
Quotes are fucking awful now. You have to change the search terms to verbatim now which takes way fucking longer. Google has enshittified almost everything. I’m just waiting for them to ruin Maps.
My guess is that a cached page is just a byproduct when the page is indexed by the crawler. The need a local copy to parse text, links etc. and see the difference to the previous page.
If anything it will keep accelerating the worse quarterly results are as they try to solve their way out of problems they made while still keeping the problems
engadget.com
Active