13ft – A site similar to 12ft.io but is self hosted
The 13 Feet Ladder project is a self-hosted server that bypasses paywalls and ads, allowing access to restricted content from sites like Medium and The New York Times.
Read original articleThe **13 Feet Ladder** project is a self-hosted server designed to bypass paywalls, ads, and other website restrictions, functioning similarly to the service provided by 12ft.io. It is particularly effective for accessing content from platforms like Medium and The New York Times. Key features include its self-hosted nature, the ability to bypass ads and paywalls by mimicking GoogleBot, and a user-friendly interface for inputting URLs. Users can set it up using Docker by cloning the repository and executing specific commands, or they can use Python by installing the necessary packages and running a script. To access restricted content, users can append the desired URL to their server's URL. While the tool facilitates unrestricted access to content, users are encouraged to support content creators whenever possible.
- The 13 Feet Ladder project allows users to bypass paywalls and ads on various websites.
- It can be self-hosted using Docker or Python, providing flexibility in setup.
- The tool mimics GoogleBot to access restricted content.
- Users can easily input URLs to retrieve content through a simple interface.
- Supporting content creators is recommended despite the tool's capabilities.
Related
Surfing the (Human-Made) Internet
The internet's evolution prompts a return to its human side, advocating for personal sites, niche content, and self-hosted platforms. Strategies include exploring blogrolls, creating link directories, and using alternative search engines. Embrace decentralized social media and RSS feeds for enriched online experiences.
Fighting Twitter's censorship of Substack (with a proxy)
Twitter censors Substack links by not displaying preview cards, leading to lower visibility. An author created a tool to host Substack content statically, gaining popularity with 500 articles. Emphasizes internet users' agency in archiving and sharing content independently.
AI crawlers need to be more respectful
Read the Docs has reported increased abusive AI crawling, leading to high bandwidth costs. They are blocking offenders and urging AI companies to adopt respectful practices and improve crawler efficiency.
- Some users express skepticism about the reliability of the tool, noting that it may only work temporarily before being blocked.
- There is a debate about the ethics of bypassing paywalls, with some arguing that it undermines the monetization of journalism.
- Several commenters suggest alternative methods for accessing content, such as using archiving services or browser extensions.
- Users share their experiences with other paywall bypassing services, indicating a general frustration with current options.
- Some advocate for supporting journalism financially rather than seeking ways to bypass payment systems.
Regarding the limitations of this approach, I'm fully aware that it isn't perfect, and it was never intended to be. It was just a quick experiment to see if the concept was feasible—and it seems that, at least sometimes, it is. Thank you all for the continued support.
But perhaps it’s necessary, as it seems Firefox no longer has an about:config option to override the user agent…am I missing it somewhere?
Edit: The about:config option general.useragent.override can be created and will be used for all requests (I just tested). I was confused because that config key doesn’t exist in a fresh install of Firefox. The user agent header string from this repo is: "Mozilla/5.0 (Linux; Android 6.0.1; Nexus 5X Build/MMB29P) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/W.X.Y.Z Mobile Safari/537.36 (compatible; Googlebot/2.1; +http://www.google.com/bot.html)"
It obviously doesn't seem that way to Google, or to the sites providing the content.
They are doing what works for them without ethical constraints (Google definitely, many content providers, eg NYT). Is it fair game to do what works for you (eg. 13ft)?!
https://archive.is/20240719082825/https://www.nytimes.com/20...
The reason is the staff at the NYT appear to be very well versed in the technical tricks people use to gain access.
If you're on iOS + Safari I recommend the "open in internet archive" shortcut, which is actually able to bypass most paywalls.
https://www.reddit.com/r/shortcuts/comments/12fbk8m/ive_crea...
Side note: paywalls are annoying but most publications are often available for free via public library.
For example, NYT is free via my public library. PL offers 3-day subs. A few other decent publications are available as well. Availability of publications is YMMV as well.
There are probably much better and more secure options, but this might be an interesting temporary kludge.
An instruction on how to specify port would be nice.
What sysadmin is so naive to rely on the Googlebot useragent??
Doesn't everyone fetch the official list of Googlebot IPs and then add those to a whitelist?
https://github.com/bpc-clone/bypass-paywalls-clean-filters
Personally I'm not a fan of this attitude. I've read and digested the arguments for it, but, for me, it runs close to "theft".
For example, read the sentence again, but in the context of a restaurant. Sure I wanna support the creators, but what if I just want a single meal and then get on with my day?
Businesses, including news web sites, need to monetize their content. There are a variety of ways they do that.
You are free to consume their content or not. You either accept their monetization method as desirable or you do not.
The "I just want to read one article" argument doesn't fly. If the article is so compelling, then follow their rules for accessing it.
Yes, some sites behave badly. So stop visiting them. There is lots of free content on the web that is well presented and lacks corporate malfeasance. Read some of that instead.
I get that I'm gonna get downvoted to oblivion with this post. HN readers are in love with ad blockers and paywall bypasses. But just because you can do something, just because you think it should be "free, no ads", does not make it right.
Creators create. They get to choose how the world sees their creation. Support it, don't support it, that's up to you. Deciding to just take it anyway, on your terms (however righteous you feel you are) is not ok.
Is there some way to support journalism across publications?
Related
Surfing the (Human-Made) Internet
The internet's evolution prompts a return to its human side, advocating for personal sites, niche content, and self-hosted platforms. Strategies include exploring blogrolls, creating link directories, and using alternative search engines. Embrace decentralized social media and RSS feeds for enriched online experiences.
Fighting Twitter's censorship of Substack (with a proxy)
Twitter censors Substack links by not displaying preview cards, leading to lower visibility. An author created a tool to host Substack content statically, gaining popularity with 500 articles. Emphasizes internet users' agency in archiving and sharing content independently.
AI crawlers need to be more respectful
Read the Docs has reported increased abusive AI crawling, leading to high bandwidth costs. They are blocking offenders and urging AI companies to adopt respectful practices and improve crawler efficiency.