diff options
-rw-r--r-- | README.md | 9 |
1 files changed, 7 insertions, 2 deletions
@@ -3,7 +3,7 @@ | |||
3 | gopherproxy is a Gopher (RFC 1436) Web Proxy that acts as a gateway into Gopherspace | 3 | gopherproxy is a Gopher (RFC 1436) Web Proxy that acts as a gateway into Gopherspace |
4 | by proxying standard Web HTTP requests to Gopher requests of the target server. | 4 | by proxying standard Web HTTP requests to Gopher requests of the target server. |
5 | 5 | ||
6 | gopherproxy is a fork of https://github.com/prologic/gopherproxy. | 6 | gopherproxy is a fork of [https://github.com/prologic/gopherproxy](https://github.com/prologic/gopherproxy). |
7 | 7 | ||
8 | Demo: https://gopher.feuerfuchs.dev/ | 8 | Demo: https://gopher.feuerfuchs.dev/ |
9 | 9 | ||
@@ -24,7 +24,12 @@ Demo: https://gopher.feuerfuchs.dev/ | |||
24 | $ gopherproxy | 24 | $ gopherproxy |
25 | ``` | 25 | ``` |
26 | 26 | ||
27 | Then simply visit: http://localhost:8000/ | 27 | Arguments: |
28 | |||
29 | - `-bind=<ip[:port]>`: The interface to bind to (default: "0.0.0.0:8000") | ||
30 | - `-uri=<uri>`: The default address to open in the proxy (default: "floodgap.com") | ||
31 | - `-robots-file=<file>`: robots.txt file to serve to protect Gopher sites from web crawlers (default: "robots.txt") | ||
32 | - `-robots-debug=<true|false>`: Print output about clients ignoring robots.txt (this includes regular browsers) (default: false) | ||
28 | 33 | ||
29 | 34 | ||
30 | ## Development | 35 | ## Development |