From: Leonardo Rodrigues Magalhães <leolistas@solutti.com.br>
> lorenor escreveu:
> > I'm searching for a method to cache websites automatically with squid.
> > The goal is to give squid a list of URLs and the proxy will cache the
> > sites.
> > I know only one way to cache a site. A client have to make a request.
> > But is there another way without client interaction?
> No, squid has no mode do that automagically.
> But ..... with some linux clients, wget for example, you can easily do that
> cd /tmp/garbage
> sites.txt should contain URL of the sites you wanna to fetch
> www.onesite.com
> www.othersite.com
> www.anything.com
> export http_proxy=http://your.squid.ip.box:3128
> wget -i sites.txt --mirror
> that should fetch in mirror style (EVERYTHING) from the informed sites and
> save them under the directory you started wget. Depending on the amount of data,
> that could take a long time to run. You can probably erase everything after wget
> finishes, but it may be intelligent to keep the files and run mirror again some
> days after, which will make MUCH less traffic being generated.
> and, in the end of the process, squid should have cached everything that is
> cacheable according to site configurations and your caching parameters as well.
> squid has no automatic mode for doing that, but that can be easily done with
> wget.
You can use the '--delete-after' option to cleanup automaticaly..
And maybe '--limit-rate' if you don't want to hammer the servers...
JD
Received on Wed Feb 11 2009 - 09:59:44 MST
This archive was generated by hypermail 2.2.0 : Wed Feb 11 2009 - 12:00:01 MST