On 21/03/2013 9:21 p.m., Sokvantha YOUK wrote:
> Dear All,
>
> I am working on distributed object cached by size to differences squid
> process. My server is Centos 6 x64 bits with squid 3.3.3.
> I don't know how to calculate maximum disk cache size squid can
> support per process.
>
> Below is my partition size:
> #df
> /dev/sdb1            206424760     65884 195873116   1% /cache1
> /dev/sdd1            206424760     60704 195878296   1% /cache3
> /dev/sde1            206424760     60704 195878296   1% /cache4
> /dev/sdf1            206424760     60704 195878296   1% /cache5
> /dev/sdg1            206424760     60704 195878296   1% /cache6
> /dev/sdh1            206424760     79192 195859808   1% /cache7
> /dev/sdi1            206424760     79200 195859800   1% /cache8
> /dev/sdc1            206424760     60704 195878296   1% /cache2
>
> I use 70% of partition size for squid cache_dir Mbytes size and
> worker=4 so my squid.conf is:
>
> workers 4
> ## 1. Handle small cache objects
> cpu_affinity_map process_numbers=1,2,3,4 cores=2,4,6,8
> cache_dir  rock /cache1/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
> cache_dir  rock /cache2/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
> cache_dir  rock /cache3/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
> cache_dir  rock /cache4/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
> cache_dir  rock /cache5/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
> cache_dir  rock /cache6/squid   170000 max-size=31000
> max-swap-rate=300 swap-timeout=300
>
> ## 2. Handle large object > 32kb < 200MB. The fourth worker handles large file
> if ${process_number}=4
> cache_dir  aufs /cache7/squid/${process_number}         170000 16 256
> min-size=31001 max-size=200000000
> cache_dir  aufs /cache8/squid/${process_number}         170000 16 256
> min-size=31001 max-size=200000000
> endif
>
>
> My question is:
> 1. At section #2, Will the fourth workers be able to handle cache_size
> larger than 300GB?
There are three limits on cache size:
1) The main limit is available RAM. 
http://wiki.squid-cache.org/SquidFaq/SquidMemory
2) Stored object count. 2^27-1 objects per directory. No exceptions. If 
you fill that 200GB with billions of 10 byte objects Squid will run out 
of file numbers before the space is full. But if you try to fill it with 
a lot of 10GB objects Squid will use it all completely and want more.
3) 32-bit and 64-bit filesystem limitations. Squid can't magically use 
more disk space than your OS can supply.
> 2. Section #1, worker number 1 will have 170Gbx6=1020Gb (cache dir
> size)? Will it be able to handle this much cache dir size?
Does it have the RAM?
Rock store is a little unusual in that it has a fixed slot size so the 
file count is a known number and less fuzziness on the RAM requirements 
for indexing.
   (170000/31000) x 6  ==>  ~33 GB cache index + all the other regular 
operating memory. Say 34-36 GB of RAM required by the worker.
Amos
Received on Thu Mar 21 2013 - 13:08:00 MDT
This archive was generated by hypermail 2.2.0 : Fri Mar 22 2013 - 12:00:05 MDT