[Rabbit-dev] ad blocking

Luis Soltero lsoltero at globalmarinenet.com
Wed Dec 15 01:51:47 CET 2010


we use image magik here which works fine for us.  I wouldn't expect imagemagik to be much better since you have to spawn
a fork and load a process for every image you want to resample.

Don't know anything about NCache...

We are running in an environment where are users are using very narrow bandwidth so proxy, image conversion, and caching
has not been an issue for us.

Good luck.

--luis


On 12/14/10 7:31 PM, Ping ni wrote:
> another question, what do you guys think the efficiency about invoking
> GraphMagics for processing picture? is it good? i have a full test
> using load runner tools for it, it looks bad. Concurrency only can
> reach 267 numbers. But if i switch off the gm module and using java
> module for pic compression, it got a good score concurrency about 600
> in 24 hours periods.
>
> but cpu and memory almost reach its limits. there is no extra system
> resources left. how to limit the cpu and memory resources for the pic
> processing module?
>
> another about the cache, is there no other compared solution like
> NCache another open source project?
>
> Thanks
> -NP
>
> On 12/15/10, Ping ni <ping.ni.bupt at gmail.com> wrote:
>> On 12/15/10, Ping ni <ping.ni.bupt at gmail.com> wrote:
>>> I think you can extend the the block listener and write your
>>> processing logistics. it can do what you want.
>>>
>>> about luis question, you are right. actually rabbit has left an
>>> interface for sql. you could put the malware host or ip into the
>>> database. get the list from databases as the map structure for
>>> accelerating efficiency.
>>>
>>> On 12/14/10, Luis Soltero <lsoltero at globalmarinenet.com> wrote:
>>>> Hello All,
>>>>
>>>> Does rabbit have the ability to block ads and malware sites using lists
>>>> from
>>>> aggregator sites such as
>>>>
>>>> http://pgl.yoyo.org/adservers/serverlist.php?hostformat=hosts&showintro=1&mimetype=plaintext
>>>> http://someonewhocares.org/hosts/hosts
>>>> http://www.montanamenagerie.org/hostsfile/hosts.txt
>>>> http://www.hosts-file.net/hphosts-partial.asp
>>>> http://www.mvps.org/winhelp2002/hosts.txt
>>>>
>>>> I know that rabbit has the following blocking facility
>>>>
>>>> ---
>>>> [rabbit.filter.BlockFilter]
>>>> # This is a filter that blocks access to resources.
>>>>
>>>> # return a 403 forbidden for these requests.
>>>> blockURLmatching=(\.sex\.|[-.]ad([sx]?)\.|/ad\.|adserving\.|ad101com-|pagead/imgad|as-us.falkag.net|clicktorrent.info)
>>>> ---
>>>>
>>>> This facility is very flexible but difficult to maintain in an every
>>>> changing internet landscape.  A better approach
>>>> might be to use lists from aggragators who's mission is to keep up to
>>>> date
>>>> lists of sites that offer ads and malware.
>>>>
>>>> My first naive approach at solving this problem was to augment
>>>> /etc/hosts
>>>> on
>>>> our proxy server with lists from the above
>>>> sites. I soon discovered rabbit ignored these.  It seems that rabbit
>>>> uses
>>>> javands to access the DNS service directly to
>>>> do queries ignoring the system resolver.   So replacing /etc/hosts does
>>>> not
>>>> work.
>>>>
>>>> A better solution would be for rabbit check against a preconfigured
>>>> "block"
>>>> list of sites and then return 403 errors
>>>> when the urls containing these hosts names are requested.  It should be
>>>> pretty simple thing to do to query the bad host
>>>> table prior to doing DNS query.  I can see two implementations of this
>>>> approach.
>>>> 1. rabbit reads the bad host table on startup and then keeps an internal
>>>> table for lookups (our current host table has
>>>> over 600K entries so this approach should be manageable)
>>>> 2. a better approach might be to query an sql table for bad hosts prior
>>>> to
>>>> the lookup.  This would be faster and more
>>>> dynamic since the table could be updated automatically from an external
>>>> process.
>>>>
>>>> I think that adding this facility to Rabbit should be pretty easy and
>>>> quite
>>>> valuable to the community.  Those of us
>>>> using rabbit are mostly running in a bandwidth limited environment and
>>>> what
>>>> better way to save bandwidth than to strip
>>>> out ads. This approach also has the benefit protecting users from known
>>>> malware sites.
>>>>
>>>> Anyone have any thoughts on this?
>>>>
>>>> Thanks,
>>>>
>>>> --luis
>>>>
>>>>
>>>> --
>>>>
>>>>
>>>> Luis Soltero, Ph.D., MCS
>>>> Director of Software Development, CTO
>>>> Global Marine Networks, LLC
>>>> StarPilot, LLC
>>>> Tel: 865-379-8723
>>>> Fax: 865-681-5017
>>>> E-Mail: lsoltero at globalmarinenet.net
>>>> Web: http://www.globalmarinenet.net
>>>> Web: http://www.starpilotllc.com
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Rabbit-dev mailing list
>>>> Rabbit-dev at khelekore.org
>>>> http://khelekore.org/cgi-bin/mailman/listinfo/rabbit-dev
>>>>


-- 


Luis Soltero, Ph.D., MCS
Director of Software Development, CTO
Global Marine Networks, LLC
StarPilot, LLC
Tel: 865-379-8723
Fax: 865-681-5017
E-Mail: lsoltero at globalmarinenet.net
Web: http://www.globalmarinenet.net
Web: http://www.starpilotllc.com





More information about the Rabbit-dev mailing list