Massive Multi Site

Massive Multi Site

Thursday 19 February 2004 8:41:24 am - 2 replies

Author Message

Bård Farstad

Friday 20 February 2004 2:03:49 am

Interesting. The only thing which I can think of which will cause overhead here is that you need to enable all 1000 siteaccesses. This means that you will use some memory for these 1000 lines. But this overhead should be much smaller than having 1000 installations by far. Each site can share the same php code and this will be cached really well.

Haven't tested though. It would be interesting to hear about some test results ( mabye create a script which generates these siteaccesses for a test first? )

--bård

Documentation: http://ez.no/doc

Björn Dieding@xrow.de

Friday 20 February 2004 8:32:02 am

The final solution could look like this...

1.) catch all request on one ip, because we do not want apache to blow up.
2.) use mod perl to provide a fast mapping from external/unkown hosts to a common internal one
3.) do the usual mod_rw stuff
4.) prodive a little kernel hack to dynamicly match against a internal uri :-)

Looking for a new job? http://www.xrow.com/xrow-GmbH/Jobs
Looking for hosting? http://hostingezpublish.com
-----------------------------------------------------------------------------
GMT +01:00 Hannover, Germany
Web: http://www.xrow.com/

You must be logged in to post messages in this topic!

Powered by eZ Publish™ CMS Open Source Web Content Management. Copyright © 1999-2014 eZ Systems AS (except where otherwise noted). All rights reserved.