Maybe you could make use of "Virtual Keyspaces". See this wiki for the idea: https://github.com/rantav/hector/wiki/Virtual-Keyspaces
Bye, Norman 2011/2/17 Frank LoVecchio <fr...@isidorey.com>: > Why not just create some sort of ACL on the client side and use one > Keyspace? It's a lot less management. > > On Thu, Feb 17, 2011 at 12:34 PM, Mimi Aluminium <mimi.alumin...@gmail.com> > wrote: >> >> Hi, >> I really need your help in this matter. >> I will try to simplify my problem and ask specific questions >> >> I am thinking of solving the multi-tenancy problem by providing a separate >> cluster per each tenant. Does it sound reasonable? >> I can end-up with one node belongs to several clusters. >> Does Cassandra support several clusters per node? Does it mean several >> Cassandra daemons on each node? Do you recommend doing that ? what is the >> overhead? is there any link that explain how to do that? >> >> Thanks a lot, >> Mimi >> >> >> On Wed, Feb 16, 2011 at 6:43 PM, Mimi Aluminium <mimi.alumin...@gmail.com> >> wrote: >>> >>> Hi, >>> We are interested in a multi-tenancy environment, that may consist of up >>> to hundreds of data centers. The current design requires cross rack and >>> cross DC replication. Specifically, the per-tenant CFs will be replicated 6 >>> times: in three racks, with 2 copies inside a rack, the racks will be >>> located in at least two different DCs. In the future other replication >>> policies will be considered. The application will decide where (which racks >>> and DC) to place each tenant's replicas. and it might be that one rack can >>> hold more than one tenant. >>> >>> Separating each tenant in a different keyspace, as was suggested >>> in previous mail thread in this subject, seems to be a good approach >>> (assuming the memtable problem will be solved somehow). >>> But then we had concern with regard to the cluster size. >>> and here are my questions: >>> 1) Given the above, should I define one Cassandra cluster that hold all >>> the DCs? sounds not reasonable given hundreds DCs tens of servers in each >>> DC etc. Where is the bottleneck here? keep-alive messages, the gossip, >>> request routing? what is the largest number of servers a cluster can bear? >>> 2) Now assuming that I can create the per-tenant keyspace only for the >>> servers that in the three racks where the replicas are held, does such >>> definition reduces the messaging transfer among the other servers. Does >>> Cassandra optimizes the message transfer in such case? >>> 3) Additional possible solution was to create a separate clusters per >>> each tenant. But it can cause a situation where one server has to run two or >>> more Cassandra's clusters. Can we run more than one cluster in parallel, >>> does it means two cassandra daemons / instances on one server? what will be >>> the overhead? do you have a link that explains how to deal with it? >>> >>> Please can you help me to decide which of these solution can work or you >>> are welcome to suggest something else. >>> Thanks a lot, >>> Mimi >>> >>> >>> >>> >>> >>> >>> > > > > -- > Frank LoVecchio > Senior Software Engineer | Isidorey, LLC > Google Voice +1.720.295.9179 > isidorey.com | facebook.com/franklovecchio | franklovecchio.com >