On Aug 29, 2014, at 2:43 AM, Marcus White <roastedseawee...@gmail.com> wrote:
> Thanks John:) > > Some more additional basic questions.. > > 1. For the multi site case, does Swift present a single end point, or > is it two separate regions with two separate IP addresses, and the > client can talk to either to get the data? I think it is the latter? Every proxy server in the cluster will be able to handle requests. There isn't a single "control node" that brokers everything. Often times, all of the proxy servers are put behind a load balancer or geo-dns or something to present a single domain for the cluster. > > 2. Are keystone databases synced across sites? I didnt know of an > option in keystone to do that, hence asking..If we need to > authenticate against keystone on different sites to access the same > object, it has to be that way? That is also my understanding, but I'm not an expert on Keystone options or deployment patterns. > > TIA, > > MW > > > On Wed, Aug 27, 2014 at 11:37 PM, John Dickinson <m...@not.mn> wrote: >> Yup, this is a totally possible failure scenario, and Swift will merge the >> data (using last-write-wins for overwrites) automatically when the partition >> is restored. But you'll still have full durability on writes, even with a >> partitioned global cluster. >> >> --John >> >> >> >> On Aug 27, 2014, at 10:49 AM, Marcus White <roastedseawee...@gmail.com> >> wrote: >> >>> Yup, thanks for the great explanation:) >>> >>> Another question, though related: If there are three regions, and two >>> get "split", there is now a partition. Both the split ones can talk to >>> the third, but not to each other. >>> >>> A PUT comes into one region, and it gets written to the local site. >>> Container information presumably gets updated here, including byte >>> count. >>> >>> Same thing happens on another site, where a PUT comes in and the >>> container information is updated with the byte count. >>> When the sites get back together, do the container servers make sure >>> its all correct in the end? >>> >>> If this is not a possible scenario, is there any case where the >>> container metadata can be different between two zones or regions >>> because of a partition, and independent PUTs can happen, and the data >>> has to be merged? Is that all done by the respective servers(container >>> or account?) >>> >>> I will be looking at the sources soon. >>> >>> Thanks again >>> MW. >>> >>> >>> >>> On Wed, Aug 27, 2014 at 8:13 PM, Luse, Paul E <paul.e.l...@intel.com> wrote: >>>> Marcus- >>>> >>>> Not sure how much nitty gritty detail you care to know as some of these >>>> answers will get into code specifics which you're better off exploring on >>>> your own so my explanation isn't potentially dated. At a high level >>>> though, the proxy looks up the nodes that are responsible for the storing >>>> of an object and its container via the rings. It passes that info to the >>>> storage nodes when it does the PUT request so when the storage node goes >>>> to update the container it's been told "and here are the nodes to send the >>>> container update to". It will send the updates to all of them. >>>> Similarly, once the container server has updated its database it goes and >>>> updates the appropriate account databases. >>>> >>>> Make sense? >>>> >>>> Thx >>>> Paul >>>> >>>> -----Original Message----- >>>> From: Marcus White [mailto:roastedseawee...@gmail.com] >>>> Sent: Wednesday, August 27, 2014 7:04 AM >>>> To: Luse, Paul E >>>> Cc: openstack >>>> Subject: Re: [Openstack] Swift questions >>>> >>>> Thanks Paul:) >>>> >>>> For the container part, you mentioned that node(meaning object >>>> server?) contacts the container server. Since you can have multiple >>>> container servers, how does the object server know which container server >>>> to contact? How and where the container gets updated is a bit confusing. >>>> With container rings and account rings being separate and in the proxy >>>> part, I am not sure I understand how that path works. >>>> >>>> MW >>>> >>>> On Wed, Aug 27, 2014 at 6:15 PM, Luse, Paul E <paul.e.l...@intel.com> >>>> wrote: >>>>> Hi Marcus, >>>>> >>>>> See answers below. Feel free to ask follow-ups, others may have more to >>>>> add as well. >>>>> >>>>> Thx >>>>> Paul >>>>> >>>>> -----Original Message----- >>>>> From: Marcus White [mailto:roastedseawee...@gmail.com] >>>>> Sent: Wednesday, August 27, 2014 5:04 AM >>>>> To: openstack >>>>> Subject: [Openstack] Swift questions >>>>> >>>>> Hello, >>>>> Some questions on new and old features of Swift. Any help would be >>>>> great:) Some are very basic, sorry! >>>>> >>>>> 1. Does Swift write two copies and then return back to the client in the >>>>> 3 replica case, with third in the background? >>>>> >>>>> PL> Depends on the number of replicas, the formula for what we call a >>>>> quorum is n/2 + 1 which is the number of success responses we get from >>>>> the back end storage nodes before telling the client that all is good. >>>>> So, yes, with 3 replicas you need 2 good responses before returning OK. >>>>> >>>>> 2. This again is a stupid question, but eventually consistent for an >>>>> object is a bit confusing, unless it is updated. If it is created, it is >>>>> either there or not and you cannot update the data within the object. >>>>> Maybe a POST can change the metadata? Or the container listing shows its >>>>> there but the actual object never got there? Those are the only cases I >>>>> can think of. >>>>> >>>>> PL> No, it's a good question because its asked a lot. The most common >>>>> scenario that we talk about for eventually consistent is the consistency >>>>> between the existence of an object and its presence in the container >>>>> listing so your thinking is pretty close. When an object PUT is complete >>>>> on a storage node (fully committed to disk), that node will then send a >>>>> message to the appropriate container server to update the listing. It >>>>> will attempt to do this synchronously but if it can't, the update may be >>>>> delayed w/o any indication to the client. This is by design and means >>>>> that it's possible to get a successful PUT, be able to GET the object w/o >>>>> any problem however it may not yet show up in the container listing. >>>>> There are other scenarios that demonstrate the eventually consistent >>>>> nature of Swift, this is just a common and easy to explain one. >>>>> >>>>> 3. Once an object has been written, when and how is the container >>>>> listing, number of bytes, account listing (if new container created) >>>>> etc updated? Is there something done in the path of the PUT to >>>>> indicate this object belongs to a particular container and the number >>>>> of bytes etc is done in the background? A little clarification would >>>>> help:) >>>>> >>>>> PL> Covered as part of last question. >>>>> >>>>> 4. For the global clusters, is the object ring across regions or is it >>>>> the same with containers and accounts also? >>>>> >>>>> PL> Check out the SwiftStack blog if you haven't already at >>>>> https://swiftstack.com/blog/2013/07/02/swift-1-9-0-release/ and there's >>>>> also some other stuff (including a demo from the last summit) that you >>>>> can find googling around a bit too. The 'Region Tier' element described >>>>> in the blog addresses the makeup of a ring so can be applied to both >>>>> container and account rings also - I personally didn't work on this >>>>> feature so will leave it to one of the other guys to comment more in this >>>>> area. >>>>> >>>>> 5. For containers in global clusters, if a client queries the >>>>> container metadata from another site, is there a chance of it getting >>>>> the old metadata? With respect to the object itself, the eventually >>>>> consistent part is a bit confusing for me:) >>>>> >>>>> PL> There's always a chance of getting old "something" whether its >>>>> metadata or data, that's part of eventually consistent. In the face of >>>>> an outage (the P in the CAP theorem) Swift will always favor availability >>>>> which may mean older data or older metadata (object or container listing) >>>>> depending on the specific scenario. If deployed correctly I don't >>>>> believe use of global clusters increases the odds of this happening >>>>> though (again will count on someone else to say more) and its worth >>>>> emphasizing the getting "old stuff" is in the face of some sort of >>>>> failure (or big network congestion) so you shouldn't think of eventually >>>>> consistent as being a system where you "get whatever you get". You'll >>>>> get the latest greatest available information. >>>>> >>>>> MW >>>>> >>>>> _______________________________________________ >>>>> Mailing list: >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>> Post to : openstack@lists.openstack.org >>>>> Unsubscribe : >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>> >>> _______________________________________________ >>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>> Post to : openstack@lists.openstack.org >>> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack