Am 01.08.2012 15:21, schrieb Paolo Bonzini: > Il 01/08/2012 15:09, Kevin Wolf ha scritto: >>> It's not about source vs. target, it's about what block device the _job_ >>> is attached too. The source is always going to be special because you >>> must do "block-job-resume source". >> >> This whole concept of a block job being attached to a single bs is >> flawed. Because in the general case it isn't. And probably it should >> have been a BackgroundJob rather than a BlockJob from the beginning, >> because I'm sure there are use cases outside the block layer as well. >> >> We really manage to get every single point messed up. :-( > > The balance between overengineering and messing up is unfortunately very > difficult to strike. Yes, migration is a BackgroundJob in itself. But > how far would we want to go specializing the monitor interface for > BackgroundJob subclasses, considering after all this is C? > > If we had a tiny C core and everything above it written in {dynamic > language of the day} (I vote for Smalltalk, FYI) perhaps we could be > more cavalier in creating new classes and management abstractions, but > keeping it simple has its advantages. > > No matter how much we messed things up, so far we decently managed to > refactor our way out (and screw up even more the next thing).
Well, the thing that disturbs me is having a BlockJob attached to a single block device. Jobs can and do operate on multiple images and should probably affect both BDSs the same. For example, bs->in_use should probably be set for both (Haven't checked in detail yet, do we have bugs here? Patch 27 doesn't call bdrv_set_in_use at least. It starts to matter when we get named targets.) So my wish was to get rid of the "main bs" in job->bs and move all needed BDSs to the subclass. Which, I then noticed, leaves you with a mere BackgroundJob. >>> is_read tells you where the error was. >> >> Rather indirect, but yes, for the mirror it works. >> >>>> Also, if you miss the event (e.g. libvirt crashed), can you still tell >>>> which bs caused the error? >>> >>> No, but how is it different from "can you still tell which bs in the >>> chain caused the error"? Which you cannot tell anyway even from the >>> event parameters we have had so far. >> >> It isn't different. We really should report the exact BDS. In practice, >> management tools care about ENOSPC which is always the top-level BDS, >> and call the admin for help in other cases. The admin can try manually >> which file is at fault, but I suppose he would be very glad to be told >> by the error message which file it is. > > Yes, and we can add it in the future as an additional argument to the > event and of query-{block,block-jobs} output. At which point we are ok > with having the target iostatus in BlockJob. Yup, target iostatus along with an ID of the BDS (QOM path? Or will all of them be named eventually?) >>>> Do we need another BlockJobInfo field for the >>>> name (*cough* ;-)) of the failed bs? >>>> >>>> If I understand it right, error reporting on the source and on the >>>> target would be completely symmetrical then, which I think is a good thing. >>> >>> Yeah, it would, but job->bs _is_ special anyway. >>> >>>> job->bs makes one image special, which isn't really useful for a generic >>>> interface. So we should keep the fact that it exists internal and get >>>> rid of it sometime. >>>> >>>>> + if (action == BDRV_ACTION_STOP) { >>>>> + block_job_pause(job); >>>>> + block_job_iostatus_set_err(job, error); >>>>> + if (bs != job->bs) { >>>>> + bdrv_iostatus_set_err(bs, error); >>>>> + } >>>>> + } >>>>> >>>>> where the bdrv_iostatus_set_err is mostly to "prepare for the future" >>>>> usage of named block devices. >>>> >>>> Again, I'd make it unconditional to get symmetric behaviour. >>> >>> Not possible, because existing clients may expect "iostatus: >>> {nospace,failed}" to imply a runstate of "not running (i/o error)". >> >> Did we make such guarantees? Does libvirt actually make use of it? > > I'm not sure libvirt relies on it, but I think it's a reasonable > expectation. Possibly, but I don't think anyone should rely on it. You first get the event, or notice that the VM is stopped, and then check query-block. Doing it the other way round and then inferring the VM state from query-block sounds highly unlikely. >>>>> Hmm, but I'm a bit wary of introducing such a big change now. We know >>>>> what it makes nicer, but we don't know of anything irremediably broken >>>>> without them, and we haven't thought enough of any warts it introduces. >>>> >>>> On the one hand, I can understand your concerns, but on the other hand, >>>> introducing an API now and then making such a big change afterwards >>>> scares me much more. >>> >>> One example of the doubts I have: is iostatus a BlockBackend or a >>> BlockDriverState thing? If it a BlockBackend thing, does the target >>> device have an iostatus at all? >> >> I think it's better to have it in BlockDriverState, but in my >> imagination the target is a BlockBackend anyway. > > Great, 0/2. :) My rhetorical questions didn't have the outcome I hoped for. Heh. :-) I was thinking that when you do it for backing files, you automatically have to use the BDS, but the iostatus/pointer model works as well, so yeah, maybe BB is better. But why wouldn't you use a BlockBackend for the target? Kevin