On Friday 30 May 2008 9:59:09 pm Sandro Tosi wrote:
> this is a really annoying bug (sometimes even grave, consider my
> situation: on a 56k line, if I ctrl+c sitecopy, all info are lost and
> then I need to reupload all the files, or reinitialize from remote,
> that are both unacceptable).
>
> The problem is that sitecopy reset to 0 bytes ~/.sitecopy/<site> file
> while uploading (I suppose keeping in memory the list, flushing only
> at end), so if you ctrl+c sitecopy the <site> package information are
> lost.
Hi Sandro,
Thanks for your inputs. This is known issue as upstream also confirmed it..
> A simple workaround is just to create a ~/.sitecopy/<site>.bak when
> upload starts, and to remove when ~/.sitecopy/<site> is flushed from
> memory.
Thanks for tips. Can you prepare patch for that. We can send upstream too.
> Please implement something: we cannot always wait for upstream to fix (it)
True. But, I am in touch with upstream to solve it. Let me look into it and
try from my side (our side indeed!).
--
Cheers,
Kartik Mistry | GPG: 0xD1028C8D | IRC: kart_
Blogs: {ftbfs,kartikm}.wordpress.com
--
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]