Hi Eric, Thank you very much for your reply.
We have thought about the event logging you mentioned in the email, and have couple of questions: 1. The event log should be stored in flash, is this correct? If it is stored, how can I read it? 2. There is an event logging component in Ctp, we are trying to use it and see whether it can solve part of our problems. Do you have any advice of using it? In today's tests we are using c-print to record some events in the application layer. We found the sendDone event had not been signaled in the lower layer when node failed. Thanks again. Xiaoyang On Tue, Oct 29, 2013 at 1:31 AM, Eric Decker <[email protected]> wrote: > > Sorry I'm not going to be more help. > > But…. in real live depoloyments it is critical to have some mechanism > of doing real time logging. This coupled with some degree of paranoid > code allows one to log information and/or take a memory (cpu state) dump > for analysis when thing go wrong like you are talking about. > > This needs to be on board and should rely on a highly robust kernel of > code. I've built this into my motes but that doesn't help what you are > doing. > > sorry. I've been doing this kind of work (embedded) for 20+ years and > unless the node is architected to assist in debugging these kinds of > problems it is very very difficult to figure out what is going on when the > node goes catatonic. > > > > On Mon, Oct 28, 2013 at 8:16 PM, Xiaoyang Zhong <[email protected]>wrote: > >> Hi all, >> >> Our group is writing an application for outdoor testbed using Ctp and >> Deluge. We are using micaz and iris motes. Before deployment, we want to do >> as much test as possible to make sure the application is working fine so >> that least maintenance is needed when deployed. It is working well most of >> the time, but we encountered node failure for certain situations (we repeat >> the tests several times, and node fails every time). >> >> We are trying to simulate the situation in tree routing that, if the >> bottleneck node fails and the rest network cannot find a valid routing to >> base station, the rest of the network should keep running until the >> bottleneck node is recovered. >> >> Our method is to turn off the base station, let the rest network running >> for a time interval (the interval we have tested ranges from tens of >> minutes to overnight), then turn on the base station again. We hope all the >> node can reconnect to the base station. >> >> In every test, one IRIS node would fail to reconnect, and from the >> sniffer, we observed no packets from that node (no routing packet, no data >> packets, no deluge packets, no anything). We think the node is "dead", but >> we don't know what would cause this. Deluge packets are not a problem, >> because we disabled deluge beacons at node booted event. And RAM size is >> also not a problem, our application is about 4K for IRIS node, while the >> RAM for IRIS node is 8K. >> >> Does anyone encountered a problem like this, a node suddenly stops >> working? >> >> Any help will be highly appreciated! >> >> Best, >> Xiaoyang >> >> _______________________________________________ >> Tinyos-help mailing list >> [email protected] >> https://www.millennium.berkeley.edu/cgi-bin/mailman/listinfo/tinyos-help >> > > > > -- > Eric B. Decker > Senior (over 50 :-) Researcher > >
_______________________________________________ Tinyos-help mailing list [email protected] https://www.millennium.berkeley.edu/cgi-bin/mailman/listinfo/tinyos-help
