1 / 57

Wireless Sensor Networks for High Fidelity Sampling

Wireless Sensor Networks for High Fidelity Sampling. Sukun Kim Electrical Engineering and Computer Sciences University of California at Berkeley Committee: David Culler, Ion Stoica, and Gregory Fenves Dissertation Talk May 14, 2007. High Fidelity Sampling.

boydk
Download Presentation

Wireless Sensor Networks for High Fidelity Sampling

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Wireless Sensor Networks for High Fidelity Sampling Sukun Kim Electrical Engineering and Computer Sciences University of California at Berkeley Committee: David Culler, Ion Stoica, and Gregory Fenves Dissertation Talk May 14, 2007

  2. High Fidelity Sampling • Three categories of WSN applications • Monitoring environments • Great duck island [11], Redwood forest [12] • Focus on low-duty cycle and low power consumption • Monitoring objects – High Fidelity Sampling • machine health monitoring [13], condition-based monitoring, earthquake monitoring [14], structural health monitoring [15] • Focus on fidelity (quality) of sample • Interactions with space and objects • Lighting control [16] • Focus on control

  3. Structural Health Monitoring Challenges • High Fidelity Data • High Frequency Sampling with Low Jitter • Time Synchronized Sampling • Large-scale Multi-hop Network • Reliable Command Dissemination • Reliable Data Collection FTSP [8] Mint [9] Broadcast [10]

  4. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion With Rodrigo Fonseca, Prabal Dutta, Arsalan Tavakoli, David Culler, Philip Levis, Scott Shenker and Ion Stoica

  5. Overview • Target applications • Where transfer completion time is more important than latency of each data point • Structural health monitoring, volcanic activity monitoring, bulk data collection • One flow at a time • Reasonable restriction for target applications • Remove inter-path interference • Easier to optimize for intra-path interference

  6. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion

  7. Algorithm • Receiver-initiated • Selective-NACK • Rate Control • Separation of Concerns • Correctness (all packets are delivered) • Performance (achieve high bandwidth)

  8. Reliability 2, 4, 5 2 4 5 4, 9 4, 9 4 9

  9. Rate Control: Conceptual Model Assuming disk model N: Number of nodes, I: Interference range Rate:

  10. Rate Control 1. At each node, Flush attempts to send as fast as possible without causing interference at the next hop along the flow 2. A node’s sending rate cannot exceed the sending rate of its successor 8 6 5 4 d8 = δ8 + H7 = δ8 + δ7 + f7 = δ8 + δ7 + δ6 + δ5 8 7 6 5 4 3

  11. Interference Range > Reception Range However, Jammer Vulnerable to Jammer No problem to Jammer Signal Strength Noise Floor Noise Floor + SNR Threshold Noise Floor + 2 X SNR Threshold SNR Threshold – minimum SNR to decode a packet Jammer – a node which can conflict with the transmission, but cannot be heard

  12. Identifying the Interference Set CDF of the difference between the received signal strength from a predecessor and the local noise floor A large fraction of interferers are detectable and avoidable Fraction of Nodes

  13. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion

  14. Implementation • RSSI is measured by snooping • Information is also snooped • δ, f, D are put into packet header, and exchanged through snooping • δ, f, D take 1 byte each, 3 bytes total • Cutoff • A node i considers a successor node (i− j) an interferer of node i+1 if, for any j > 1, rssi(i+1) − rssi(i−j) < 10 dBm • The threshold of 10 dBm was chosen after empirically evaluating a range of values

  15. Implementation • 16-deep Rate-limited Queue • Enforces departure delay D(i) • When a node becomes congested (depth 5), it doubles the delay advertised to its descendants • But continues to drain its own queue with the smaller delay until it is no longer congested

  16. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion

  17. Packet Throughput of Different Fixed Rates Effective Throughput (pkt/s) Packet throughput of fixed rate streams over different hop counts The optimal fixed rate depends on the distance from the sink

  18. Packet Throughput of Flush Effective Throughput (pkt/s) Effective packet throughput of Flush compared to the best fixed rate at each hop Flush tracks the best fixed packet rate

  19. Bandwidth of Flush Effective Bandwidth (B/s) Effective bandwidth of Flush compared to the best fixed rate at each hop Flush’s protocol overhead reduces the effective data rate

  20. Fraction of Data Transferred in Different Phases • Fraction of data transferred from the 6th hop during the transfer phase and acknowledgment phase • Greedy best-effort routing is unreliable, and exhibits a loss rate of 43.5%. A higher than sustainable rate leads to a high loss rate

  21. Amount of Time Spent in Different Phases • Fraction of time spent in different stages • A retransmission during the acknowledgment phase is expensive, and leads to a poor throughput

  22. Packet Throughput at Transfer Phase Effective Goodput (pkt/s) Effective goodput during the transfer phase Flush provides comparable goodput as a lower loss rate which reduces the time spent in the expensive acknowledgment phase, which increases the effective bandwidth

  23. Packet Rate over Time for a Source • Source is 7 hops away, Data is smoothed by averaging 16 values • Flush approximates the best fixed rate with the least variance

  24. Maximum Queue Occupancy across All Nodes for Each Packet • Flush exhibits more stable queue occupancies than Flush-e2e

  25. Sending Rate at Lossy Link 6 3 4 5 2 1 0 Packets were dropped from hop 3 to hop 2 with 50% probability between 7 and 17 seconds Both Flush and Flush-e2e adapt while the fixed rate overflows its queue

  26. Queue Length at Lossy Link Flush and Flush-e2e adapt while the fixed rate overflows its queue

  27. Route Change Experiment 4 5 • We started a transfer over a 5 hop path • Approximately 21 seconds into the experiment forced the node 4 hops from the sink to switch its next hop • Node 4’s next hop is changed, changing all nodes in the subpath to the root • No packets were lost, and Flush adapted quickly to the change • Flush adapts when the next hop changes suddenly 3a 3b 2a 1a 2b 1b 0

  28. Scalability Test Effective bandwidth from the real-world scalability test where 79 nodes formed 48 hop network Flush closely tracks or exceeds the best possible fixed rate across at all hop distances that we tested Effective Bandwidth (B/s)

  29. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion

  30. Discussion • High-power node • reduces hop count and interference • Not an option on the Golden Gate Bridge due to power and maintenance problems • Interactions with Routing • Link estimator of a routing layer breaks down under heavy traffic

  31. Related Work • Li et al – capacity of a chain of nodes limited by interference using 802.11 • ATP, W-TCP – rate-based transmission in the Internet • Wisden – concurrent transmission without a mechanism for a congestion control • Fetch – single flow, selective-NACK, no mention about rate control

  32. Conclusion • A reasonable assumption (single flow) simplifies a problem (eliminates inter-path congestion control) • Light-weight solution reduces complexity • Overhearing is used to measure interference and to exchange information • Two rules to determine a rate • At each node, Flush attempts to send as fast as possible without causing interference at the next hop along the flow • A node’s sending rate cannot exceed the sending rate of its successor

  33. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion With Shamim Pakzad, David Culler, James Demmel, Gregory Fenves, Steve Glaser, and Martin Turon

  34. Node Layout (1st phase) SF (south) Sausalito (north) • Distance between nodes on the span is either 100ft or 50ft • Initially designed as 150ft • Difference in MicaZ radio output power was up to 7.5dBm 500 ft 1125 ft 4200 ft 246 ft 56 nodes 8 nodes

  35. Environment Fog Strong and salty wind Rapidly changing ... high and scary

  36. Node Battery (4 X 6V Lantern Battery) Bi-directional Patch Antenna Node (Mote + Accelerometer Board)

  37. Node Zip tie around Antenna Extreme Rusting of C-clamp

  38. Base Station Laptop Students At Work

  39. Installation Crawling and Installing Hard Hat Done! Strong Wind Harness However… Ouch Sharp Edge

  40. Table of Contents • Introduction • Flush: A Reliable Bulk Transport Protocol • Algorithm • Implementation • Evaluation • Discussion • Related Work • Conclusion • Deployment at the Golden Gate Bridge • Data from the Golden Gate Bridge • Conclusion

  41. Reliable Data Collection at GGB Data is collected reliably over a 46-hop network, with a bandwidth of 441B/s at the 46th hop

  42. Vibration Data of GGB (1) (2) (3) The vertical modal properties match among (1) simulation model, (2) previous study, and (3) this study

  43. Conclusion • As a concrete example of HFS, SHM is designed, implemented and deployed • Requirements are identified and solutions are proposed • The system satisfied requirements, and provided meaningful data for the research of structural analysis

  44. Bonus – Spectacular Views Questions

  45. Acknowledgement • David Culler • GGB – Shamim Pakzad, James Demmel, Gregory Fenves, Steve Glaser, and Martin Turon • Reliable Data Collection – Rodrigo Fonseca, Prabal Dutta, Arsalan Tavakoli, Philip Levis, Scott Shenker and Ion Stoica • Jaein Jeong, Xiaofan Jiang, Jay Taneja, Jorge Ortiz, Robert Szewczyk, Tom Oberheim, Anthony Joseph, Joe Polastre, Alec Woo, Kamin Whitehouse, Phil Buonadonna

  46. Average Number of Transmissions per node for sending 1,000 packets

  47. Bandwidth at Transfer Phase Effective Goodput (B/s) Effective goodput during the transfer phase Effective goodput is computed as the number of unique packets received over the duration of the transfer phase

  48. Details of Queue Length for Flush-e2e

  49. Memory and Code Footprint

More Related