Type of Document Master's Thesis Author Chong, Justin Brandon Author's Email Address firstname.lastname@example.org URN etd-09212008-181754 Title Activity Recognition Processing in a Self-Contained Wearable System Degree Master of Science Department Electrical and Computer Engineering Advisory Committee
Advisor Name Title Martin, Thomas L. Committee Chair Jones, Mark T. Committee Member Lockhart, Thurmon E. Committee Member Keywords
- Singular Value Decomposition
- Activity Recognition
Date of Defense 2008-09-12 Availability unrestricted AbstractElectronic textiles provide an eective platform to contain wearable computing elements, espe-
cially components geared towards the application of activity recognition. An activity recogni-
tion system built into a wearable textile substrate can be utilized in a variety of areas including
health monitoring, military applications, entertainment, and fashion. Many of the activity
recognition and motion capture systems previously developed have several drawbacks and lim-
itations with regard to their respective designs and implementations. Some such systems are
often times expensive, not conducive to mass production, and may be dicult to calibrate.
An eective system must also be scalable and should be deployable in a variety of environ-
ments and contexts. This thesis presents the design and implementation of a self-contained
motion sensing wearable electronic textile system with an emphasis toward the application of
activity recognition. The system is developed with scalability and deployability in mind, and
as such, utilizes a two-tier hierarchical model combined with a network infrastructure and
wireless connectivity. An example prototype system, in the form of a jumpsuit garment, is
presented and is constructed from relatively inexpensive components and materials.
Filename Size Approximate Download Time (Hours:Minutes:Seconds)
28.8 Modem 56K Modem ISDN (64 Kb) ISDN (128 Kb) Higher-speed Access jchongthesis.pdf 4.35 Mb 00:20:08 00:10:21 00:09:03 00:04:31 00:00:23
If you have questions or technical problems, please Contact DLA.