|
|
|
|
LEADER |
02642 am a22002653u 4500 |
001 |
103641 |
042 |
|
|
|a dc
|
100 |
1 |
0 |
|a Zhao, Nan
|e author
|
100 |
1 |
0 |
|a Massachusetts Institute of Technology. Department of Architecture
|e contributor
|
100 |
1 |
0 |
|a Massachusetts Institute of Technology. Media Laboratory
|e contributor
|
100 |
1 |
0 |
|a Massachusetts Institute of Technology. Sustainable Design Lab
|e contributor
|
100 |
1 |
0 |
|a Zhao, Nan
|e contributor
|
100 |
1 |
0 |
|a Aldrich, Matthew Henry
|e contributor
|
100 |
1 |
0 |
|a Reinhart, Christoph
|e contributor
|
100 |
1 |
0 |
|a Paradiso, Joseph A.
|e contributor
|
700 |
1 |
0 |
|a Aldrich, Matthew Henry
|e author
|
700 |
1 |
0 |
|a Reinhart, Christoph
|e author
|
700 |
1 |
0 |
|a Paradiso, Joseph A.
|e author
|
245 |
0 |
0 |
|a A Multidimensional Continuous Contextual Lighting Control System Using Google Glass
|
260 |
|
|
|b Association for Computing Machinery (ACM),
|c 2016-07-18T15:44:21Z.
|
856 |
|
|
|z Get fulltext
|u http://hdl.handle.net/1721.1/103641
|
520 |
|
|
|a An increasing number of internet-connected LED lighting fixtures and bulbs have recently become available. This development, in combination with emerging hardware and software solutions for activity recognition, establish an infrastructure for context-aware lighting. Automated lighting control could potentially provide a better user experience, increased comfort, higher productivity, and energy savings compared to static uniform illumination. The first question that comes to mind when thinking about context-aware lighting is how to determine the relevant activities and contexts. Do we need different lighting for reading a magazine and reading a book, or maybe just different lighting for reading versus talking on the phone? How do we identify the relevant situations, and what are the preferred lighting settings? In this paper we present three steps we took to answer these questions and demonstrate them via an adaptive five-channel solid-state lighting system with continuous contextual control. We implemented a multidimensional user interface for manual control as well as an autonomous solution using wearable sensors. We enable a simple set of sensors to manipulate complicated lighting scenarios by indirectly simplifying and reducing the complexity of the sensor-lighting control space using human-derived criteria. In a preliminary user study, we estimated significant energy savings of up to 52% and showed multiple future research directions, including behavioral feedback.
|
546 |
|
|
|a en_US
|
655 |
7 |
|
|a Article
|
773 |
|
|
|t Proceedings of the 2nd ACM International Conference on Embedded Systems for Energy-Efficient Built Environments - BuildSys '15
|