Abstract
AbstractIndoor Environmental Quality (IEQ) concerns several aspects of environmental comforts, such as thermal, visual and acoustics comfort. In particular, IEQ plays a relevant role in workers’ satisfaction since it strongly influences health, well-being, and productivity. Specifically, it has been demonstrated that the furniture configuration in working spaces affects the occupant’s comfort perception. Nevertheless, IEQ has been either neglected or partially addressed in the context of interior design. The contribution of this paper is to introduce a novel method for furniture layout optimisation in terms of IEQ requirements in multi-occupant offices. In particular, we explore the furniture arrangement task as a Multi-Objective Markov Decision Process (MOMDP), which is solved by a reinforcement learning (RL) agent. The goal is to determine optimal workstation positions that maximise workers’ IEQ satisfaction and functional aspects of working spaces under analysis. Firstly, we formulated the furniture layout task as a MOMDP problem by defining reward functions in terms of thermal, acoustics and visual comfort. Then, we train the RL agent to produce optimal/suboptimal layout patterns through a Q-learning-based algorithm. We conducted experiments in two different offices. The experimental results demonstrated that the proposed multi-objective RL approach is able to determine optimal furniture arrangements that provide a balance among office occupants in terms of IEQ satisfaction. Moreover, numerical results show that the proposed approach can be a valuable tool for evaluating the conformity to the environmental comfort standard of working environments during the furniture layout design phase instead of applying corrections during the post-occupancy evaluation.
Publisher
Springer Science and Business Media LLC
Reference55 articles.
1. Abadi M, Barham P, Chen J, et al (2016) $$\{$$TensorFlow$$\}$$: a system for $$\{$$Large-Scale$$\}$$ machine learning. In: 12th USENIX symposium on operating systems design and implementation (OSDI 16), pp 265–283
2. Akazawa Y, Okada Y, Niijima K (2006) Interactive learning interface for automatic 3d scene generation. In: 7th International Conference on Intelligent Games and Simulation, GAME-ON 2006, pp 30–35
3. Arulkumaran K, Deisenroth MP, Brundage M et al (2017) Deep reinforcement learning: a brief survey. IEEE Signal Process Mag 34(6):26–38
4. Bechikh S, Kessentini M, Said LB et al (2015) Preference incorporation in evolutionary multiobjective optimization: a survey of the state-of-the-art. In: Advances in computers, vol 98. Elsevier, pp 141–207
5. Beranek LL, Blazier WE, Figwer JJ (1971) Preferred noise criterion (pnc) curves and their application to rooms. J Acoust Soc Am 50(5A):1223–1228
Cited by
1 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献