Preprint
Review

Human Factors Considerations for Quantifiable Human States in Physical Human-Robot Interaction: A Literature Review

Altmetrics

Downloads

295

Views

180

Comments

0

A peer-reviewed article of this preprint also exists.

This version is not peer-reviewed

Submitted:

01 June 2023

Posted:

02 June 2023

You are already at the latest version

Alerts
Abstract
As the global population rapidly ages with longer life expectancy and declining birth rates, the need for healthcare services and caregivers for older adults is increasing. Current research envisions addressing this shortage by introducing domestic service robots to assist with daily activities. The successful integration of robots as domestic service providers in our lives requires them to possess efficient manipulation capabilities, provide effective physical assistance, and have adaptive control frameworks that enable them to develop social understanding during human-robot interaction. In this context, human factors, especially quantifiable ones, represent a necessary component. The objective of this paper is to conduct an unbiased review encompassing studies on human factors studied in research involving physical interactions and strong manipulation capabilities. We identified the prevalent human factors in physical human-robot interaction (pHRI), noted the factors typically addressed together and determined the frequently utilized assessment approaches. Additionally, we gathered and categorized proposed quantification approaches based on measurable data for each human factor. We also formed a map of common contexts and applications addressed in pHRI for a comprehensive understanding and easier navigation of the field. We found out that most of the studies in direct pHRI (when there is direct physical contact) focus on social behaviors with belief being the most commonly addressed human factor type. Task collaboration is moderately investigated, while physical assistance is rarely studied. In contrast, indirect pHRI (when the physical contact is mediated via a third item) studies often involve industrial settings, with physical ergonomics being the most frequently investigated human factor. More research is needed on human factors in direct and indirect physical assistance applications, including studies that combine physical social behaviors with physical assistance tasks. We also found that while the predominant approach in most studies involves the use of questionnaires as the main method of quantification, there is a recent trend that seeks to address quantification approaches based on measurable data.
Keywords: 
Subject: Computer Science and Mathematics  -   Robotics

1. Introduction

As the global population continues to age rapidly due to increased life expectancy and declining birth rates, the demand for healthcare services and caregivers for older adults is growing [1]. In Japan, over 30% of the population is already over 60, while it is projected that by 2050, one in four individuals in North America and Europe could be aged 65 or older [2]. From 2015 to 2050, the proportion of the global population aged over 60 years is estimated to almost double from 12% to 22% [1].
As a result, there is an increasing demand for solutions that enhance the quality of life and independence of older adults to address the looming shortage of caregivers [3]. To achieve this, the Society 5.0 vision aims to introduce robots as domestic service providers to assist with daily living activities and create novel values in healthcare and caregiving domains [4]. Given that caregiving is closely associated with physical interaction tasks in daily activities such as showering, eating, toileting, and transferring [5], robots introduced into these application domains must be capable of efficient physical support. It is noteworthy that physical support does not involve physical assistance only, but also emotional support through physical comfort such as patting or hugging.
To ensure the successful integration of robots into these fields, a practical approach would be to equip them with the capability to provide multiple services instead of limiting them to a single task (e.g. a robot that only cleans the floor, or only performs feeding), i.e., robotic multiple service providers. The success of this approach has been demonstrated by the widespread of cellphones and computers which has proven that technology can be harnessed to enhance our daily lives by providing a range of functions through a single device. A crucial characteristic of multiple service robots is the possession of dexterity and manipulation capabilities, as it enables them to perform various tasks efficiently, similar to how humans operate.
Furthermore, a seamless assimilation of robots would require them to develop a social understanding of humans, which will help robots make decisions according to their human partners’ state, such as their level of anxiety, emotional state, physical comfort, social distancing or mental workload. This is anticipated to endow robots with the ability to engage with humans in a socially acceptable manner. As a first step towards achieving this goal, robots need to possess the capability of predicting the human state, i.e. quantification of human factors. It’s worth mentioning that this goal is part of a comprehensive vision to develop adaptive control frameworks that can make optimal decisions by taking into account not only the human state but also environmental conditions and a collaborative task goal.
First and foremost, to aid in the development of multiple service robots, it’s necessary to identify research efforts that have: 1) examined quantifiable and measurable factors that can represent the human state, 2) addressed efficient physical interaction, and 3) exploited robots with strong manipulation capabilities. To this end, our goal in this manuscript is to review studies that have quantifiably assessed human factors in physical human-robot interaction (pHRI) scenarios, with a specific focus on robots that are capable of manipulation, meaning those equipped with multi-degree-of-freedom arm(s).
The definitions of pHRI and related human factors in the literature are often ambiguous. The literature on human factors in human-robot interaction (HRI) is characterized by great diversity in its usage and a lack of agreement on unified definitions or terminologies, as noted in [6]. For this review, we characterize human factors as data drawn from humans that can quantifiably interpret the human state to be used in a human-in-the-loop control framework, such as trust, anxiety, mental workload, physical comfort, and perceived safety. As far as the authors are aware, no previous reviews have focused on investigating the studies that have specifically addressed such human factors considerations in the context of pHRI.
Similar to the ambiguity surrounding the definition of human factors, the field of pHRI also grapples with the difficulty of reaching a consensus on a unified definition. In [7], the authors define pHRI as the domain that focuses on the physical aspects of the interaction between humans and robots in a shared workspace, including factors such as robot speed and distance. Whereas in [8], pHRI referred to a broad area of applications that include cooperation, assistance, teleoperation, and entertainment. In [9], the authors categorized the physical connection between humans and robots as either proximally, where there is an exchange of forces between parties, i.e., physical coupling, or remotely, referring to teleoperation. Further restrictions on the area that pHRI refers to are adopted by [10], where authors considered only physical coupling (direct physical contact) between humans and robots, including those mediated by an object. Building upon the perspective of the field presented in [10] on pHRI, we define indirect pHRI as physical coupling that is mediated by an object and direct pHRI as physical coupling that occurs directly between parties without object mediation.
To gain a comprehensive understanding of our topic of interest, it is imperative to discern research gaps as well as evaluate the advancements achieved in the literature. In pursuit of this, we adopted a novel holistic conceptual model for human factors classification in HRI [11], which enabled us to provide an unbiased overview of the most commonly studied human factors and the approaches used to quantify them. However, to facilitate deeper analysis and enhance our understanding of these factors, we first identified the various contexts in which these human factors have been investigated in pHRI, which can serve as a potential mapping scheme for the field. While this review does not aim to provide a strict systematic literature review, it offers an objective summary of the progress made in human factors in pHRI over the last 15 years. The purpose of this summary is to assist in accelerating the development of a streamlined adaptive control framework for multiple service robots.

1.1. Paper Organization

The rest of this review article is organized as follows:
  • Section 2, Related Work, presents literature reviews with similar objectives to ours, but with different focus areas.
  • Section 3, Research Questions, presents the research questions guiding this review.
  • Section 4, Methodology, presents the reviewing process used to develop a potential mapping scheme for the pHRI field and human factors classification.
  • Section 5, Results, presents the classification results, which provide answers to the research questions formulated and identify trends and gaps in the literature.
  • Section 6, Discussion, proposes suggestions for filling the identified literature gaps.
  • Section 7, Limitations of the review, addresses the validity and limitations of this review.
  • Section 8, Conclusion, provides a summary of the key findings and their implications for future research in pHRI.

2. Related Work

Remarkable efforts have been expended to comprehensively review HRI studies pertaining to human factors. Hopko et al. [12] conducted a systematic literature review, where they assessed the most frequently studied human factors and quantification methods used in industrial shared spaces. Their findings revealed that trust, cognitive workload, and anxiety were the most commonly investigated human factors. While Hopko et al. dealt with perceived safety as a distinct human factor, Rubagotti et al. [7] incorporated related human factors, such as trust, anxiety, and comfort, as sub-factors under the umbrella of perceived safety. Their primary objective was to understand how physical factors, such as distance and robot speed, and different types of robots, such as industrial manipulators, mobile robots, and humanoids, can influence the perception of safety. Similarly, Ortenzi et al. [13] took a holistic approach to address different human factors by utilizing "user experience" as an umbrella term instead of examining each factor separately. Their study focused on reviewing efforts made in handovers between humans and robots. Likewise, Prasad et al. [14] examined previous studies in handshake scenarios to identify the most commonly addressed human factors. Coronado et al. [11] conducted a literature review to provide an overview of the quantification methods used in HRI industrial applications to evaluate performance and measurable dimensions of human factors. Their work resulted in a taxonomy of performance aspects and a holistic conceptual model for the most commonly addressed human factors in HRI. Their review also revealed that, after physical safety, trust was the most frequently addressed human factor that influences human performance. Simone et al. [15] studied the impact of human-robot collaboration on human factors in industrial settings. Their review highlighted that successful human-robot interaction depends on several crucial factors, including trust, usability, and acceptance. Additionally, they concluded that the interaction can affect stress and mental workload either positively or negatively. Lorenzini et al. [16] conducted a narrative review to identify the most effective assessment methods for physical and cognitive ergonomics in workplace environments. Their review focused on available monitoring technologies that endow robots with the ability to adapt online to the mental states and physical needs of workers. Notably, each of the aforementioned reviews found that questionnaires were the predominant assessment approach utilized in studies across diverse scenarios.
Evidently, a dearth of research has specifically examined human factors in pHRI scenarios or investigated the studies that quantified the impact of human factors within domestic contexts. Moreover, past attempts to classify human factors in this domain have faced numerous challenges, necessitating the use of various methodological workarounds. In light of these gaps, we endeavor to provide an objective overview of the progress made in this field, in order to promote the development of adaptive control frameworks for multiple service robots. Efficiently predicting the human state is a crucial first step in this development.

3. Research Questions

Although the literature on HRI has addressed human factors considerations, there has been a notable lack of dedicated attention to pHRI as a specific area of study requiring a structured analysis of trends and developments. In this review paper, we aim to identify the different contexts in which human factors have been studied in the field, with a specific emphasis on those that involve robots with manipulation capabilities. Therefore, by creating a mapping scheme, we can help researchers navigate and better understand the field. As a result, we formulated the following research question:
  • RQ1: What are the existing applications or contexts in the literature, and how can they be categorized?
To answer this question, we will identify the applications and scenarios addressed in the literature on both direct and indirect pHRI.
In Section 2, Related Work, numerous reviews have attempted to overcome the challenge of diverse terminologies used to describe various human factors. Some authors, such as Rubagotti et al. [7], have employed the term "perceived safety," while others, like Ortenzi et al. [13], have used the broader concept of "user experience" to encompass multiple human factors, without delving into specific classifications. Other reviews, like Hopko et al. [12], addressed individual human factors but did not provide clear definitions or outlines.
Similarly, classifying human factors has been a major challenge in this review due to the lack of systematic or standardized definitions in HRI literature, making it difficult to determine the most commonly addressed human factor. For example, some studies use terms like stress [17], strain [18], and frustration [19] without providing clear definitions, making it inaccurate to group them together under one umbrella term. Likewise, legibility [20], and predictability [21] of robot’s motion; and physical ergonomics [22], physical comfort [20], fatigue [23], health and safety [24] are two sets of examples of different human factors that overlap but cannot be lumped under a single comprehensive term within the set. Additionally, perceived enjoyment [19], and enjoyability [25]; and psychological safety, mental safety, and subjective safety [7] are also examples of human factors that are distinct but interconnected. We anticipate that this may have caused an unintentional research bias in favor of human factors that have limited synonyms like trust. Trust has been identified as a common human factor in several HRI reviews and has even been given special attention in dedicated reviews [26].
Furthermore, the challenge of assessing progress in human factors in HRI is exacerbated when studies used generic terms like preference [27] and enjoyment [25] without providing clear definitions for each term. This highlights the importance of defining and classifying human factors in a more precise and standardized way to avoid unintentional research bias. To address this challenge, we explored various classification schemes, such as dividing human factors into cognitive and physical ergonomics, as proposed by Lorenzini [16], or further categorizing them into load and perceptual factors. However, we found that human factors in pHRI, which involved both domestic and industrial scenarios, are more complex and diverse than these schemes can account for, as they may involve emotional factors in addition to cognitive and physical ones.
Therefore, we chose to build upon the holistic conceptual model proposed by Coronado et al. [11] for human factors in HRI. By adopting this model, we aim to develop a standardized approach that can systematically assess the progress of human factors research in HRI, including pHRI scenarios.
Coronado et al. proposed a holistic conceptual model adapted for HRI, based on [28], to reduce confusion between the concepts of usability and user experience in HRI literature. This model illustrates the relationships among usability, user experience, hedonomics, and ergonomics, and identifies four distinct human factor types: cognitive ergonomics, physical ergonomics, belief, and affect. Below we will briefly describe the definition of each concept.
User experience, according to the ISO 9241-11:2018 (ergonomics of human-system interaction) [29], includes "all user’s emotions, beliefs, preferences, perceptions, physical and psychological responses, behaviors, and accomplishments that occur before, during, and after use". Whereas, usability is considered as "the extent to which a system, product or service can be used by specified users to achieve specified goals with effectiveness, efficiency, and satisfaction in a specified context of use".
Ergonomics, which is classified into physical and cognitive ergonomics in this review, aims to mitigate injuries, pain, and suffering, while hedonomics, which is classified into belief and affect in this review, focuses on the pleasant and enjoyable aspects of interactions.
Physical ergonomics encompasses the potential adverse effects on the human body during an interaction, such as postures, repetitive movements, heavy workloads, or forces. Cognitive ergonomics is concerned with designing systems that align with the perceptual and psychological abilities of users. Affect pertains to emotional-related terms, whereas belief refers to cognitive responses that can trigger emotions (i.e., affective responses).
By utilizing this conceptual model, we anticipate addressing the deficiency of unified definitions of human factors by classifying them under one of the four human factor types. Following the naming convention presented in Coronado et al., we use the term measurable dimensions to refer to human factors that can be quantified and measured, such as trust, anxiety, and fatigue. In contrast, we use human factor types to refer to broader categories of human factors, such as cognitive ergonomics, physical ergonomics, belief, and affect. We are confident that this approach will facilitate the investigation of relationships between measurable dimensions within each human factor type. This will allow us to gain a more nuanced understanding of how different aspects of human factors interact with each other.
In order to identify the most common human factor type addressed in pHRI studies that involve robots with manipulation capabilities, we formulated the following research question:
  • RQ2: What are the most commonly studied human factors that can be used to develop quantitative measures of the human state?
We hypothesize that belief will be the answer to this question, as it encompasses trust as one of its measurable dimensions, which is consistently identified as one of the most frequently addressed human factors in HRI.
Quantification of human factors based on measurable data plays a key role in the development of adaptive control frameworks that can effectively predict human states and make decisions based on them. As a result, we have formulated the following research question:
  • RQ3: What are the quantification approaches employed to evaluate human factors?
To identify the quantification approaches based on measurable data, we will categorize the gathered studies into two categories based on the objectives of the studies that addressed human factors. We call this categorization Human Factor Usage which includes impact and quantification. The impact category explores the effect of manipulating robot parameters on human factors, i.e. the impact of those manipulations, while the quantification category will explore proposed quantification approaches for human factors or their correlation with measurable entities such as physical data (e.g., body posture and interaction forces) and physiological data. We hypothesize that, based on the literature, questionnaires will be the most commonly used approach for quantifying human factors in pHRI. Our objective is to identify promising quantification techniques that have been introduced or adapted specifically for pHRI.

4. Methodology

The methodology used to conduct this review follows the guidelines proposed by Petersen et al. [30], which were adapted to the context of HRI by Coronado et al. [11]. The methodology consists of several steps:
  • Justifying the need for conducting a literature review, as described in Section 1, Introduction.
  • Defining the research questions, as explained in Section 3, Research Questions.
  • Determining a search strategy, assessing its comprehensiveness, establishing selection criteria, and conducting a quality assessment. These steps are detailed in this section.
  • Extracting data from the identified studies, organizing and categorizing the information obtained, and visualizing the results. These steps are detailed in Section 5, Results.
  • Addressing the limitations of the review by conducting a validity assessment, as described in Section 7, Limitations of the Review.

4.1. Search Strategy

Given that pHRI is a relatively young field and the overall population is not well-known beforehand, conducting a comprehensive database search for human factors is challenging. Unlike more established fields, the literature on pHRI is not yet fully developed, which makes it harder to determine suitable search terms. Additionally, as noted earlier, there is a lack of standardization in the reporting of human factors in HRI, which further complicates the search process. To address this challenge, we employed a combination of search strategies to generate a suitable set of keywords for a comprehensive and effective search string.
As a first step, a combination of manual search [31] and snowballing [32] was used. We identified relevant studies from the references and citations of the following HRI reviews that specifically addressed human factors: [7,12,15,33,34]. In addition to examining the proceedings of the following 2022 conferences which were not indexed yet at the time of writing this manuscript: IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), ACM/IEEE International Conference on Human-Robot Interaction (HRI), IEEE International Conference on Intelligent Robots and Systems (IROS), and IEEE International Conference on Advanced Robotics and Its Social Impacts (ARSO). Furthermore, we included some prominent papers based on expert recommendations.
Initially, keywords were extracted from the set of papers obtained from the previous step, which formed the initial set of papers. However, it was found that barely half of them used the word "physical" in any of their search fields, despite the physical coupling occurring in the interaction. Moreover, the most common words detected in the papers were too generic for a database search, such as “human-robot interaction”. Additionally, the term "human factors" was not commonly used in the papers, as researchers would not necessarily use it when evaluating factors such as trust [26,35], anxiety [36], or other related concepts [37].
The search string was developed based on combining the generic term detected and the human factors terms identified in the initially considered studies, such as comfort, cognitive load, mental load, fatigue, stress, anxiety, fear, acceptance, perception, emotions, trust, safety, ergonomics, and hedonomics. The string was refined by excluding topics covered under the broad domain of HRI, such as wearables, teleoperation, telepresence, rehabilitation, and autonomous vehicles. The search was conducted on Scopus and Web of Science search engines, and the resulting papers were evaluated using a test-set approach. Each iteration was assessed by checking for the presence or absence of the initial set of papers obtained from the manual and snowballing search approaches. We ended up with 711 papers, after the removal of duplicates.

4.2. Study Selection and Quality Assessment

The study selection process involved the following steps:
  • Initial screening: Abstracts of papers were checked against the inclusion criteria, with full texts read for papers that were in doubt. This resulted in 128 papers.
  • Full-text assessment: Quality assessment was conducted during this phase to ensure that each paper included information relevant to the research questions. Exclusion criteria were applied, resulting in 99 papers being included in the study.
The following points summarize the characteristics of each study included in this review, i.e., inclusion criteria:
  • Physical coupling between humans and robots takes place, direct or indirect pHRI.
  • Robot used has manipulation capabilities (i.e. has multi degrees of freedom arm(s)).
  • Human factors are evaluated during physical interaction.
While our primary objective is to focus on pHRI in domestic scenarios, we did not discard studies in industrial settings, as the field of pHRI is still young, and studying empirical data from various contexts can provide valuable insights.
Although some papers met the inclusion criteria, they did not directly contribute to the main objectives of this study. As a result, the following exclusion criteria were established:
  • Physical coupling with robots for intents other than collaboration [38,39], including rehabilitation.
  • Human factor evaluations of individuals with mental disorders were excluded [40], as experimental verification is required to generalize the factors of these special populations to the neurotypical population.
  • Unclear identification of physical coupling between agents within a study.
  • Manifestly missing methodology of human factors assessment or analysis [41,42].
  • Physical interaction with baby-like robots, since this type of robot is not expected to contribute to the main objective of this study [43,44].
  • Studies that are not in English [45].
  • Inaccessible full text.

5. Results

The following subsections present the results of our analysis, which aim to address the research questions we formulated. We first present a mapping scheme for pHRI in Section 5.1, followed by the classification of human factors based on our conceptual model in Section 5.2. Then, in Section 5.3, we identify the most frequently addressed human factors, and examine both the commonly employed quantification approaches and the proposed quantification approaches.

5.1. Mapping Scheme for pHRI

Due to the vast applications of direct and indirect pHRI categories, it is crucial to provide subcategories that facilitate a more detailed understanding and easier navigation of the field. To address this need, we developed a mapping scheme for pHRI that categorizes the gathered studies based on the common contexts or scenarios identified in direct and indirect pHRI, as shown in Table 1, where each subcategory is described. These subcategories are based on the specific characteristics of the studies in each category, thereby addressing RQ1.

5.1.1. Direct pHRI

Figure 1 shows that direct pHRI has received considerably more attention in human factor studies than indirect pHRI, with over two-thirds of the studies conducted in this area.
The subcategories for direct pHRI proposed in this study are based on the interaction context, including the purpose and duration of contact. These classifications provide a preliminary overview of the research in this area. As the field matures, these subcategories can be further refined to provide a more detailed understanding.
Figure 2 shows that inconsistent contact has been explored in more studies than consistent contact, while Figure 3 indicates that the topic of non-functional contact has been studied almost twice as often as the topic of functional contact.
Many of the studies examined in the context of inconsistent contact were also included in the context of non-functional contact, such as studies that have investigated the NAO robot’s reactive behavior in response to social touch, including one conducted by [46]. Another study, by [49], compared the instrumental and affective touch of the NAO robot to that of a Stretch robot in a caregiving scenario. Moreover, several studies have focused on the use of NAO in caregiving scenarios that involve inconsistent and non-functional touch, such as those by [54,73,75]. These findings are consistent with several reviews in the HRI field that have highlighted the potential of NAO in caregiving scenarios. Fitter and Kuchenbecker have conducted research on different clapping contexts using a Baxter robot. Their studies include teaching [70], and playing with the robot [25,63]. Meanwhile, Shiomi et al. used a female-looking Android to investigate the effects of its subtle reactions to touch [62,74] and the impact of pre-touch reaction distance on humans [66]. Additionally, Hu et al. investigated the impact of unexpected physical actions applied by a Sawyer robot in a gaming context with the aim of helping participants accomplish higher scores [17,36].
In contrast, many studies in the context of consistent contact were also included in the context of the functional touch. Some of those studies are conducted in dancing context [96]; direct manipulation of the arm to follow certain trajectory [90,91,93,103,105], different from kinesthetic teaching; leading robots with mobile platforms using their manipulators in a nursing context [108] and a social context [107].
Fewer studies were identified as common between the consistent contact and non-functional touch categories. These studies primarily focused on hugging and handshake scenarios with humanoids. Some studies on handshake scenarios were conducted with the Meka robot [79] and HRP-2 robot [52]. Studies conducted on hugging scenarios investigated the impact of manipulating visual and tactile appearance using an ARMAR-IIIb robot [77], the duration and style of hugs using a CASTOR robot [71], and the impact of the perceived gender using a Metahug system [57,74].
The least common studies were found between the inconsistent contact and functional touch categories. Many of these studies were conducted in a nursing scenario, where the impact of instrumental and affective touch was investigated using the Cody robot [83] and NAO robot [49]. Additionally, the NAO robot was used to investigate the impact of touch style in a nursing scenario [101], and gaze height and speech timing before and after touches were examined. However, few studies have been conducted in an industrial setting, with the majority focusing on the impact of tapping a robotic manipulator as an indicator to execute a certain command [94,95].
It is worth noting that among the studies that examined the impact of touch on human factors, only two studies ([49,83]) explored the effects of functional and non-functional touch in HRI, both of which were conducted in a nursing context and categorized under inconsistent contact. No studies have examined the effects of consistent versus inconsistent contact on human factors in direct pHRI.
Table 2 shows the common studies between all subcategories of direct pHRI, including functional touch, non-functional touch, consistent contact, and inconsistent contact.

5.1.2. Indirect pHRI

In contrast, to direct pHRI, indirect pHRI has clear classifications based on real-world applications like assembly, handover, co-manipulation, and atypical scenarios. These subcategories can be useful for researchers seeking to understand the potential use cases and challenges of HRI in different practice settings. As depicted in Figure 4, there has been a particular focus on addressing human factors in handovers, which has received the most attention compared to other forms of indirect pHRI applications.
Handover scenarios in industrial settings have been the subject of several studies. For instance, some investigations have used Universal Robot (UR) manipulators to compare comfortable handovers to those that encourage dynamic behaviors [22], while others have employed UR to explore scenarios where a manipulator is completing a preliminary task and an operator requests the handover of a tool through natural communication, such as voice commands [117]. Similarly, some studies have examined industrial contexts that incorporate handovers using a Panda manipulator [118]. Other research has focused on determining the most ergonomic tool orientation for handovers [115,121]. Various handover parameters have also been explored, including physical signs of apprehension [119], the initial position of the robot arm before handover, the robot’s grasp method, and the retraction speed of the arm following handover [116].
In the context of domestic scenarios, there has been limited research into the impact of robot parameters on human factors during handovers. One study examined comfortable robot-to-human handovers in a caregiving context for older adults, using a Pepper robot [122]. Another investigation explored affective and efficient interaction styles in a cooking context, where both faulty and successful handovers were performed during the cooking task [35].
In a general context where handovers may occur in domestic and industrial environments, a study investigated handovers using a mobile manipulator that transferred objects to humans without stopping [21].
In the atypical category, numerous studies have focused on industrial applications of assistive holding, drilling, and cutting, using manipulators [129,130,132,134,135]. A few studies have explored the potential of robots as dressing assistants [133]. Another study investigated the repositioning of patients using robots in a caregiving context [131].
Co-manipulation occurs in both domestic and industrial scenarios and is usually investigated in similar ways. It typically involves investigating ways to reduce joint overload when humans collaborate with robots in lifting heavy objects [104,112,125,127,128]. Some studies also focus on mutual adaptation in this collaboration [124], as well as human intention and behavior [126].
Studies in the assembly category are typically conducted in an industrial setting due to the nature of the application. Some of these studies have explored the use of two robotic arms to hold an object for an operator while they work on it [114]. Other studies have focused on typical assembly tasks where a robot holds an object and the user assembles another part into it [112,113], or welds a wire into it [111].
Only one study in the indirect pHRI category, conducted in an industrial setting, investigated both assembly and co-manipulation tasks [112]. This finding suggests that more research is needed in this area to gain a better understanding of the challenges and opportunities of the indirect pHRI field, since many industrial applications may require more than one type of indirect pHRI.
Among the studies reviewed, only one ([64]) investigated both direct and indirect forms of pHRI. This study incorporated social touch and handover, respectively, into a collaborative task of building a towering toy. This research shows promising directions for addressing contexts where social behavior and physical assistance are combined.

5.2. Human Factors Classification

To address RQ2, we adopted the holistic conceptual model of human factors in HRI proposed by Coronado et al. [11] and adapted it to demonstrate the relationships between the most relevant measurable dimensions found in the pHRI literature. Figure 5 illustrates the relationships between usability, user experience (UX), hedonomics, and ergonomics, based on Coronado et al. definitions, which are discussed in Section 3.
Although the conceptual model used for human factor types does not fully resolve the issue of inconsistent terminologies or definitions in the field, grouping human factors/measurable dimensions, such as trust, robot perception, and anxiety, into subcategories/human factor types, such as belief, can help address this issue. This is because presenting the trends of a category instead of individual human factors can reduce bias that arises from the usage of synonyms or variations in the ways researchers refer to human factors.
On a lower level, to address the lack of unified terminologies in the literature, this study adopts a strategy of lumping some human factors under one measurable dimension, as described in Table 3. To avoid ambiguity for the reader, a brief description of some dimensions is provided. The adopted HRI conceptual model allows for some overlap between dimensions within a human factor type, as it is difficult to define sharp outlines for each dimension. The number of times a dimension is used, whether it is lumped under a broader one, detected individually, or intersects with another in definition, will not affect the overall result of the addressed frequency of a human factor type. However, if two or more human factor types share a dimension, their overall results will be biased. Therefore, in this review, all dimensions considered for each human factor type do not intersect over more than one type. If a dimension intersects over multiple types, it is reviewed and divided into more specific dimensions to avoid bias.
In order to identify the most commonly addressed human factors, i.e., address RQ2, we present two sets of figures: Figure 6 and Figure 7 show the emerged human factor types and measurable dimensions across all gathered studies, with Table 4 showing the characteristics and the studies detected in each human factor type. It is notable how belief and the measurable dimension, robot perception, have drawn the most attention in the pHRI field. This is reflected in Figure 8 to a certain extent, which shows the number of studies that examined human factor types in different direct pHRI categories. Analogous to the considerable overlap between the non-functional touch and inconsistent contact categories, Figure 8 shows that both categories addressed hedonomics more than ergonomics. Similarly, the functional touch and consistent contact categories exhibited a similar trend, with both categories emphasizing belief more than other human factor types. However, Figure 9, which illustrates the number of studies that examined human factor types within subcategories of indirect pHRI, shows that physical ergonomics is the leading human factor type in these scenarios. Similarly, physical ergonomics is the only aspect that receives attention in the atypical category. Likewise, cognitive ergonomics and affect were not addressed in the co-manipulation category. Table 5 shows the studies detected in each measurable dimension and pHRI type.
During the analysis of the results, we observed that several studies evaluated two human factor types simultaneously within the same experiment. Notably, the highest number of studies focused on the evaluation of belief and affect, including [36,46,50,53,65,70,83,90,91,114,135]. Following this, several studies explored the relationship between cognitive ergonomics and belief, such as [17,25,35,36,57,70,90,114,118]. Fewer number of studies examined cognitive ergonomics with physical ergonomics, including [20,22,90,114] and physical ergonomics with belief, such as [55,74,90,114]. Lastly, it was found that the simultaneous consideration of affect and cognitive ergonomics within the same experiment was relatively rare. Only a few studies, such as [36,70], specifically explored this combination. Notably, there were no studies found that examined measurable dimensions from physical ergonomics and affect in the same experiment.
Additionally, some studies, such as Amanhoud et al. [114] and Wang et al. [90], assessed three human factor types at once, namely cognitive ergonomics, physical ergonomics, and belief. Similarly, Fitter et al. [70] and Hu et al. [36] investigated the effects of cognitive ergonomics, belief, and affect.

5.3. Quantification Approaches of Human Factors

Less than half of the studies included in this analysis correlated dimensions with measurable entities or proposed quantification approaches for different human factors, as revealed by the human factor usage classification, as depicted in Figure 10. Some studies validated their proposed quantification approaches by manipulating robot parameters and observing their impact on the human state, including studies by [36,46,53,92,104,106,109,116,118,119,122,129,130,134]. As a result, these studies were also included in the impact category, which contained more studies than the quantification category. Table 5 includes the studies that emerged in each human factor usage category, organized according to human factor types and measurable dimensions.
Confirming the hypothesis for RQ3, questionnaires were found to be the most commonly used approach for human factors quantification in pHRI, as shown in Figure 11. However, the remaining approaches were used with much lower frequencies compared to questionnaires. To determine the role of questionnaires in pHRI, either proposed as a quantification approach or employed as an assessment approach, further analysis was conducted. Despite the fact that questionnaires were heavily employed for assessment, Figure 12 indicates that only three studies proposed or validated standardized questionnaires for pHRI. On the other hand, the proposed quantification approaches heavily relied on physical data (e.g., force, pressure, body posture, gaze direction, etc.), either through direct correlation or mathematical and machine learning models. Table 6 shows the studies that employed quantification approaches for each measurable dimension and highlighted the studies that proposed quantification approaches. For a comprehensive understanding and easier navigation of the quantification approaches in the field, we organized the studies that utilized mathematical models and machine learning models in Table 5 and Table 6 based on the input requirements of each model, including task parameters, physical data, and physiological signals.
Consistent with the overall trend, the questionnaire approach was found to be the leading method for evaluating human factors in all human factor types, as depicted in Figure 13. Although belief was heavily assessed using questionnaires and physiological signals, no studies developed machine learning models to quantify any of its measurable dimensions. Cognitive ergonomics did not see any studies that addressed physical data, while in affect there was a moderate reliance on physiological signals.

6. Discussion

In light of the literature gaps identified in the preceding section, this section proposes suggestions on how to address these gaps and highlights the importance of addressing them in order to enhance the development of adaptive controllers for multiple service robots.

6.1. Mapping Scheme for pHRI

The proposed mapping for pHRI was essential to draw conclusions about human factors considerations in the literature. However, one cannot disregard the fact that the papers gathered by this review are limited to the ones that have addressed human factors, meaning that mapping for pHRI as a domain would require a broader range of studies. Thus, this mapping can be considered a starting point that is extensible to include more pHRI subcategories. For instance, as the field develops and more experiments are conducted, it is foreseen that the consistent subcategory in direct pHRI could be further broken down to have passive and active interaction, where robots may comply with human motion or not by applying forces against them (e.g. imposing the robot’s task as prioritized with respect to the human intention/action). It is expected that such a category would impact robot perception, and usability [36].

6.1.1. Direct pHRI

The difference in the number of studies between direct and indirect pHRI domains reflects the research community’s greater focus on evaluating the human state during direct interaction conditions compared to indirect interaction conditions. This emphasis on direct pHRI may be attributed to the fact that many of the applications that have emerged are socially oriented, as evidenced by the greater number of studies on non-functional touch compared to functional touch. For example, handshaking, [52,58,67,68,72,79,80,82,85]; hugging [51,59,71,74,77]; clapping [25,63,70]; nursing [49,83,100,101,108]; and playing virtual games [17,36,105,106] have received significant attention. This trend is consistent with the idea that touch can serve important social and emotional functions in HRI, which in turn contributes to the realization of the Society 5.0 vision to a certain extent. However, it is important to note that the Society 5.0 vision encompasses assistive applications as well. Similar to the ongoing efforts to realize Society 5.0, the emergence of Industry 5.0, which emphasizes collaboration between humans and robots instead of full automation, is expected to drive the attention given to human factors in industrial settings. This includes the adoption of direct pHRI scenarios in industrial environments.
The identification of common studies between consistent contact and non-functional touch, such as handshaking and hugging, emphasizes the importance of investigating social behaviors in a realistic and broader context that includes more types of interactions. Conversely, the considerable overlap of studies between the categories of inconsistent contact and non-functional touch, especially in nursing scenarios, suggests a maturation of the field of pHRI in which researchers design experiments that encompass a wider range of interaction types beyond touch. Additionally, the studies that have investigated the purpose of direct pHRI (i.e., functional touch vs. non-functional touch) have mainly focused on nursing scenarios, indicating the relatively greater attention given to this area. However, the common studies detected between inconsistent contact and non-functional touch, which mainly comprise studies addressing social touches, may be due to the fact that touches occur spontaneously in human interactions and do not necessarily trigger or stop them if not maintained, unlike functional touches where many studies were exploratory and used the robot as a tool to follow certain paths.
The significant overlap between consistent contact and functional touch, and between inconsistent contact and non-functional touch suggest that they may be redundant, indicating a need to consider either the purpose or duration categories for future study classification in reviews. Further refinement of these classifications could enhance our understanding of direct pHRI as the field matures.
Our main objective is to bring robotic multiple service providers to life, and it is promising that most direct studies of pHRI focus on non-industrial applications. However, this also highlights a gap in current research for Industry 5.0. While non-industrial applications have explored social aspects of pHRI, such as handshaking, hugging, and playing, they have not adequately investigated physical assistance applications in daily life activities such as toileting, sit-and-stand, transferring, and others. Without proper physical assistance, the potential benefits of social aspects are limited, and pHRI may not be useful in domestic or institutional environments. Therefore, it is important to investigate more physical assistive applications, while taking into consideration the human state, to fully realize the potential of robotic multiple service providers.

6.1.2. Indirect pHRI

Although human factors in handover scenarios have been explored in various contexts, there has been insufficient attention given to handovers in domestic settings. Handovers in the industrial context are being thoroughly investigated, including the orientation of tools and the human body posture. However, the objective of comfortable and efficient handovers is likely similar in both industrial and domestic settings. This suggests that results from studies conducted in one setting could be applicable to the other.
However, most studies on atypical scenarios and assembly tasks are industrial-focused, highlighting the need to address domestic applications beyond those currently studied. Atypical scenarios in domestic settings, such as dressing, grooming, showering, or household chores, are complex and technically challenging. Similarly, assembly contexts in domestic settings, such as putting pillows into pillowcases or performing certain cleaning duties, could also pose technical challenges.
Studies on co-manipulation scenarios are limited, and the focus is mainly on reducing human joint overload during collaboration. Few studies have investigated mutual adaptation to achieve a common goal. While addressing joint overload is a logical first step in co-manipulation, mutual adaptation between humans and robots is also crucial in domestic scenarios where robots can assist with tasks such as furniture reorganization.
The scarcity of research in the indirect pHRI category underscores the insufficient attention that human factors researchers have paid to this field. Additionally, The lack of studies that addressed different applications simultaneously suggests that more research is needed in this area to gain a better understanding of the challenges and opportunities of indirect pHRI, since many industrial settings may require multiple types of applications. Notably, domestic environments have received minimal to no attention in any of the subcategories of indirect pHRI. It could be argued that the term pHRI is more commonly used in industrial settings than in domestic contexts in the indirect pHRI domain.

6.2. Human Factor Classification

The holistic human factors model adapted to pHRI and proposed in Figure 5 is focused on our specific research questions and therefore differs from the original in some respects, i.e., in [11]. In particular, we did not include measurable dimensions related to "performance" as a human factor type under "ergonomics" since it was not a primary focus of our analysis. Based on user experience definition, we found it appropriate to categorize all measurable dimensions related to physical and cognitive ergonomics under the "user experience" umbrella, whereas the original model depicts "user experience" overlapping with these human factor types but not entirely encompassing them. It is worth noting that for a more comprehensive understanding of the conceptual model, we recommend readers to refer to the original adapted HRI model.
Given the objectives of our review, we focused on factors that are affected by the interaction. Therefore, our review did not detect factors that appear out of the scope of the definition of user experience, such as workplace design and awareness, which are not impacted during an interaction [11].
We ensured that each measurable dimension we identified was classified under only one human factor type, as overlapping dimensions between types would have contradicted the holistic approach of the conceptual model we adopted. To achieve this, we based our classification on the definitions of usability, user experience, hedonomics, ergonomics, and their respective sub-categories.
Similarly, based on the definition of usability and the main objectives of this review, all of the measurable dimensions detected in usability belong to the hedonomics category of human factors. This is because the objective aspects of usability, which are more related to performance, were not a primary focus of the review.
Although defining each measurable dimension in detail was not our primary focus, we classified them based on the provided descriptions of human factor types and the usability definition. This approach allowed us to present an accurate and unbiased overview of the most relevant human factors that have been addressed in pHRI literature.
Our primary goal is to expedite the development of adaptive control frameworks for multiple service robots. We believe that investigating the most frequent human factor types evaluated together in a single experiment, along with exploring the relationships between measurable dimensions within each human factor type, will aid in this development by identifying the human factors that may have significant correlations for researchers to consider when designing adaptive control frameworks.
Since affect and belief address humans’ emotions from different perspectives, it is not surprising that most of the studies that have evaluated more than 1 human factor type belonged to belief and affect categories. Similarly, cognitive ergonomics and belief come next in the most common number of studies between human factor types, since both evaluate cognitive aspects but from different perspectives. In contrast, no studies have evaluated humans’ emotions during an assessment of their physical ergonomics. Therefore, it is anticipated that there could be an unrevealed relationship between the 3 human factor types and their measurable dimensions that is worth investigating. On a broader view, unrevealed relationships between measurable dimensions within a human factor type are very likely to exist. However, a lack of definitions may not support conducting such a study, unlike correlating human factor types to each other, which can be perceived as more feasible.
No wonder direct pHRI studies addressed a wider variety of human factor types in comparison with indirect pHRI studies since many more papers have emerged in the former than the latter. As expected, affect is the most addressed human factor type among the non-functional touch studies, and accordingly, the inconsistent studies, given the big number of common studies among both categories. Along the same lines, physical ergonomics is the least addressed human factor type among them since those type of studies is more socially directed. It was foreseeable that belief would be highly addressed among most of the studies since HRI literature has pointed out that dimensions such as trust were the most frequently addressed in comparison with any other dimensions. In contrast, cognitive ergonomics neither attracted most of the research community nor was completely ignored by it.
As previously mentioned, the majority of indirect pHRI applications are focused on industry. Consequently, it is not surprising that physical ergonomics is the most frequently addressed category, with limited attention given to affect. In fact, affect was hardly mentioned in any of the indirect categories. Belief, however, which takes emotions into account, received more attention than affect due to its evaluation being influenced by cognitive aspects.
To fully realize the potential of pHRI in domestic scenarios, it is essential to address various gaps in the literature. In direct pHRI scenarios, the robotics research community must prioritize assessing physical assistance experiments, such as toileting, and sit-and-stand, from a human factors perspective. While social interaction experiments like hugging, handshaking, and playing are important, it is crucial to combine both physical interaction types to understand their influence on user perception during physical assistance. Additionally, applications that may be considered industrial or non-industrial, such as handovers and co-manipulation, should also prioritize hedonomics and cognitive ergonomics, in addition to physical ergonomics. Moreover, atypical and assembly applications such as eating assistance, grooming, showering, and household chores require attention to various human factors. Addressing these literature gaps will enhance the overall HRI experience and increase the usefulness of pHRI in domestic settings.

6.3. Human Factors Quantification

Due to the lack of consensus on human factors terminology and definitions in the domain of HRI, the large number of experiments conducted to understand the impact of human factors, as shown in Figure 10, is not expected to yield efficient and uniform progress in the field. In other words, without reliable and agreed-upon quantification approaches, the results of experiments on the impact of human factors may not be reliable enough to facilitate the development of the field. However, A promising aspect is that a significant number of studies proposed quantification approaches and validated them, demonstrating the research community’s awareness of the importance of such validation. The development of standardized and reliable quantification methods has the potential to facilitate the detection of progress in the field of pHRI, thus accelerating its advancement.
It is clear that the human factor types addressed in Figure 6 follow a trend similar to the measurable dimensions addressed in Figure 7. However, the human factor types are comprehensive, reflect the research direction in a broader view and give all measurable dimensions, within a human factor type, the same level of importance despite the terminology used. Thus, adopting the conceptual model is expected to yield efficient uniform progress in the field.
Figure 11 and Figure 12 demonstrate that subjective approaches are the primary evaluation method used across all human factor types. Questionnaires are commonly used due to their simplicity and fast administration, whereas other assessment approaches may require specialized equipment and complex data processing. However, questionnaires should be considered a preliminary assessment approach that provides an initial understanding of human states or validates other objective approaches. Otherwise, the development of adaptive control frameworks for human states is unlikely to become a reality in the near future. Hedonomics relies primarily on questionnaires, whereas ergonomics has investigated objective approaches more frequently than subjective ones. Physical ergonomics parameters are quantifiable, which explains why objective measures have been used in physical ergonomics studies more than in cognitive ergonomics. Future efforts should focus on developing objective measures for the other three human factor types to enable the implementation of real-time adaptive control frameworks in the future. Nevertheless, developing real-time objective quantification approaches for HRI, including its subcategory pHRI, is challenging, which explains why these approaches are rarely addressed in these fields.

7. Limitations of the Review

This section is dedicated to demonstrating the validity of this review by discussing its limitations. It should be noted that limitations and future work of the field are discussed within Section 6, Discussion. In order to evaluate the validity of this review, 3 validity factors proposed by [30] can be applied: theoretical validity, interpretive validity, and descriptive validity.
A common limitation in literature reviews is the possibility of existing studies that are related to the main topic but could not be detected due to unfortunate choices of keywords either from the review’s authors or the related studies. However, the adopted guidelines, which included the test-set step, could make us argue that this type of threat is eliminated to a decent extent. Petersen et al. [30] have referred to this validity as theoretical validity which they defined as “our ability to be able to capture what we intend to capture”. Therefore, theoretical validity includes data extraction validation. In order to restrain data extraction bias, full-text checkout over all the gathered papers was performed iterative, until a relatively proper lumping of measurable dimensions is achieved.
Another validity factor is interpretive validity, which “is achieved when the conclusions drawn are reasonable given the data" [30]. To limit interpretive validity threats, HRI experts got involved during each step of conducting this review, including the interpretation of the extracted data. However, one can argue that broad dimensions, such as robot perception, emotions, and physical comfort are misinterpreted as they can be divided into many smaller dimensions. For instance, robot perception can be divided into attitudes, acceptance, perceived performance, etc. However, according to the HRI conceptual model adopted, measurable dimensions that do not intersect between human factor types cannot bias our results as previously explained in Section 5.2. Nevertheless, breaking down factors can provide a clearer understanding of the measurable dimensions addressed within a human factors type. Therefore, it is highly encouraged for future reviewers to consider this approach. Furthermore, it can be argued that one common human factor, such as perceived safety, did not emerge as a measurable dimension within any human factor type, which might be regarded as another misinterpretation. However, perceived safety is a broad term that is shared between different human factor types as it includes dimensions such as stress, anxiety, trust, fear, and psychological comfort[7]. Therefore, considering it as a measurable dimension is not reasonable.
To maintain descriptive validity, which “is the extent to which observations are described accurately and objectively”, the adapted model of human factors in HRI is considered. The conceptual model provides clear outlines for usability, user experience, and each of the human factors types, while still allowing for accurate classification of measurable dimensions within each category. However, it can be noticed that performance as a human factor type within ergonomics is discarded, which can rise a descriptive validity threat by arguing that ergonomics is not accurately conducted throughout this review. As discussed in Section 5.2, performance, as defined by Coronado et al. [11], includes the evaluation of the performance of the robot, the user, and the overall system. However, since our main objective is to quantify the human state, these performance evaluations are not directly relevant to our review. Especially that perceived performance, which contributes to the human state, is included in robot perception, as shown in Table 3. As this is a human-centered review, research has shown that in domestic contexts, humans tend to prefer robots that exhibit human-like characteristics over high-performance robots [35]. Therefore, since our primary objective is to quantify the human state, discarding performance evaluations do not negatively impact the main focus of this review. Therefore, we are confident that the conclusions of this review can serve as a reliable foundation for future research in this field.

8. Conclusions

For multiple service robots to coexist with humans, they must meet users’ physical and social needs. Socially, robots must be able to predict and quantify human states to interact appropriately. Physically, manipulation is a key capability. This article focused on studies of robots with manipulation abilities and their impact on human factors in pHRI. The goal is to advance the development of pHRI control frameworks that can predict human states and make decisions accordingly.
To achieve the objective of this review, we identified the most common human factors in pHRI, noted the factors typically addressed together, and determined the most commonly used assessment approaches. We also collected and classified proposed assessment approaches for each human factor and created an initial map of the common contexts and applications in pHRI. Our aim was to provide an unbiased overview of the field, identify research gaps, and facilitate the search for adaptive control framework advancements.
Most studies in the direct pHRI category focus on social behaviors, with belief being the most commonly addressed human factor type. Task collaboration is moderately investigated, while physical assistance is rarely studied. In contrast, indirect pHRI studies often involve industrial settings, with physical ergonomics being the most frequently investigated human factor. More research is needed on human factors in direct and indirect physical assistance applications, including studies that combine physical social behaviours (e.g., comforting touch) with physical assistance tasks (e.g., sit-to-stand). This will enable robots to exhibit appropriate social skills while providing physical assistance. Expectedly, the predominant approach in most studies involves the use of questionnaires as the main method of quantification. However, it is worth noting a recent trend that seeks to address quantification approaches based on measurable data.
To advance service robots and Society 5.0, future reviews should include a wider range of HRI scenarios beyond physical collaboration. These reviews should prioritize studies that use measurable data to quantify human factors, as well as those that explore correlations between demographic and personal information and human factors. By identifying relationships between measurable dimensions of each human factor, we can create more robust adaptive control frameworks for various pHRI scenarios.

Author Contributions

Conceptualization, Nourhan Abdulazeem and Yue Hu; methodology, Nourhan Abdulazeem; formal analysis, Nourhan Abdulazeem; resources, Nourhan Abdulazeem; data curation, Nourhan Abdulazeem; writing—original draft preparation, Nourhan Abdulazeem; writing—review and editing, Yue Hu; supervision, Yue Hu; funding acquisition, Yue Hu. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC). We acknowledge the support of the Natural Sciences and Engineering Research Council of Canada (NSERC), [funding reference number RGPIN-2022-03857].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ageing and health. Available online: https://www.who.int/news-room/fact-sheets/detail/ageing-and-health. (accessed on 20 October 2022).
  2. Ageing. Available online: https://www.un.org/en/global-issues/ageing. (accessed on 17 May 2023).
  3. Building The Caregiving Workforce Our Aging World Needs. Glob. Coalit. Aging 2021.
  4. Society 5.0. Available online: https://www8.cao.go.jp/cstp/english/society5_0/index.html. (accessed on 29 November 2022).
  5. Schweighart, R.; O’Sullivan, J.L.; Klemmt, M.; Teti, A.; Neuderth, S. Wishes and Needs of Nursing Home Residents. Healthcare 2022, 10, 854. [Google Scholar] [CrossRef] [PubMed]
  6. What is Human Factors and Ergonomics | HFES. Available online: https://www.hfes.org/About-HFES/What-is-Human-Factors-and-Ergonomics. (accessed on 30 November 2022).
  7. Rubagotti, M.; Tusseyeva, I.; Baltabayeva, S.; Summers, D.; Sandygulova, A. Perceived safety in physical human robot interaction—A survey. Rob. Auton. Syst. 2022, 104047. [Google Scholar] [CrossRef]
  8. Pervez, A.; Ryu, J. Safe physical human robot interaction-past, present and future. J. Mech. Sci. 2008, 22, 469–483. [Google Scholar] [CrossRef]
  9. Selvaggio, M.; Cognetti, M.; Nikolaidis, S.; Ivaldi, S.; Siciliano, B. Autonomy in physical human-robot interaction: A brief survey. IEEE Robot. Autom. Lett. 2021, 6, 7989–7996. [Google Scholar] [CrossRef]
  10. Losey, D.P.; McDonald, C.G.; Battaglia, E.; O’Malley, M.K. A Review of Intent Detection, Arbitration, and Communication Aspects of Shared Control for Physical Human–Robot Interaction. Appl. Mech. Rev. 2018, 70, 010804. [Google Scholar] [CrossRef]
  11. Coronado, E.; Kiyokawa, T.; Ricardez, G.A.G.; Ramirez-Alpizar, I.G.; Venture, G.; Yamanobe, N. Evaluating quality in human-robot interaction: A systematic search and classification of performance and human-centered factors, measures and metrics towards an industry 5.0. J. Manuf. Syst. 2022, 63, 392–410. [Google Scholar] [CrossRef]
  12. Hopko, S.; Wang, J.; Mehta, R. Human Factors Considerations and Metrics in Shared Space Human-Robot Collaboration: A Systematic Review. Front. Robot. AI. 2022, 9. [Google Scholar] [CrossRef]
  13. Ortenzi, V.; Cosgun, A.; Pardi, T.; Chan, W.P.; Croft, E.; Kulić, D. Object Handovers: A Review for Robotics. IEEE Trans. Robot. 2021, 37, 1855–1873. [Google Scholar] [CrossRef]
  14. Prasad, V.; Stock-Homburg, R.; Peters, J. Human-Robot Handshaking: A Review. Int. J. Soc. Robot. 2022, 14, 277–293. [Google Scholar] [CrossRef]
  15. Simone, V.D.; Pasquale, V.D.; Giubileo, V.; Miranda, S. Human-Robot Collaboration: an analysis of worker’s performance. Procedia Comput. Sci. 2022, 200, 1540–1549. [Google Scholar] [CrossRef]
  16. Lorenzini, M.; Lagomarsino, M.; Fortini, L.; Gholami, S.; Ajoudani, A. Ergonomic Human-Robot Collaboration in Industry: A Review. Front. Robot. AI. 2023, 9, 813907. [Google Scholar] [CrossRef] [PubMed]
  17. Hu, Y.; Benallegue, M.; Venture, G.; Yoshida, E. Interact With Me: An Exploratory Study on Interaction Factors for Active Physical Human-Robot Interaction. IEEE Robot. Autom. Lett. 2020, 5, 6764–6771. [Google Scholar] [CrossRef]
  18. Fujita, M.; Kato, R.; Tamio, A. Assessment of operators’ mental strain induced by hand-over motion of industrial robot manipulator. 19th International Symposium in Robot and Human Interactive Communication, 2010, pp. 361–366. [CrossRef]
  19. Gualtieri, L.; Fraboni, F.; De Marchi, M.; Rauch, E. Development and evaluation of design guidelines for cognitive ergonomics in human-robot collaborative assembly systems. Appl. Ergon. 2022, 104, 103807. [Google Scholar] [CrossRef] [PubMed]
  20. Dehais, F.; Sisbot, E.A.; Alami, R.; Causse, M. Physiological and subjective evaluation of a human–robot object hand-over task. Appl. Ergon. 2011, 42, 785–791. [Google Scholar] [CrossRef] [PubMed]
  21. He, K.; Simini, P.; Chan, W.P.; Kulić, D.; Croft, E.; Cosgun, A. On-The-Go Robot-to-Human Handovers with a Mobile Manipulator. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), 2022, pp. 729–734. [CrossRef]
  22. Zolotas, M.; Luo, R.; Bazzi, S.; Saha, D.; Mabulu, K.; Kloeckl, K.; Padir, T. Productive Inconvenience: Facilitating Posture Variability by Stimulating Robot-to-Human Handovers. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 122–128. [Google Scholar] [CrossRef]
  23. Peternel, L.; Fang, C.; Tsagarakis, N.; Ajoudani, A. Online Human Muscle Force Estimation for Fatigue Management in Human-Robot Co-Manipulation. 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2018, pp. 1340–1346. [CrossRef]
  24. Mizanoor Rahman, S.M.; Ikeura, R. Cognition-based variable admittance control for active compliance in flexible manipulation of heavy objects with a power-assist robotic system. Robotics biomim. 2018, 5, 7. [Google Scholar] [CrossRef] [PubMed]
  25. Fitter, N.; Kuchenbecker, K. Synchronicity Trumps Mischief in Rhythmic Human-Robot Social-Physical Interaction. Springer Proceedings in Adv. Robot. 2020, 10, 269–284. [Google Scholar] [CrossRef]
  26. Langer, A.; Feingold-Polak, R.; Mueller, O.; Kellmeyer, P.; Levy-Tzedek, S. Trust in socially assistive robots: Considerations for use in rehabilitation. Neurosci. Biobehav. Rev. 2019, 104, 231–239. [Google Scholar] [CrossRef]
  27. Chen, T.L.; King, C.H.; Thomaz, A.L.; Kemp, C.C. Touched by a robot: An investigation of subjective responses to robot-initiated touch. 2011 6th ACM/IEEE International Conference on Human-Robot Interaction (HRI), 2011, pp. 457–464. [CrossRef]
  28. Sauer, J.; Sonderegger, A.; Schmutz, S. Usability, user experience and accessibility: towards an integrative model. Ergonomics 2020, 63, 1207–1220. [Google Scholar] [CrossRef]
  29. ISO 9241-11:2018(en), Ergonomics of human-system interaction — Part 11: Usability: Definitions and concepts. https://www.iso.org/obp/ui/#iso:std:iso:9241:-11:ed-2:v1:en.
  30. Petersen, K.; Vakkalanka, S.; Kuzniarz, L. Guidelines for conducting systematic mapping studies in software engineering: An upyear. Inf. Softw. Technol. -01. [CrossRef]
  31. Kitchenham, B.; Brereton, P. A systematic review of systematic review process research in software engineering. Inf. Softw. Technol. 2013, 55, 2049–2075. [Google Scholar] [CrossRef]
  32. Wohlin, C. Guidelines for snowballing in systematic literature studies and a replication in software engineering. the 18th International Conference; ACM Press: London, England, United Kingdom, 2014; pp. 1–10. [Google Scholar] [CrossRef]
  33. Sigurjónsson, V.; Johansen, K.; Rösiö, C. Exploring the operator’s perspective within changeable and automated manufacturing – A literature review. Procedia CIRP 2022, 107, 369–374. [Google Scholar] [CrossRef]
  34. Simões, A.C.; Pinto, A.; Santos, J.; Pinheiro, S.; Romero, D. Designing human-robot collaboration (HRC) workspaces in industrial settings: A systematic literature review. J. Manuf. Syst. 2022, 62, 28–43. [Google Scholar] [CrossRef]
  35. Hamacher, A.; Bianchi-Berthouze, N.; Pipe, A.G.; Eder, K. Believing in BERT: Using expressive communication to enhance trust and counteract operational error in physical Human-robot interaction. 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), 2016, pp. 493–500. [CrossRef]
  36. Hu, Y.; Abe, N.; Benallegue, M.; Yamanobe, N.; Venture, G.; Yoshida, E. Toward Active Physical Human–Robot Interaction: Quantifying the Human State During Interactions. IEEE Trans. Hum. Mach. Syst. 2022. [Google Scholar] [CrossRef]
  37. Naneva, S.; Sarda Gou, M.; Webb, T.L.; Prescott, T.J. A Systematic Review of Attitudes, Anxiety, Acceptance, and Trust Towards Social Robots. Int J of Soc Robotics 2020, 12, 1179–1201. [Google Scholar] [CrossRef]
  38. Fraboni, F.; Gualtieri, L.; Millo, F.; De Marchi, M.; Pietrantoni, L.; Rauch, E. Human-Robot Collaboration During Assembly Tasks: The Cognitive Effects of Collaborative Assembly Workstation Features. Proceedings of the 21st Congress of the International Ergonomics Association (IEA 2021); Springer International Publishing: Cham, Switzerland, 2022; pp. 242–249. [Google Scholar] [CrossRef]
  39. Naggita, K.; Athiley, E.; Desta, B.; Sebo, S. Parental Responses to Aggressive Child Behavior towards Robots, Smart Speakers, and Tablets. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 337–344. [Google Scholar] [CrossRef]
  40. Sandygulova, A.; Amirova, A.; Telisheva, Z.; Zhanatkyzy, A.; Rakhymbayeva, N. Individual Differences of Children with Autism in Robot-assisted Autism Therapy. 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI); IEEE: Sapporo, Japan, 2022; pp. 43–52. [Google Scholar] [CrossRef]
  41. Ficuciello, F.; Villani, L.; Siciliano, B. Variable Impedance Control of Redundant Manipulators for Intuitive Human-Robot Physical Interaction. IEEE Trans. Robot. 2015, 31, 850–863. [Google Scholar] [CrossRef]
  42. Ma, X.; Quek, F. Development of a child-oriented social robot for safe and interactive physical interaction. IEEE/RSJ 2010 International Conference on Intelligent Robots and Systems, IROS 2010 - Conference Proceedings, 2010, pp. 2163–2168. [CrossRef]
  43. Kunimura, H.; Ono, C.; Hirai, M.; Muramoto, M.; Matsuzaki, W.; Uchiyama, T.; Shiratori, K.; Hoshino, J. Baby robot "YOTARO"; Vol. 6243 LNCS, 2010. [CrossRef]
  44. Yamashita, Y.; Ishihara, H.; Ikeda, T.; Asada, M. Path Analysis for the Halo Effect of Touch Sensations of Robots on Their Personality Impressions. SOCIAL ROBOTICS, (ICSR 2016); Agah, A.; Cabibihan, J.; Howard, A.; Salichs, M.; He, H., Eds., 2016, Vol. 9979, pp. 502–512. [CrossRef]
  45. Ishii, S. Meal-assistance Robot “My Spoon”. JRSJ 2003, 21, 378–381. [Google Scholar] [CrossRef]
  46. Okuda, M.; Takahashi, Y.; Tsuichihara, S. Human Response to Humanoid Robot That Responds to Social Touch. Appl. Sci. 2022, 12. [Google Scholar] [CrossRef]
  47. Jørgensen, J.; Bojesen, K.; Jochum, E. Is a Soft Robot More “Natural”? Exploring the Perception of Soft Robotics in Human–Robot Interaction. Int. J. Soc. Robot. 2022, 14, 95–113. [Google Scholar] [CrossRef]
  48. Au, R.H.Y.; Ling, K.; Fraune, M.R.; Tsui, K.M. Robot Touch to Send Sympathy: Divergent Perspectives of Senders and Recipients. 2022 17th ACM/IEEE International Conference on Human-Robot Interaction (HRI); IEEE: Sapporo, Japan, 2022; pp. 372–382. [Google Scholar]
  49. Mazursky, A.; DeVoe, M.; Sebo, S. Physical Touch from a Robot Caregiver: Examining Factors that Shape Patient Experience. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 1578–1585. [Google Scholar] [CrossRef]
  50. Kunold, L. Seeing is not Feeling the Touch from a Robot. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 1562–1569. [Google Scholar] [CrossRef]
  51. Block, A.E.; Seifi, H.; Hilliges, O.; Gassert, R.; Kuchenbecker, K.J. In the Arms of a Robot: Designing Autonomous Hugging Robots with Intra-Hug Gestures. ACM Trans. Hum.-Robot Interact. 2022, 12, 47. [Google Scholar] [CrossRef]
  52. Bainbridge, W.; Nozawa, S.; Ueda, R.; Okada, K.; Inaba, M. A Methodological Outline and Utility Assessment of Sensor-based Biosignal Measurement in Human-Robot Interaction A System for Determining Correlations Between Robot Sensor Data and Subjective Human Data in HRI. Int. J. Soc. Robot. 2012, 4, 303–316. [Google Scholar] [CrossRef]
  53. Hayashi, A.; Rincon-Ardila, L.; Venture, G. Improving HRI with Force Sensing. Machines 2022, 10. [Google Scholar] [CrossRef]
  54. Giorgi, I.; Tirotto, F.; Hagen, O.; Aider, F.; Gianni, M.; Palomino, M.; Masala, G. Friendly But Faulty: A Pilot Study on the Perceived Trust of Older Adults in a Social Robot. IEEE Access 2022, 10, 92084–92096. [Google Scholar] [CrossRef]
  55. Burns, R.; Lee, H.; Seifi, H.; Faulkner, R.; Kuchenbecker, K. Endowing a NAO Robot With Practical Social-Touch Perception. Front. Robot. AI. -19. [CrossRef]
  56. Zheng, X.; Shiomi, M.; Minato, T.; Ishiguro, H. Modeling the Timing and Duration of Grip Behavior to Express Emotions for a Social Robot. IEEE Robot. Autom. Lett. 2021, 6, 159–166. [Google Scholar] [CrossRef]
  57. Shiomi, M.; Hagita, N. Audio-Visual Stimuli Change not Only Robot’s Hug Impressions but Also Its Stress-Buffering Effects. Int. J. Soc. Robot. 2021, 13, 469–476. [Google Scholar] [CrossRef]
  58. Law, T.; Malle, B.; Scheutz, M. A Touching Connection: How Observing Robotic Touch Can Affect Human Trust in a Robot. Int. J. Soc. Robot. 2003; 13. [Google Scholar] [CrossRef]
  59. Gaitan-Padilla, M.; Maldonado-Mejia, J.; Fonseca, L.; Pinto-Bernal, M.; Casas, D.; Munera, M.; Cifuentes, C. Physical Human-Robot Interaction Through Hugs with CASTOR Robot. SOCIAL ROBOTICS, ICSR 2021; Li, H.; Ge, S.; Wu, Y.; Wykowska, A.; He, H.; Liu, X.; Li, D.; PerezOsorio, J., Eds., 2021, Vol. 13086, pp. 814–818. [CrossRef]
  60. Zheng, X.; Shiomi, M.; Minato, T.; Ishiguro, H. What Kinds of Robot’s Touch Will Match Expressed Emotions? IEEE Robot. Autom. Lett. 2020, 5, 127–134. [Google Scholar] [CrossRef]
  61. Zamani, N.; Moolchandani, P.; Fitter, N.; Culbertson, H. ; IEEE. Effects of Motion Parameters on Acceptability of Human-Robot Patting Touch. 2020 IEEE HAPTICS SYMPOSIUM (HAPTICS), 2020, pp. 664–670. [CrossRef]
  62. Shiomi, M.; Minato, T.; Ishiguro, H. Effects of Robot’s Awareness and its Subtle Reactions Toward People’s Perceived Feelings in Touch Interaction. J. Robot. Mechatron. 2020, 32, 43–50. [Google Scholar] [CrossRef]
  63. Fitter, N.; Kuchenbecker, K. How Does It Feel to Clap Hands with a Robot? Int. J. Soc. Robot. 2020, 12, 113–127. [Google Scholar] [CrossRef]
  64. Claure, H.; Khojasteh, N.; Tennent, H.; Jung, M. Using expectancy violations theory to understand robot touch interpretation. 2020, pp. 163–165. [CrossRef]
  65. Willemse, C.; van Erp, J. Social Touch in Human-Robot Interaction: Robot-Initiated Touches can Induce Positive Responses without Extensive Prior Bonding. Int. J. Soc. Robot. 2019, 11, 285–304. [Google Scholar] [CrossRef]
  66. Shiomi, M.; Shatani, K.; Minato, T.; Ishiguro, H. How Should a Robot React Before People’s Touch?: Modeling a Pre-Touch Reaction Distance for a Robot’s Face. IEEE Robot. Autom. Lett. 2018, 3, 3773–3780. [Google Scholar] [CrossRef]
  67. Orefice, P.; Ammi, M.; Hafez, M.; Tapus, A. Pressure Variation Study in Human-Human and Human-Robot Handshakes: Impact of the Mood. 2018 27TH IEEE International Symposium on Robot and Human Interactive Communication IEEE RO-MAN 2018); Cabibihan, J.; Mastrogiovanni, F.; Pandey, A.; Rossi, S.; Staffa, M., Eds., 2018, pp. 247–254.
  68. Midorikawa, R.; Niitsuma, M. Effects of Touch Experience on Active Human Touch in Human-Robot Interaction. IFAC-PapersOnLine, 2018, Vol. 51, pp. 154–159. 2: Issue; 22. [CrossRef]
  69. Hirano, T.; Shiomi, M.; Iio, T.; Kimoto, M.; Tanev, I.; Shimohara, K.; Hagita, N. How Do Communication Cues Change Impressions of Human-Robot Touch Interaction? Int. J. Soc. Robot. 2018, 10, 21–31. [Google Scholar] [CrossRef]
  70. Fitter, N.; Kuchenbecker, K. Teaching a robot bimanual hand-clapping games via wrist-worn IMUs. Front. Robot. AI. 2018, 5. [Google Scholar] [CrossRef]
  71. Block, A.; Kuchenbecker, K. Emotionally Supporting Humans Through Robot Hugs. ACM/IEEE International Conference on Human-Robot Interaction, 2018, pp. 293–294. [CrossRef]
  72. Avelino, J.; Correia, F.; Catarino, J.; Ribeiro, P.; Moreno, P.; Bernardino, A.; Paiva, A.; Kosecka, J. The Power of a Hand-shake in Human-Robot Interactions. 2018 IEEE/RSJ International Conference on Intelligent Robots and systems (IROS); Maciejewski, A.; Okamura, A.; Bicchi, A.; Stachniss, C.; Song, D.; Lee, D.; Chaumette, F.; Ding, H.; Li, J.; Wen, J.; Roberts, J.; Masamune, K.; Chong, N.; Amato, N.; Tsagwarakis, N.; Rocco, P.; Asfour, T.; Chung, W.; Yasuyoshi, Y.; Sun, Y.; Maciekeski, T.; Althoefer, K.; AndradeCetto, J.; Chung, W.; Demircan, E.; Dias, J.; Fraisse, P.; Gross, R.; Harada, H.; Hasegawa, Y.; Hayashibe, M.; Kiguchi, K.; Kim, K.; Kroeger, T.; Li, Y.; Ma, S.; Mochiyama, H.; Monje, C.; Rekleitis, I.; Roberts, R.; Stulp, F.; Tsai, C.; Zollo, L., Eds., 2018, pp. 1864–1869.
  73. Andreasson, R.; Alenljung, B.; Billing, E.; Lowe, R. Affective Touch in Human-Robot Interaction: Conveying Emotion to the Nao Robot. Int. J. Soc. Robot. 2018, 10, 473–491. [Google Scholar] [CrossRef]
  74. Shiomi, M.; Hagita, N. Do Audio-Visual Stimuli Change Hug Impressions? SOCIAL ROBOTICS, ICSR 2017; Kheddar, A.; Yoshida, E.; Ge, S.; Suzuki, K.; Cabibihan, J.; Eyssel, F.; He, H., Eds., 2017, Vol. 10652, pp. 345–354. [CrossRef]
  75. Alenljung, B.; Andreasson, R.; Billing, E.; Lindblom, J.; Lowe, R. User Experience of Conveying Emotions by Touch. 2017 26TH IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN); Howard, A.; Suzuki, K.; Zollo, L., Eds., 2017, pp. 1240–1247.
  76. Willemse, C.; Huisman, G.; Jung, M.; van Erp, J.; Heylen, D. Observing Touch from Video: The Influence of Social Cues on Pleasantness Perceptions. Haptics : Perception, Devices, Control, and Applications, Eurohaptics 2016, PT II; Bello, F.; Kajimoto, H.; Visell, Y., Eds., 2016, Vol. 9775, pp. 196–205. [CrossRef]
  77. Trovato, G.; Do, M.; Terlemez, O.; Mandery, C.; Ishii, H.; Bianchi-Berthouze, N.; Asfour, T.; Takanishi, A. Is hugging a robot weird? Investigating the influence of robot appearance on users’ perception of hugging. IEEE-RAS International Conference on Humanoid Robots, 2016, pp. 318–323. [CrossRef]
  78. Martinez-Hernandez, U.; Rubio-Solis, A.; Prescott, T. ; IEEE. Bayesian perception of touch for control of robot emotion. 2016 International Joint Conference on Neural Networks (IJCNN), 2016, pp. 4927–4933.
  79. Tsalamlal, M.; Martin, J.; Ammi, M.; Tapus, A.; Amorim, M. ; IEEE. Affective Handshake with a Humanoid Robot: How do Participants Perceive and Combine its Facial and Haptic Expressions? 2015 International Conference on Affective Computing and Intelligent Interaction (ACII), 2015, pp. 334–340.
  80. Jindai, M.; Ota, S.; Ikemoto, Y.; Sasaki, T. Handshake request motion model with an approaching human for a handshake robot system. Proceedings of the 2015 7th IEEE International Conference on Cybernetics and Intelligent Systems, CIS 2015 and Robotics, Automation and Mechatronics, RAM 2015, 2015, pp. 265–270. [Google Scholar] [CrossRef]
  81. Haring, K.; Watanabe, K.; Silvera-Tawil, D.; Velonaki, M.; Matsumoto, Y. Touching an Android robot: Would you do it and how? Proceedings - 2015 International Conference on Control, Automation and Robotics, ICCAR 2015, 2015, pp. 8–13. [Google Scholar] [CrossRef]
  82. Ammi, M.; Demulier, V.; Caillou, S.; Gaffary, Y.; Tsalamlal, Y.; Martin, J.; Tapus, A. ; ACM. Haptic Human-Robot Affective Interaction in a Handshaking Social Protocol. Proceedings of the 2015 ACM/IEEE International Conference on Human-Robot Interaction (HRI’15), 2015, pp. 263–270. [CrossRef]
  83. Chen, T.; King, C.H.; Thomaz, A.; Kemp, C. An Investigation of Responses to Robot-Initiated Touch in a Nursing Context. Int. J. Soc. Robot. 2014, 6, 141–161. [Google Scholar] [CrossRef]
  84. Nie, J.; Park, M.; Marin, A.; Sundar, S. ; Assoc Comp Machinery. Can You Hold My Hand? Physical Warmth in Human-Robot Interaction. Proceeding of the Seventh annual ACM/IEEE International Conference on Human-Robot Interaction, 2012, pp. 201–202.
  85. Jindai, M.; Watanabe, T. Development of a handshake request motion model based on analysis of handshake motion between humans. IEEE/ASME International Conference on Advanced Intelligent Mechatronics, AIM, 2011, pp. 560–565. [CrossRef]
  86. Evers, V.; Winterboer, A.; Pavlin, G.; Groen, F. The Evaluation of Empathy, Autonomy and Touch to Inform the Design of an Environmental Monitoring Robot. SOCIAL ROBOTICS, ICSR 2010; Ge, S.; Li, H.; Cabibihan, J.; Tan, Y., Eds., 2010, Vol. 6414, pp. 285–294.
  87. Cramer, H.; Kemper, N.; Amin, A.; Wielinga, B.; Evers, V. ’Give me a hug’: the effects of touch and autonomy on people’s responses to embodied social agents. Comput. Animat. 2009, 20, 437–445. [Google Scholar] [CrossRef]
  88. Nomura, T.; Suzuki, T.; Kanda, T.; Kato, K. Measurement of Anxiety toward Robots. Proceedings - IEEE International Workshop on Robot and Human Interactive Communication, 2006, pp. 372–377. [CrossRef]
  89. Zheng, X.; Shiomi, M.; Minato, T.; Ishiguro, H. How Can Robots Make People Feel Intimacy Through Touch? J. Robot. Mechatron. 2020, 32, 51–58. [Google Scholar] [CrossRef]
  90. Wang, Q.; Liu, D.; Carmichael, M.G.; Aldini, S.; Lin, C.T. Computational Model of Robot Trust in Human Co-Worker for Physical Human-Robot Collaboration. IEEE Robot. Autom. Lett. 2022, 7, 3146–3153. [Google Scholar] [CrossRef]
  91. Eimontaite, I.; Gwilt, I.; Cameron, D.; Aitken, J.M.; Rolph, J.; Mokaram, S.; Law, J. Language-free graphical signage improves human performance and reduces anxiety when working collaboratively with robots. Int. J. Adv. Manuf. 2019, 100, 55–73. [Google Scholar] [CrossRef]
  92. Memar, A.H.; Esfahani, E.T. Objective Assessment of Human Workload in Physical Human-robot Cooperation Using Brain Monitoring. ACM Trans. Hum.-Robot Interact. 2020, 9, 1–21. [Google Scholar] [CrossRef]
  93. Yamada, Y.; Umetani, Y.; Hirasawa, Y. Proposal of a psychophysiological experiment system applying the reaction of human pupillary dilation to frightening robot motions. IEEE SMC’99 Conference Proceedings. 1999 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.99CH37028), 1999, Vol. 2, pp. 1052–1057 vol.2. [CrossRef]
  94. Maurtua, I.; Ibarguren, A.; Kildal, J.; Susperregi, L.; Sierra, B. Human–robot collaboration in industrial applications: Safety, interaction and trust. Int. J. Adv. Robot 2017, 14, 1729881417716010. [Google Scholar] [CrossRef]
  95. Sahin, M.; Savur, C. Evaluation of Human Perceived Safety during HRC Task using Multiple Data Collection Methods. 2022 17th Annual System of Systems Engineering Conference (SOSE), 2022, pp. 465–470. [CrossRef]
  96. Paez Granados, D.F.; Yamamoto, B.A.; Kamide, H.; Kinugawa, J.; Kosuge, K. Dance Teaching by a Robot: Combining Cognitive and Physical Human–Robot Interaction for Supporting the Skill Learning Process. IEEE Robot. Autom. Lett. 2017, 2, 1452–1459. [Google Scholar] [CrossRef]
  97. Materna, Z.; Kapinus, M.; Beran, V.; Smrž, P.; Zemčík, P. Interactive Spatial Augmented Reality in Collaborative Robot Programming: User Experience Evaluation. 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), 2018-08, pp. 80–87. [CrossRef]
  98. Schmidtler, J.; Bengler, K.; Dimeas, F.; Campeau-Lecours, A. A questionnaire for the evaluation of physical assistive devices (quead): Testing usability and acceptance in physical human-robot interaction. 2017 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2017, 2017, Vol. 2017–January, pp 876. [Google Scholar] [CrossRef]
  99. Reyes-Uquillas, D.; Hsiao, T. Safe and intuitive manual guidance of a robot manipulator using adaptive admittance control towards robot agility. Robot. Comput. Integr. Manuf. 2021, 70. [Google Scholar] [CrossRef]
  100. Shiomi, M.; Hirano, T.; Kimoto, M.; Iio, T.; Shimohara, K. Gaze-Height and Speech-Timing Effects on Feeling Robot-Initiated Touches. J. Robot. Mechatron. 2022, 32, 68–75. [Google Scholar] [CrossRef]
  101. Bock, N.; Hoffmann, L.; Pütten, R.V. Your Touch Leaves Me Cold, Robot. ACM/IEEE International Conference on Human-Robot Interaction, 2018, pp. 71–72. [CrossRef]
  102. Li, J.; Ju, W.; Reeves, B. Touching a Mechanical Body: Tactile Contact With Body Parts of a Humanoid Robot Is Physiologically Arousing. J. hum. robot interact. 2017, 6, 118–130. [Google Scholar] [CrossRef]
  103. Gopinathan, S.; Otting, S.; Steil, J. A User Study on Personalized Stiffness Control and Task Specificity in Physical Human-Robot Interaction. Front. Robot. AI. 2017, 4. [Google Scholar] [CrossRef]
  104. Der Spaa, L.; Gienger, M.; Bates, T.; Kober, J. Predicting and Optimizing Ergonomics in Physical Human-Robot Cooperation Tasks. Proceedings - IEEE International Conference on Robotics and Automation, 2020, pp. 1799–1805. [CrossRef]
  105. Carmichael, M.; Aldini, S.; Liu, D. Human user impressions of damping methods for singularity handling in human-robot collaboration. Australasian Conference on Robotics and Automation, ACRA, 2017, Vol. 2017-December, pp. 107–113.
  106. Novak, D.; Beyeler, B.; Omlin, X.; Riener, R. Workload Estimation in Physical Human-Robot Interaction Using Physiological Measurements. Interact. Comput. 2015, 27, 616–629. [Google Scholar] [CrossRef]
  107. Ferland, F.; Aumont, A.; Letourneau, D.; Michaud, F. Taking Your Robot For a Walk: Force-Guiding a Mobile Robot Using Compliant Arms. Proceedings of the 8TH ACM/IEEE International on Human-Robot Interaction (HRI 2013); Kuzuoka, H.; Evers, V.; Imai, M.; Forlizzi, J., Eds., 2013, pp. 309–316.
  108. Chen, T.; Kemp, C. Lead me by the hand: Evaluation of a direct physical interface for nursing assistant robots. 5th ACM/IEEE International Conference on Human-Robot Interaction, HRI 2010, 2010, pp. 367–374. [Google Scholar] [CrossRef]
  109. Yazdani, A.; Novin, R.; Merryweather, A.; Hermans, T. ; IEEE. DULA and DEBA: Differentiable Ergonomic Risk Models for Postural Assessment and Optimization in Ergonomically Intelligent pHRI. 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 9124–9131. [CrossRef]
  110. Behrens, R.; Pliske, G.; Umbreit, M.; Piatek, S.; Walcher, F.; Elkmann, N. A Statistical Model to Determine Biomechanical Limits for Physically Safe Interactions With Collaborative Robots. Front. Robot. AI. 2022, 8. [Google Scholar] [CrossRef]
  111. Tan, J.T.C.; Duan, F.; Kato, R.; Arai, T. Safety Strategy for Human–Robot Collaboration: Design and Development in Cellular Manufacturing. Adv. Robot . 2010, 24, 839–860. [Google Scholar] [CrossRef]
  112. Michalos, G.; Kousi, N.; Karagiannis, P.; Gkournelos, C. Seamless human robot collaborative assembly - An automotive case study | Elsevier Enhanced Reader. Mechatronics 2018. [Google Scholar] [CrossRef]
  113. Eyam, A.; Mohammed, W.; Martinez Lastra, J. Emotion-driven analysis and control of human-robot interactions in collaborative applications. Sensors 2021, 21. [Google Scholar] [CrossRef]
  114. Amanhoud, W.; Hernandez Sanchez, J.; Bouri, M.; Billard, A. Contact-initiated shared control strategies for four-arm supernumerary manipulation with foot interfaces. Int. J. Rob. Res. 2021, 40, 986–1014. [Google Scholar] [CrossRef]
  115. Aleotti, J.; Micelli, V.; Caselli, S. Comfortable robot to human object hand-over. 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication, 2012, pp. 771–776. [CrossRef]
  116. Pan, M.K.; Croft, E.A.; Niemeyer, G. Evaluating Social Perception of Human-to-Robot Handovers Using the Robot Social Attributes Scale (RoSAS). Proceedings of the 2018 ACM/IEEE International Conference on Human-Robot Interaction; ACM: Chicago IL USA, 2018; pp. 443–451. [Google Scholar] [CrossRef]
  117. Lambrecht, J.; Nimpsch, S. Human Prediction for the Natural Instruction of Handovers in Human Robot Collaboration. 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), 2019, pp. 1–6. [CrossRef]
  118. Lagomarsino, M.; Lorenzini, M.; Balatti, P.; Momi, E.D.; Ajoudani, A. Pick the Right Co-Worker: Online Assessment of Cognitive Ergonomics in Human-Robot Collaborative Assembly. IEEE Trans. Cogn. Develop. I: 1–1. Conference Name. [CrossRef]
  119. Hald, K.; Rehm, M.; Moeslund, T.B. Proposing Human-Robot Trust Assessment Through Tracking Physical Apprehension Signals in Close-Proximity Human-Robot Collaboration. 2019 28th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), 2019, pp. 1–6. [CrossRef]
  120. Strabala, K.W.; Lee, M.K.; Dragan, A.D.; Forlizzi, J.L.; Srinivasa, S.; Cakmak, M.; Micelli, V. Towards Seamless Human-Robot Handovers. J. hum. robot interact. -01. [CrossRef]
  121. Qin, M.; Brawer, J.; Scassellati, B. Task-Oriented Robot-to-Human Handovers in Collaborative Tool-Use Tasks. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 1327–1333. [Google Scholar] [CrossRef]
  122. Nowak, J.; Fraisse, P.; Cherubini, A.; Daures, J.P. Assistance to Older Adults with Comfortable Robot-to-Human Handovers. 2022 IEEE International Conference on Adv. Robot. and Its Social Impacts (ARSO), 2022-05, pp. 1–6. [CrossRef]
  123. Huber, M.; Radrich, H.; Wendt, C.; Rickert, M.; Knoll, A.; Brandt, T.; Glasauer, S. ; IEEE. Evaluation of a novel biologically inspired Trajectory Generator in Human-Robot Interaction. RO-MAN 2009: The 18TH IEEE International Symposium on Robot and Human Interactive Communication, 2009, pp. 712–+.
  124. Nikolaidis, S.; Hsu, D.; Srinivasa, S. Human-robot mutual adaptation in collaborative tasks: Models and experiments. Int. J. Rob. Res. S: 36, 618–634. Publisher; -01. [CrossRef]
  125. Muller, T.; Subrin, K.; Joncheray, D.; Billon, A.; Garnier, S. Transparency Analysis of a Passive Heavy Load Comanipulation Arm. IEEE Trans. Hum. Mach. Syst. [CrossRef]
  126. Wang, Y.; Lematta, G.J.; Hsiung, C.P.; Rahm, K.A.; Chiou, E.K.; Zhang, W. Quantitative Modeling and Analysis of Reliance in Physical Human–Machine Coordination. J Mech Robot . 2019, 11. [Google Scholar] [CrossRef]
  127. Lorenzini, M.; Kim, W.; De Momi, E.; Ajoudani, A. A synergistic approach to the Real-Time estimation of the feet ground reaction forces and centers of pressure in humans with application to Human-Robot collaboration. IEEE Robot. Autom. Lett. 2018, 3, 3654–3661. [Google Scholar] [CrossRef]
  128. Kim, W.; Lee, J.; Peternel, L.; Tsagarakis, N.; Ajoudani, A. Anticipatory Robot Assistance for the Prevention of Human Static Joint Overloading in Human-Robot Collaboration. IEEE Robot. Autom. Lett. 2018, 3, 68–75. [Google Scholar] [CrossRef]
  129. Peternel, L.; Fang, C.; Tsagarakis, N.; Ajoudani, A. A selective muscle fatigue management approach to ergonomic human-robot co-manipulation. Robot. Comput. Integr. Manuf. 2019, 58, 69–79. [Google Scholar] [CrossRef]
  130. Peternel, L.; Tsagarakis, N.; Caldwell, D.; Ajoudani, A. Robot adaptation to human physical fatigue in human–robot co-manipulation. Auton. Robot. 2018, 42, 1011–1021. [Google Scholar] [CrossRef]
  131. Kowalski, C.; Brinkmann, A.; Hellmers, S.; Fifelski-von Bohlen, C.; Hein, A. Comparison of a VR-based and a rule-based robot control method for assistance in a physical human-robot collaboration scenario. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN); IEEE: Napoli, Italy, 2022; pp. 722–728. [Google Scholar] [CrossRef]
  132. Figueredo, L.; Aguiar, R.; Chen, L.; Chakrabarty, S.; Dogar, M.; Cohn, A. Human Comfortability: Integrating Ergonomics and Muscular-Informed Metrics for Manipulability Analysis during Human-Robot Collaboration. IEEE Robot. Autom. Lett. 2021, 6, 351–358. [Google Scholar] [CrossRef]
  133. Gao, Y.; Chang, H.; Demiris, Y. User Modelling Using Multimodal Information for Personalised Dressing Assistance. IEEE Access 2020, 8, 45700–45714. [Google Scholar] [CrossRef]
  134. Chen, L.; Figueredo, L.; Dogar, M. Planning for Muscular and Peripersonal-Space Comfort during Human-Robot Forceful Collaboration. IEEE-RAS International Conference on Humanoid Robots, 2019, pp. 409–416. [CrossRef]
  135. Madani, A.; Niaz, P.; Guler, B.; Aydin, Y.; Basdogan, C. Robot-Assisted Drilling on Curved Surfaces with Haptic Guidance under Adaptive Admittance Control. IEEE International Conference on Intelligent Robots and Systems, 2022, Vol. 2022-October, pp. 3723–3730. [CrossRef]
Figure 1. Number of studies emerged in each pHRI category.
Figure 1. Number of studies emerged in each pHRI category.
Preprints 75342 g001
Figure 2. Number of Studies emerged in each category of the direct pHRI classified based on duration.
Figure 2. Number of Studies emerged in each category of the direct pHRI classified based on duration.
Preprints 75342 g002
Figure 3. Number of studies emerged in each category of the direct pHRI classified based on purpose.
Figure 3. Number of studies emerged in each category of the direct pHRI classified based on purpose.
Preprints 75342 g003
Figure 4. Number of studies emerged in each indirect pHRI category.
Figure 4. Number of studies emerged in each indirect pHRI category.
Preprints 75342 g004
Figure 5. Most representative measurable dimensions in the pHRI literature according to the results of the literature review. This diagram is adapted to pHRI from the holistic conceptual model adapted for HRI proposed in [11].
Figure 5. Most representative measurable dimensions in the pHRI literature according to the results of the literature review. This diagram is adapted to pHRI from the holistic conceptual model adapted for HRI proposed in [11].
Preprints 75342 g005
Figure 6. Number of studies emerged in each human factor type.
Figure 6. Number of studies emerged in each human factor type.
Preprints 75342 g006
Figure 7. Number of studies emerged in each measurable dimension.
Figure 7. Number of studies emerged in each measurable dimension.
Preprints 75342 g007
Figure 8. Number of studies that evaluated each human factor type in each direct pHRI category.
Figure 8. Number of studies that evaluated each human factor type in each direct pHRI category.
Preprints 75342 g008
Figure 9. Number of studies that evaluated each human factor type in each indirect pHRI category.
Figure 9. Number of studies that evaluated each human factor type in each indirect pHRI category.
Preprints 75342 g009
Figure 10. Number of studies emerged in each human factor usage category.
Figure 10. Number of studies emerged in each human factor usage category.
Preprints 75342 g010
Figure 11. Number of studies relied on each quantification approach.
Figure 11. Number of studies relied on each quantification approach.
Preprints 75342 g011
Figure 12. Quantification approaches proposed in the quantification category. MM stands for Mathematical Model and ML stands for Machine Learning.
Figure 12. Quantification approaches proposed in the quantification category. MM stands for Mathematical Model and ML stands for Machine Learning.
Preprints 75342 g012
Figure 13. Number of studies relied on each assessment approach in each human factor type.
Figure 13. Number of studies relied on each assessment approach in each human factor type.
Preprints 75342 g013
Table 1. The characteristics and studies of each subcategory in pHRI.
Table 1. The characteristics and studies of each subcategory in pHRI.
Categories Characteristics Studies
Direct pHRI Purpose Non-Functional Touch includes studies that involve physical touch between the robot and the human with the intention of communicating a psychological state, such as social touch [46], or for the purpose of exploration or curiosity, as in [47]. [25,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89]
Functional Touch Includes studies that involved physical contact between the robot and the human for a specific purpose, such as manipulation, assistance, or control, i.e. instrumental touch. [17,36,49,83,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109]
Duration Consistent Contact Includes studies requiring continuous direct contact throughout the interaction. [17,36,52,56,57,59,71,74,76,77,79,82,90,91,92,93,96,98,99,102,103,104,105,106,107,108,109,110]
Inconsistent Contact Includes studies where direct contact is not necessary throughout the entire interaction. [25,46,47,48,49,50,51,53,54,58,60,61,62,63,64,65,66,67,68,69,70,72,73,75,78,80,81,83,84,85,86,87,88,89,94,95,97,100,101]
Indirect pHRI Assembly Includes studies where one agent holds a part while the other agent assembles it with another part. [109,111,112,113,114]
Handover Includes studies where one agent is handing over an object to the other agent. [20,21,22,35,64,109,115,116,117,118,119,120,121,122,123]
Co-manipulation Includes studies where both human and robot agents manipulate an object in the environment with the goal of changing its position or orientation. [104,109,112,124,125,126,127,128]
Atypical Includes studies that do not fall into any of the other categories of indirect pHRI, such as assistive holding/drilling, or dressing assistance. [23,109,129,130,131,132,133,134,135]
Table 2. Common studies detected across direct pHRI categories.
Table 2. Common studies detected across direct pHRI categories.
Functional Touch Non-Functional Touch
Consistent Contact [17,36,90,91,92,93,96,98,99,102,103,104,105,106,107,108] [109] [52,56,57,59,71,74,76,77,79,82]
Inconsistent Contact [49,83,94,95,97,100,101] [25,46,47,48,49,51,53,54,58,60,61,62,63,64,65,66,67,68,69,70,72,73,75,78,80,81,83,84,85,86,87,88,89]
Table 3. A brief description of some measurable dimensions.
Table 3. A brief description of some measurable dimensions.
Measurable Dimension Description
Tactility Indicates the perceived pleasantness when touching a robot.
Physical Comfort Includes studies that have evaluated human posture, muscular effort, joint torque overloading, peri-personal space, comfortable handover, legibility, and physical safety.
Mechanical Transparency “Quantifies the ability of a robot to follow the movements imposed by the operator without noting any resistant effort” [125]. It includes predictability of the robot’s motion in following user physical instructions, naturalness and smoothness of the motion, sense of being in control, responsiveness to physical instruction of participants, feeling of resistive force, and frustration.
Robot Perception Indicates the user’s perception towards the robot. It includes attitudes, impressions, opinions, preferences, favourability, likeability, willingness for another interaction, behaviour perception, politeness, anthropomorphism, animacy, vitality, perceived naturalness, agency, perceived intelligence, competence, perceived safety, emotional security, harmlessness, toughness, familiarity, friendship, companionship, friendliness, warmth, psychological comfort, helpfulness, reliable alliance, acceptance, ease of use, and perceived performance.
Perceived Intuition Includes goal perception, whether the robot understands the goal of the task or not, robot intelligence, willingness to follow the robot’s suggestion, dependability, understanding of robot intention and perceived robot helpfulness.
Conveying Emotions Indicates humans’ perspective on how they should convey their emotions to robots by physical touch.
Receiving Emotions Indicates humans’ perspective of how humans expect to receive a robot’s emotions through physical touch.
Emotional State Indicates recognition of a human’s emotional state during interaction without necessarily conveying their emotions using physical touch.
Table 4. The characteristics and studies of human factor type categories.
Table 4. The characteristics and studies of human factor type categories.
Human Factor Type Measurable Dimension Studies
Cognitive Ergonomics Mental Workload [17,22,25,35,36,70,90,92,97,106,108,114,117,118]
Stress [17,20,36,57,101,118]
Physical Ergonomics Pain Sensitivity [110]
Tactility [55,74,76]
Physical Comfort [20,21,22,23,90,104,109,112,115,120,121,122,127,128,129,130,131,132,133,134]
Mechanical Transparency [104,105,107,114,125,135]
Belief Robot Perception [17,25,35,36,46,49,50,51,52,53,54,55,57,58,59,62,65,66,68,69,71,72,74,77,80,81,83,84,85,86,87,88,91,94,95,96,98,100,103,116,123,135]
Trust [54,58,84,86,90,114,118,119,124,126]
Perceived Intuition [36,70,87,114]
Enjoyability [25,70]
Anxiety [17,36,59,88,91]
Affect Emotional state [36,50,51,53,61,63,64,65,67,70,71,77,83,91,93,102,111,113]
Conveying Emotions [46,48,53,73,75,78]
Receiving Emotions [56,60,79,82,89]
Table 5. Measurable dimensions addressed within each pHRI type and human factor usage.
Table 5. Measurable dimensions addressed within each pHRI type and human factor usage.
Human Factor Type Measurable Dimension Direct pHRI Indirect pHRI
Purpose Duration
Functional Touch Non-functional Touch Consistent Inconsistent Assembly Handover Co-manipulation Atypical
Cognitive Ergonomics Mental Workload [90]* [92]* [106]* [36]* [17] [97] [25] [70] [108] [36]* [17] [92]* [106]* [90]* [97] [25] [70] [114] [35] [117] [22] [118]*
Stress [36]* [17] [57] [57] [101] [20] [118]*
Physical Ergonomics Pain Sensitivity [110]*
Tactility [55][76] [74] [76] [74]
Physical Comfort [90]* [109]* [90]* [109]* [112] [109]* [115] [20] [120] [121] [21][122]* [22] [109]* [112] [104]* [127]* [128]* [109]* [131] [129]* [23]* [132]*[130]* [133] [134]* [109]*
Mechanical Transparency [104] [105] [107] [104] [105] [107] [114] [125] [125] [135]
Belief Robot Perception [36]* [54] [52]* [46]* [53] [55] [57] [58] [62][25] [65] [66] [68] [69] [72] [77] [80] [81]* [83] [84] [85] [86] [87] [88]* [51] [49] [50] [47] [59] [71] [74] [90] [17] [93] [36]* [96] [52]* [56] [57] [99] [59] [71] [102][76] [91] [95] [49] [50] [51] [46]* [53]* [58] [62][100] [25] [65] [66] [68] [69] [72] [80] [81]* [83] [84] [85] [86] [87] [88]* [47] [54] [70] [35][116]* [123] [135]
Trust [90]* [54] [58] [84] [86] [54] [58] [84] [86] [114] [119]* [118]* [124][126]*
Perceived Intuition [36]* [70] [87] [36]* [70] [87] [114]
Enjoyability [25] [70] [25] [70]
Anxiety [36]* [17] [88]* [59] [91] [36]* [17] [88]* [59]
Affect Emotional State [91] [36]* [83] [93] [102] [50] [53]*[61] [63] [64] [65] [67]* [70] [83] [51] [77] [36]* [91] [77] [17] [50] [53]*[61] [63] [64][65] [67]* [70] [83] [51] [113]* [111] [64]
Conveying Emotions [48] [75] [46]* [53]* [73]* [78]* [48] [46]* [73]*; [53]* [75] [78]*
Receiving Emotions [60] [79] [82] [89] [56]* [79] [82] [60] [89]
14.1cm * studies that have proposed quantification approaches.
Table 6. Quantification approaches employed within each measurable dimension.
Table 6. Quantification approaches employed within each measurable dimension.
HF MD Q Machine Learning Model PS Mathematical Model PD
PD PS PD TP PS
Cognitive Ergonomics Mental Workload [114] [25] [70] [108] [97] [35] [117] [22] [106]* [92]* [36]* [17] [118]* [90]*
Stress [57] [101] [36]* [17] [20] [118]*
Physical Ergonomics Pain Sensitivity [110]*
Tactility [55] [76] [74]
Physical Comfort [115] [20] [120] [121] [21] [23]* [129]* [109]* [20] [131] [132] [133] [104] [134] [127] [128][90]* [122] [130]* [22] [125]*
Mechanical Transparency [114][104][105][107][135] [112]
Belief Robot Perception [46]* [47] [53]* [55] [57] [58] [59] [62][100] [25] [65] [66] [68] [69] [71] [72] [74] [77] [80] [81] [83] [84] [85] [86] [87] [123] [88] [91] [36]* [17] [51] [96] [35][116]* [49] [50] [95] [94] [98]*[103] [99] [135] [52]* [36]* [36]* [54] [81]
PD PS PD TP PS
Trust [54] [58] [114] [84] [86] [124] [90]* [90]* [119]* [126]*
Perceived Intuition [114] [70] [87] [36]* [36]* [36]*
Enjoyability [25] [70]
Anxiety [88] [91] [59] [36]* [17]
Affect Emotional State [53]* [61] [63] [64] [65] [70] [83] [36]* [50] [111] [71] [51] [77] [74] [91] [113]* [102] [83] [36]* [93] [111] [67]*
Conveying Emotions [75] [48] [78]* [46] [53]* [73]*
Receiving Emotions [56]*
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated