SlideShare ist ein Scribd-Unternehmen logo
1 von 8
Ergonomics-for-One in a Robotic Shopping Cart for the
                            Blind
                      Vladimir A. Kulyukin                                                    Chaitanya Gharpure
    Computer Science Assistive Technology Laboratory                         Computer Science Assistive Technology Laboratory
           Department of Computer Science                                           Department of Computer Science
                 Utah State University                                                    Utah State University
                 vladimir.kulyukin@usu.edu                                                      cpg@cc.usu.edu

ABSTRACT                                                                     homes and communities as long as possible. This situation brings
Assessment and design frameworks for human-robot teams                       a unique challenge and opportunity to assistive robotics: is it
attempt to maximize generality by covering a broad range of                  possible to develop robotic devices that will enable older and
potential applications. In this paper, we argue that, in assistive           disabled individuals to maintain their independence and thereby
robotics, the other side of generality is limited applicability: it is       reduce the cost of institutionalized medical care?
oftentimes more feasible to custom-design and evolve an                      Vision is a sensory modality that deteriorates with age. As of
application that alleviates a specific disability than to spend              now, there are 11.4 million visually impaired individuals living in
resources on figuring out how to customize an existing generic               the U.S. [8]. Grocery shopping is an activity that presents a
framework. We present a case study that shows how we used a                  barrier to independence for many visually impaired people who
pure bottom-up learn-through-deployment approach inspired by                 either do not go grocery shopping at all or rely on sighted guides,
the principles of ergonomics-for-one to design, deploy and                   e.g., friends, spouses, and partners. Traditional navigation aids,
iteratively re-design a proof-of-concept robotic shopping cart for           such as guide dogs and white canes, are not adequate in such
the blind.                                                                   dynamic and complex environments as modern supermarkets.
                                                                             These aids cannot help their users with macro-navigation, which
Categories and Subject Descriptors                                           requires topological knowledge of the environment. Nor can they
H.1.2 [Models and Principles]: User/Machine Systems – human                  assist with carrying useful payloads.
factors.                                                                     In summer 2004, the Computer Science Assistive Technology
                                                                             Laboratory (CSATL) of the Department of Computer Science
General Terms                                                                (CS) of Utah State University (USU) launched a project whose
Performance, Design, Experimentation, Human Factors.                         objective is to build a robotic shopping cart for the visually
                                                                             impaired. In our previous publications, we examined several
                                                                             technical aspects of robot-assisted navigation for the blind, such
Keywords                                                                     as RFID-based localization, greedy free space selection, and
assistive technology, navigation and wayfinding for the blind,               topological knowledge representation [6, 7]. In this paper, we
assistive robotics, ergonomics-for-one.                                      focus on how the ergonomic aspects of the system have evolved
                                                                             through fitting trials in two dynamic and complex environments.
1. INTRODUCTION                                                              The paper is organized as follows. In Section 2, we review
Current demographic trends in the U.S. signify a demographic
                                                                             relevant research on human-robot interaction (HRI). In Section 3,
shift from a population where most people are relatively young to
                                                                             we discuss the basic principles of ergonomics-for-one and present
a population where most people are relatively old. In 2000, U.S.
                                                                             an ergonomics-for-one analysis to identify the key elements of the
residents aged 65 and older constituted approximately 12 percent
                                                                             performance gap between the blind individual and the task of
of the population. It is projected that by 2030 people aged 65 and
                                                                             independent grocery shopping. In Section 4, we present our initial
older will make up 22 percent of the U.S. population [11]. In
                                                                             design aimed at bridging several elements of the performance gap.
essence, older adults will make up an increasingly larger percent
                                                                             In Section 5, we present our navigation trials with a small sample
of the population.
                                                                             of visually impaired participants. We describe our experiment
The primary concern for aging adults is the decline in their                 design, analyze the collected data, and present the participants'
sensory-motor abilities. Surveys show that a great number of U.S.            feedback. In Section 6, we return to bridging the performance
residents would like to maintain their independent status in their           gap. We describe how we deployed the robotic shopping cart in a
                                                                             supermarket and discuss what ergonomic modifications we made
                                                                             in the system after several fitting trials in that environment. In
 Permission to make digital or hard copies of all or part of this work for   Section 7, we focus on two major challenges for the future. In
 personal or classroom use is granted without fee provided that copies are   Section 8, we give our conclusions.
 not made or distributed for profit or commercial advantage and that
 copies bear this notice and the full citation on the first page. To copy
 otherwise, or republish, to post on servers or to redistribute to lists,    2. RELATED WORK
 requires prior specific permission and/or a fee.                            In recent years, many researchers have asked the question of what
 HRI’04, March 2–3, 2006, Salt Lake City, Utah, USA.                         it means to design a human-robot team and to measure its
 Copyright 2006 ACM 1-58113-000-0/00/0004…$5.00.
performance from the standpoint of human-robot interaction              3. ERGONOMICS-FOR-ONE
(HRI). Fong et al. [1] identify common metrics for task-oriented        We discovered the field of ergonomics-for-one during a literature
HRI through a thorough analysis of existing HRI applications.           search for conceptual frameworks to help guide the design of the
Several task-specific metrics are proposed and suggested for            robotic shopping cart for the blind and assess its performance with
standardization. It is claimed that the metrics are applicable at any   human participants. The term ergonomics-for-one was first coined
point of the HRI spectrum starting at pure teleoperation and            by McQuistion in 1993 [9]. In brief, ergonomics-for-one is the
ending with full autonomy. Howard [5] proposes a systematic             science of fitting the task to a particular individual who wants to
approach for assessing performance of a human-robot team. The           repeatedly accomplish the task in a given environment.
approach takes into account the capabilities of both human and
robotic agents and integrates the effect of cognitive stress during     Although the use of the term is recent, the ideas underlying it are
continuous operation. Goodrich and Olsen [4] propose several            not novel: occupational therapists have been devising solutions to
principles of efficient HRI based on the lessons from evaluating        find better fits between individuals and their environments for
neglect tolerance and interface efficiency. Each principle, e.g.,       decades [13]. In ergonomics-for-one, a solution to a specific
manipulate the world instead of the robot, is motivated by              fitting task is referred to as an accommodation system that
relevant factors from cognitive information processing. Olsen and       consists of five components: 1) essential task functions; 2)
Goodrich [10] propose several HRI metrics for leveraging human          equipment used to accomplish the task; 3) inputs and outputs; 4)
attention to develop HRI interfaces that enhance the task               environment in which the task is accomplished; and 5) individual
effectiveness of the human-robot team. Scerri et al. [12] discuss       with a disability who desires to accomplish the task.
criteria to determine how to change the autonomy level of the
                                                                        A performance gap is identified by comparing the essential task
robot to enhance the performance of the human-robot team on the
                                                                        functions to the disabled individual's capabilities. The assistive
basis of decision costs. Yanco and Drury [14] propose qualitative
                                                                        device is designed to bridge the performance gap. It should be
taxonomies and qualitative and quantitative metrics for human-
                                                                        noted that ergonomics-for-one does not define the order in which
robot performance evaluation. Fong et al. [2] offer a detailed
                                                                        the components must be considered so long as eventually all of
survey of socially interactive robots and a taxonomy of design
                                                                        them are taken into account.
methods and system components.
Given these approaches, it is natural to ask whether any of them        At this point one may ask the question of what exactly
can be readily applied to assess the performance or guide the           ergonomics-for-one brings to the table that HRI does not. The
design of a robotic shopping cart for the blind. We do not believe      answer is a different research methodology. First, ergonomics-for-
that, at this point in time, this question has a positive answer.       one does not assume the existence a common framework that can
There are several reasons to justify our belief. First, these           be used to design and assess every assistive device imaginable. To
approaches assume that the operator is capable of maintaining           be sure, there are standard procedures that evaluate the extent of
visual contact with the robot, either continuously, when the            individual disabilities, e.g., standard vision or hearing tests. But
operator is collocated in the same task space, or part of the time,     these tests are used only as inputs to the design process. Second,
when the operator is remote and interacts with the robot through        ergonomics-for-one is inherently bottom-up in that it places a
an interface intermediary. Second, existing assessment                  great deal of emphasis on fitting trials and learning through
frameworks focus on interfaces, autonomy, and task efficiency           deployment [13]. The objective of such trials, also known as
and do not take into account the ergonomic interaction between          initial usability tests or walk-throughs, is to ascertain the user's
the human and the robot. Third, the scope of many frameworks is         comfort, ease of use, preference, and other psychosocial elements.
simply too broad. The other side of generality is inapplicability: it   We used a group of five visually impaired individuals from the
is oftentimes more feasible to custom-design a new approach than        local visually impaired community in Logan, Utah, to help us
to spend resources on figuring out how to customize an existing         analyze the components of an accommodation system that could
generic framework. It was this realization that prompted us to          help them do grocery shopping independently. The youngest
look for inspiration outside of the traditional HRI realms.             individual was 13, the oldest was 47. Two participants were white
                                                                        cane users. The other three participants used both guide dogs and
                                                                        white canes. We met with the individuals in an informal setting
The paper is organized as follows. In Section 2, we review
                                                                        and asked them about what it would take them to shop
relevant research on human-robot interaction (HRI). In Section 3,
                                                                        independently. To minimize peer pressure, we met with each
we discuss the basic principles of ergonomics-for-one and present
                                                                        individual separately.
an ergonomics-for-one analysis to identify the key elements of the
performance gap between the blind individual and the task of            Essential task functions: The interviews helped us identify five
independent grocery shopping. In Section 4, we present our initial      essential task functions: 1) getting to a supermarket; 2) finding the
design aimed at bridging several elements of the performance gap.       needed grocery items; 3) getting through a cash register; 4)
In Section 5, we present our navigation trials with a small sample      leaving the store; and 5) getting home. None of the individuals
of visually impaired participants. We describe our experiment           had any problems getting to a supermarket and getting home from
design, analyze the collected data, and present the participants'       a supermarket. Logan has a free bus system with a network of bus
feedback. In Section 6, we return to bridging the performance           stops all over the city and several suburbs. All places where one
gap. We describe how we deployed the robotic shopping cart in a         can buy groceries have bus stops close by.
supermarket and discuss what ergonomic modifications we made
in the system after several fitting trials in that environment. In      Function 2 was refined into three sub-functions: 1) navigating to a
Section 7, we focus on two major challenges for the future. In          shelf section with a needed grocery item; 2) finding the needed
Section 8, we give our conclusions.                                     item on the shelf; and 3) placing the item in a shopping basket.
Function 3 was refined into four sub-functions: 1) navigating to a    next to other Lays items or other potato chip brands. This is even
cash register; 2) placing the items from the basket on the belt; 3)   more of a problem with smaller items, like small bags of
paying for the items; and 4) placing the bagged items back into       sunflower seeds. Assuming that the individual does not have
the shopping basket. Function 4 was refined into three sub-           manual dexterity problems, once the item is found, the individual
functions: 1) getting to the exit; 2) leaving the basket in a         can place it into a shopping basket.
designated place; and 3) exiting the store.
                                                                      Fifth, the same navigational challenges apply to the function of
Equipment: None of the participants did any grocery shopping          getting through a cash register. Additional challenges are knowing
on their own. They either did not do any grocery shopping or used     when it is time to start placing the items on the conveyor belt,
sighted guides: parents, siblings, or partners. The only equipment    paying for the items, and putting the bagged items back into the
used by the participants were white canes, guide dogs, and            basket.
shopping baskets. They could not use shopping carts, because
                                                                      Sixth, when the shopper is ready to leave the store, she again has
they could not simultaneously handle guide dogs or white canes
                                                                      to navigate to the exit, thus confronting the navigational
and push the carts.
                                                                      challenges identified above, and place the basket in the properly
Inputs and outputs: When asked how they would prefer to               designated place.
interact with an assistive grocery shopping device, if they had
one, the participants suggested speech and keypad as input            4. BRIDGING THE GAP: PART I
options and speech and dynamic Braille as output options.
                                                                      4.1 On to a Robotic Shopping Cart
Environment: The target environment was a typical supermarket.        After considering the first performance gap component,
There are several features that make this environment particularly    independent use of a shopping cart, we concluded that the
challenging. First, there is always some shopper traffic. On          navigation performance of the shopping cart had to be automated.
certain days, e.g. Saturday, and during certain hours, e.g.,          Effectively, the robotic shopping cart would act as a supermarket
between 6 and 8 pm, the shopper traffic is at its highest. Second,    guide for blind shoppers. This is, by no means, a novel idea as the
there are indigenous processes already in place, e.g., shelf re-      field of AI robotics had built robotic guides before [3]. None of
stocking, cleaning, product scanning, etc., that cannot be            the guides, however, were specifically built for blind shoppers in
disrupted. Third, the products are periodically re-shuffled and re-   supermarkets. As far as we could see, we had two options:
arranged, and free open spaces are occupied with temporary            building a new robotic base with a shopping cart mounted on top
displays and stands.                                                  of it or mounting a shopping cart on top of an existing robotic
Individual: Two participants were completely blind. Three             base. We chose the second option, because the first option, after a
participants had light perception, i.e., they were able to            preliminary cost analysis, looked prohibitively expensive for a
distinguish between light and dark. All participants were             research prototype. In addition, we already had experience
ambulatory, did not have any serious speech impediments,              mounting equipment on our Pioneer 2DX robotic base from the
hearing problems, or cognitive disabilities.                          ActivMedia Corporation when we experimented with our robot-
                                                                      assisted navigation for the blind in indoor environments [6, 7].
After the interviews, we identified the performance gap that had
to be addressed by the accommodation system. First, using a
guide dog and/or a white cane with a shopping cart is not feasible.
Neither guide dogs nor white canes would help avoid front
obstacles if the blind shopper has to push the cart in addition to
handling a guide dog or using a white cane. Of course, it is
possible to use a basket, but the shopper would then be restricted
to buying a small number of items.
Second, since the shopper cannot independently navigate, she
needs to communicate her intentions to a sighted guide. This
would be ordinarily done in natural language if the guide is
human.
Third, the visually impaired participant needs assistance with
                                                                                Figure 1: RoboCart in Lee's MarketPlace.
navigating to shelf sections with specific grocery items. In an
environment where end points of routes remain static, many guide      Thus, we built a polyvinyl chloride (PVC) pipe structure, securely
dog handlers and cane users can learn routes after several trials.    mounted it on top of the Pioneer 2DX robotic base, and then
However, this assumption does not hold in supermarkets due to         placed a large shopping basket into that structure. The resulting
constant re-arrangements and re-shufflings of products.               design, which we called RoboCart, is shown in Figure 1. As one
                                                                      can easily see from Figure 2, the RoboCart design is a
Fourth, even if it is assumed that the blind shopper can find her     modification of RG, our indoor robotic guide for the blind that we
way to the correct shelf section, she still needs to pick the right   built in 2003-2004 on top of another Pioneer 2DX base. It should
item. For example, suppose that the blind shopper wants to buy a      be noted that this is a proof-of-concept design. The back
bag of Lays Classic and finds her way to the correct shelf section    directional wheel of the base is small, which results in the
with Lays potato chips. There is always a chance that the shopper     inherent imbalance of this design. While we have not observed
will pick a wrong bag as Lays Classic bags are typically placed       any accidents in which RoboCart tipped over, the future design
will be modified to have a four-wheel base so that the device will   below, there is another reason why Braille may not be a viable
never tip over and injure the blind shopper.                         option for some users.

                                                                     4.3 How Do We Navigate?
                                                                     When we started thinking about bridging the navigational
                                                                     component of the performance gap, we realized that we had little
                                                                     knowledge about what aspects of navigation might be important
                                                                     to the blind navigator. We also did not know if our
                                                                     communication choices described in the previous section would
                                                                     be ergonomically acceptable to blind individuals. Finally, we
                                                                     wanted to find out whether the presence or absence of the human
                                                                     navigator behind the robot affects the robot's navigation.
                                                                     To answer these questions, we decided to conduct a series of
                                                                     fitting trials. We had to find a suitable environment for the trials.
                                                                     We had started negotiations with Lee's MarketPlace, a
     Figure 2: RG, an indoor robotic guide for the blind.            supermarket in Logan, Utah, about the possibility of testing
                                                                     RoboCart in their supermarket. But the negotiations were still in
4.2 How Do We Communicate?                                           progress. We ruled out tests in our CS Department, because we
Upon entering the supermarket, the shopper needs to                  had already tested our robotic guide in the CS Department rather
communicate her wishes to RoboCart. The input options that we        extensively and had achieved satisfactory results [6, 7].
considered were automatic speech recognition (ASR) and keypad.
                                                                     We chose to conduct fitting trials at the USU Center for Persons
When using ASR, the blind shopper would wear a wireless
                                                                     with Disabilities (CPD). The CPD occupies an entire building on
microphone coupled to an over-the-ear headphone and
                                                                     the North USU Campus. The building has an area of 40,000
communicate her intentions to the robot through speech. We will
                                                                     square feet. It has numerous offices, classrooms, laboratories,
not go into details here on why we ruled out speech as an input
                                                                     lounges, and bathrooms. Another challenging aspect of this
option, because we have described our reasons in detail in our
                                                                     environment that makes it similar to a supermarket is numerous
previous publications [6]. In brief, our ASR experiments, both in
                                                                     activities occur there during its working hours. Thus, other people
noisy and noise-free environments, had recognition rates of below
                                                                     going about their business, i.e., human traffic, are an integral part
50 percent even though all of our participants were native
                                                                     of the environment.
speakers of American English. Our decision to rule out ASR as an
input option should not be construed as a general argument
against ASR as an HRI mode. Rather, we concluded that, given         5. FITTING TRIALS
the current state of the art in commercial ASR and the constraints
of our problem, we should explore the keypad first.
                                                                     5.1 Experiment Design
                                                                     We used the paired differences strategy to design our pilot
The input option that we chose was a small 10-key Belkin             experiments. In a paired difference experiment, one is interested
numeric keypad. The layout of keys on the keypad is the same as      to find the mean difference between two methods of conducting
the layout of keys on the cell phone. Since many visually            some activity, which, in our case, is navigation. A data point is
impaired people use cell phones, our thinking was that the           obtained by numerically measuring the performances of two
learning curve would not be steep. In addition, the number 5 key     participants, say X and Y, from two different samples doing a
on Belkin keypads has a small plastic protrusion that the visually   designated activity and computing the difference between the two
impaired can sense through touch. Once the number 5 key is           measurements. When a sample of differences is obtained, one can
found, it is easy to find the other keys.                            test two hypotheses: the null hypothesis, H 0 : µ D = 0 , against
                                                                     one of the three alternative hypotheses, H a : µ D > 0 ,
When compared to ASR, the keypad does reduce input ambiguity.
However, even with the keypad the proverbial problem of shared
vocabulary does not go away. The user still must know what to        H a : µ D < 0 , and H a : µ D ≠ 0 , where µ D is the mean
type into the robot to make the robot do what the user wants. To
overcome this problem, we decided to create a Braille directory.     difference. Essentially, H 0 suggests that there is no difference in
The directory was to be realized as a Braille sheet with
                                                                     performance, whereas H a ’s suggest that there may be a
instructions that map each destination to a short sequence of
numbers. The semantics of each line was to be as follows: if you     difference.    The      test    statistic    is    a     one-sample
want to go to destination X, please type this numerical sequence
into the keypad.
                                                                     t = x D / s D / n D , where x D is the sample mean
The next element of the communication gap is output. The             difference, s D is the standard deviation of the differences, and
options that we considered were synthetic speech and dynamic         n D is the number of differences.
Braille displays. As we investigated dynamic Braille displays, we
found out that they were expensive: the cheapest option we could     We selected a total of 9 routes in the environment. Each route was
find was approximately 5K USD. Originally, the cost was the          more than 40 meters in length and had 3 to 5 turns. In our case,
main reason why we decided on synthetic speech. As we discuss        our first sample consisted of the robot. Since we focused on
10              11         12         13        14        15      16       17      18

                                         3.64            3.60       -1.96      0.90      -4.07     -4.14   3.24     3.91    1.90

                                                                Table 1: T-statistics at α = 0.05 and df=4.


              Route,Part       0                           1                   2                    3               4                5
              10               65.79, 65.84                59.87, 60.89        62.90, 63.45         61.76, 65.08    61.94, 63.75     65.49, 67.14
              11               70.83, 72.27                55.93, 57.33        56.91, 59.27         55.45, 59.16    56.42, 59.22     72.29, 73.67
              12               70.94, 72.25                72.56, 73.68        75.78, 79.43         71.79, 98.53    73.96, 75.47     69.96, 72.46
              13               87.88, 89.93                87.06, 87.93        89.45, 91.03         86.29, 88.78    86.29, 88.55     87.70, 90.17
              14               55.76, 56.29                82.21, 83.71        84.60, 86.46         83.22, 84.89    83.61, 84.86     55.81, 57.12
              15               57.35, 60.30                79.23, 80.15        78.85, 81.29         79.88, 81.85    85.91, 88.43     56.11, 64.49
              16               120.74, 123.34              93.11, 97.67        95.54, 102.48        90.91, 93.09    98.70, 101.48    122.87, 129.00
              17               124.72, 123.34              83.93, 103.48       87.10, 103.16        91.17, 94.08    90.25, 92.14     125.10, 126.82
              18               129.11, 130.61              130.89, 139.79      97.35, 100.58        84.46, 86.63    88.14, 92.67     130.89, 139.77
                                                                   Table 2: 95% confidence intervals.

navigation and guidance, we used the robotic guide shown in                                   own and the robot navigating with a visually impaired human. On
Figure 2. Our second sample consisted of five visually impaired                               the other routes, i.e., 12, 13, and 18, there appears to be
participants. To obtain the measurements, we ran the robot five                               insufficient evidence to reject H 0 . In other words, the presence
times on each of the designated routes and recorded the time-to-
completion, e.g., the amount of time it took the robot to complete                            of the human navigator behind the robot does not appear to affect
the route. For each route, the average time-to-completion was                                 the robot's performance. Since, in computing µ D , we subtracted
computed from the five runs.                                                                  the robot's time-to-completion from a participant's time-to-
We then had each participant use the robot to navigate the same                               completion, the positive t-statistics that exceed 2.776 suggest that
routes. The robot would inform the participant through synthetic                              the robot was slower without the navigator than with the
speech about its present location. We told each participant the                               navigator. On the contrary, the negative t-statistics smaller than
keypad codes for all destinations. The participant would type in                              2.776 suggest that the robot was slower with the navigator than by
the destination code through the keypad attached to a pole on the                             itself.
back of the robot. Each route was navigated five times and the                                To verify the validity of these observations, we analyzed the data
time-to-completion measurements were taken for each participant.                              through confidence intervals. We computed 95% time-to-
For each participant we computed the average time-to-                                         completion confidence intervals for each route and each
completion. A sample of differences that we used to test the                                  participant, including the robot. Table 2 gives the confidence
hypothesis was obtained by computing the difference between the                               intervals for all routes and participants. The robot is listed as
robot's average times-to-completion and the participants' times-to-                           participant 0. The interval table verifies the conclusions of the
completion.                                                                                   hypothesis tests. For example, both ends of the robot's confidence
We     chose       to   test       the        third      alternative     hypothesis,          interval for route 10 given in column 0 are greater than the
H a : µ D ≠ 0 , at α = 0.05
                                                                                              corresponding ends of participants 1 through 4 and are essentially
                                              as the level of significance. The               the same as the ends of participant 5. The same observations can
rejection          region          for            this          hypothesis          is        be made on routes 11, 16, and 17. This seems to verify the test of
| t |> tα / 2 = t 0.025 = 2.776 ,             and has 4 degrees of freedom.                   hypothesis conclusion that on these routes the robot without the
                                                                                              navigator appeared to be slower than the robot with the navigator.
Table 1 contains the sample t- statistics for each of the 9 routes                            The same technique can be applied to routes 14 and 15 on which,
numbered 10 through 18. These statistics should not be viewed as                              according to the test of hypothesis conclusion, the robot appeared
definitive. The paired differences design requires that the sample                            to be faster without the navigator than with the navigator. The
of differences be random. This assumption may not be satisfied in                             robot's confidence intervals for these routes are to the left of the
our case, because we did not choose the five individuals                                      confidence intervals of participants 1 through 4 and coincide with
randomly. Their names were given to us by referral.                                           the ends of participant 5.
                                                                                              To understand what was causing these differences, we looked at
5.2 Results                                                                                   the video footage of the runs. The video footage of the robot runs
The results in Table 1 tell us that on routes 10, 11, 14, 15, 16, and
                                                                                              without the navigator on routes 10, 11, 16, and 17 showed that
17,   H0    is rejected, because the absolute value of the t-statistic is                     there was quite a bit of human traffic in the hallways. The video
larger than 2.776. In other words, on these routes there appears to                           footage of the robot runs with the navigator on the same routes
be a significant difference between the robot navigating on its                               showed that in the cases of participants 1 through 4, the amount of
human traffic in the hallways declined. The exception was               was on the right or left. It also told me when it was turning left or
participant 5 for whom the amount of human traffic remained             right. I would appreciate voice messages being spoken more
essentially the same. Since the robot's speed decreases with the        loudly. I understand that you cannot make it too loud without
number of obstacles present in front, the robot traveled more           making it obnoxious to the people around me. Perhaps, it could
slowly in the presence of human traffic.                                be done with one over-ear headphone or a shoulder speaker so
                                                                        that I have my other ear available to me.
The situation was reversed on routes 14 and 15. During the robot
runs without the navigator the amount of human traffic was              Comment 6: Overall, I felt very comfortable navigating with the
minimal. However, when we ran the robot with the human                  robot. I felt even more comfortable after I learned on one of the
navigators, human traffic picked up considerably. The exception         runs that the robot can recover from situations when it gets lost
again was participant 5 for whom the amount of human traffic did        by finding an alternate route. Self-correction is a valuable feature
not change. Our conclusion was that the amount of human traffic,        of this device.
i.e., the number of people on route, is a nuisance variable that
may have contributed to the differences in robot performance.           Comment 7: Make sure that there is no chance of the robot going
                                                                        off the course.
Another interesting observation that we made as we watched the
video footage was the effect of the occasional mismatch between
the verbalized intent of the robot and the robot's actual actions. At
several T-intersections the robot would tell the navigator that it
was turning left and then, due to the presence of people, it started
drifting to the right before actually making a left turn. When that
happened, we observed that several human navigators pulled hard
on the robot's handle, sometimes driving the robot to a virtual
halt. We conjecture that when a communication mismatch occurs,
i.e., when the robot starts doing something other than what it said
it would do, the human navigators become apprehensive and try
to stop the robot. Since these mismatches happened on the routes
where the robot performed better without the navigator than with
the navigator, we concluded that the mismatches may have
contributed to the performance difference.
                                                                                    Figure 3:RoboCart's Handle, Design 1.
While watching the video footage, we also observed a different
kind of communication problem that occurred several times
during u-turns. The robot would inform the navigator that it had
started making a u-turn after it had already started executing the
maneuver. Although the robot's message was accurate, it came a
bit too late and, as discussed in the next section, caused some
                                                                        6. BRIDGING THE GAP: PART II
                                                                        In fall 2004, we received permission from Lee's MarketPlace to
discomfort on the part of the participants.
                                                                        use their supermarket as a test site for our experiments. We asked
                                                                        two visually impaired individuals to participate in a series of
5.3 Participants Speak                                                  fitting trials in the store. On several occasions we ran RoboCart
After the experiments, we conducted informal verbal interviews          on its own. The objective was to learn through deployment what
with the participants and recorded their responses. The interviews      modifications in ergonomic design and navigation were required.
consisted of several questions about navigation safety and user
comfort. The objective was to let the participants give us feedback
on their experiences. Below we give several comments verbatim.
                                                                        6.1 Ergonomic Modifications
                                                                        As shown in Figure 2, our original design included a guide leash.
Comment 1: There was some abruptness in the robot motion.               However, the participants expressed a wish that the dog leash be
Stops and slows down too suddenly. Sometimes it accelerates too         replaced with a static handle. When asked why, the participants
fast.                                                                   said that the dog leash did not give them sufficient feedback as to
                                                                        what direction the robot was taking them. This wish was
Comment 2: Sometimes the robot tells you too late when it is            expressed both by the cane users and guide dog handlers. It was
about to make a u-turn. This is a problem if you have a guide dog       quite understandable that cane users expressed this wish because
and need to tell him to get out of the robot's way.                     the cane is firm and does resemble a static handle. We were
Comment 3: A little more user training up front would help. Let         surprised, however, to hear the same complaint from the guide
me touch the robot and give me some time to get comfortable with        dog handlers. As we took a closer look at how the guide dogs are
the keypad.                                                             handled, the explanation presented itself immediately. It turns out
                                                                        that guide dog handlers do not use the leash when their dogs are at
Comment 4: The robot slows down at turns and then it kicks into         work. They use a firm leather handle attached to a special harness
high gear too abruptly. I have a back injury and so such changes        on the back of the dog. The handle enables the handler to give
in speed were felt a lot.                                               directions to the animal as well as to receive immediate haptic
                                                                        feedback about the animal's movement. The leash is used only
Comment 5: The communication was clear and helpful. The
                                                                        when the dog is not at work and is being treated as a pet.
robot told me when I got to a destination, whether the destination
The above lesson led to our first modification - the addition of a     boxes and similar movable objects being placed into them by the
static handle shown in Figure 3. The keypad hangs on the right         store staff. After investigating the possibility of using Markov
pole of the handle. After several trials in Lee's MarketPlace, we      localization [3], we decided against it because of safety concerns.
realized that the keypad's position was inconvenient for the user.     Most applications of Markov localization indoors are based on
It is difficult to access the keypad quickly when the robot is         laser range finding. Laser range finding does not perform well in
moving. To reach for the keypad requires letting go of the handle.     large open spaces or environments with large glassy surfaces that
Using the other hand is impossible as it is occupied with a cane or    absorb laser signals. The performance of Markov localization is
a leash.                                                               not predictable in dynamic environments and degrades in the
                                                                       presence of numerous dynamic obstacles.
                                                                       We considered extending our RFID-based navigation to open
                                                                       spaces by putting portable towers with RFID tags. We rejected
                                                                       this idea, too, because it called for a great deal of calibration and
                                                                       instrumentation and could be too disruptive to the indigenous
                                                                       business processes. We discussed our problem with the
                                                                       supermarket's owner and a senior store manager. They suggested
                                                                       that we put masking tape lines on the floor and use them for
                                                                       navigating large open spaces. In their opinion, if the system were
                                                                       to be deployed in their store permanently, they could easily paint
                                                                       such lines on the floor. As long as the paint was resistant to the
                                                                       floor wax, the lines were not a problem.

            Figure 4: RoboCart's Handle, Design 2.
Figure 4 shows how we modified this design by changing the
position of the keypad. We purchased the wireless version of the
same keypad, attached it to a small plastic rectangle, and then
attached the rectangle to the handle's bar. This position allows the
navigator to quickly reach for the keypad during the navigation
without letting go of the handle.
We also learned that Braille may not be feasible. Of the seven
visually impaired people that we informally polled about the
possibility of using Braille on the robot only 2 were comfortable
with the idea. As we investigated the matter further and talked                      Figure 6: RoboCart following a line.
with the assistive technology specialists at the USU Center for        RoboCart was equipped with a small LogiTech web camera.
Persons with Disabilities, we learned that only a small fraction of    Figure 5 shows how the camera was added to the robotic base.
visually impaired people use Braille. This fraction consists mostly    We put one masking tape line from the lobby and up to the aisles.
of people who are blind from birth. People who lose vision later       A simple vision-based line following algorithm was written and
in their lives due to accident, illness, or age either never learn     successfully tested on several runs. Figure 6 shows how RoboCart
Braille or use it rather slowly.                                       follows the line to reach an aisle. Once in the aisles, our original
                                                                       RFID-based navigation algorithm was used. One aisle has 5
                                                                       shelves on both sides. An RFID tag is placed every 3 meters on
                                                                       the 2nd or 3rd shelf on both sides of the aisle so that the robot's
                                                                       RFID antenna can detect it. Thus, every aisle in which we tested
                                                                       RoboCart is equipped with 10 RFID tags: 5 on the left side and 5
                                                                       on the right side. There is also a designated cash register where
                                                                       RoboCart takes the blind shopper. The cash register is equipped
                                                                       with two RFID tags. The first tag makes RoboCart stop and
                                                                       inform the blind shopper that the products can be unloaded onto
                                                                       the belt on the right. The second tag informs the shopper that she
                                                                       has to wait for the bagger to put the bags into the cart. The store
                                                                       management was comfortable with this instrumentation plan.
                 Figure 5: RoboCart's Camera.

6.2 Navigation Modifications                                           7. A Glimpse of the Future
Several important modifications were made to our navigation            When we learned that Braille may not be a viable option, we
algorithm. The original algorithm was designed for structured          replaced Braille with a voice-based directory based on synthetic
indoor environments [7], which was fine for navigating                 speech. Instead of reading Braille, a blind person uses the keypad
supermarket aisles. The algorithm did not work in large open           to scroll up and down the voice menu in which each line is spoken
spaces, such as supermarket lobbies. Besides having a lot of           to the user by the speech synthesis software. Modern grocery
customer traffic, supermarket lobbies constantly change in terms       stores carry thousands of items. One challenge that we are
of their layout due to promotion displays, flower stands, product
currently investigating is how to organize the directory for easy      10. REFERENCES
browsing.                                                              [1] Fong, T., Kaber, D., Lewis, M., Scholtz, J., Schultz, A., and
Since the RFID tags must be placed on both sides of each aisle,            Steinfield, A. Common Metrics for Human-Robot
the approximate layout of the store must be known in advance.              Interaction. In citeseer.csail.mit.edu/667916.html.
This layout must be maintained by the store. Thus, another             [2] Fong, T., Nourbakhsh, I., and Dautenhahn, K. A Survey of
challenge is how the store maintains the layout so that the robotic        Socially Interactive Robots. Robotics and Autonomous
shopping cart always guides the blind shopper to the appropriate           Systems, 42:143-166, 2003.
shelf.
                                                                       [3] Fox, D., Burgard, W., and Thrun, S. Markov Localization for
Another ergonomic challenge is access to individual items.                 Mobile Robots in Dynamic Environments. Journal of AI
RoboCart leads the blind person to shelf sections, not to                  Research, 11:391-427, 1999.
individual items. For example, it will guide the person to the shelf   [4] Goodrich, M. and Olsen, D. Seven Principles of Efficient
section with Lays potato chips, but the person still has to pick up        Human-Robot Interaction. In Proceedings of the IEEE
an individual bag and put it into the robot's shopping basket. To          International Conference on Systems, Man, and Cybernetics,
address this problem, we have integrated a small portable barcode          pp. 3943-3948. IEEE, October 2003.
reader into the system. Grocery stores already use barcode
reading technologies to keep track of their price inventories. The     [5] Howard, A. A Methodology to Assess Performance of
scenario that we are currently experimenting with is as follows:           Human-Robotic Systems in Achievement of Collective
RoboCart gets the blind shopper to a shelf section with a bunch of         Tasks. In Proceedings of the International Conference on
individual items, the shopper then uses a handheld barcode reader          Intelligent Robots and Systems (IROS). IEEE/RSJ, July 2005.
to read the barcodes on the shelf until the barcode of the right       [6] Kulyukin, V., Gharpure, C., De Graw, N., Nicholson, J., and
item is found. Under this scenario, the shopper has to find the            Pavithran, S. A Robotic Wayfinding System for the Visually
shelf and then slide the barcode reader along the shelf and listen         Impaired. In Proceedings of the Innovative Applications of
until a speech message tells the user that the proper barcode is           Artificial Intelligence Conference (IAAI), pp. 864-869.
read.                                                                      AAAI, July 2004.
                                                                       [7] Kulyukin, V., Gharpure, C., Nicholson, J., and S. Pavithran.
8. CONCLUSIONS                                                             RFID in Robot-Assisted Indoor Navigation for the Visually
In this paper we showed how the basic principles of ergonomics-            Impaired. In Proceedings of the IEEE International
for-one were applied to the design and development of a proof-of-          Conference on Intelligent Robots and Systems (IROS).
concept robotic shopping cart for the blind. We identified the             IEEE/RSJ, October 2004.
performance gap that must be overcome by an accommodation              [8] LaPlante, M. and Carson, D. Disability in the United States:
system that allows the blind to shop independently. We described           Prevalence and Causes. U.S. Department of Education,
our initial usability tests and showed how the tests shaped the            Washington, DC, 2000.
ergonomic modifications of the system.
                                                                       [9] McQuistion, L. Rehabilitation Engineering: Ergonomics for
                                                                           One. Ergonomics in Design, January:9-10, 1993.
9. ACKNOWLEDGMENTS
The first author would like to acknowledge that this research has      [10] Olsen, D and Goodrich, M. Metrics for Evaluating Human-
been supported, in part, through NSF CAREER grant (IIS-                     Robot Interactions. In Performance Metrics for Intelligent
0346880) and two Community University Research Initiative                   Systems (PERMIS). NIST, September, 2003.
(CURI) grants (CURI-04 and CURI-05) from the State of Utah.            [11] Pollack, M. Intelligent Technology for the Aging Population.
We would like to thank Mr. Lee Badger, the owner of Lee's                   AI Magazine, 26(2):9-24, 2005.
MarketPlace, for allowing us to use his supermarket in Logan,
                                                                       [12] Scerri, P., Pynadath, D., and Tambe, M. Toward Adjustable
Utah, as a research site. We are grateful to John Nicholson, our
                                                                            Autonomy for the Real World. Journal of AI Research,
research colleague at USU CSATL, for helping us to conduct
                                                                            17:171-228, 2002.
many fitting trials. We would like to thank Ying Bing, a CS
graduate student, for implementing the line following algorithm.       [13] E. Berg. Ergonomics in Health Care and Rehabilitation.
Finally, we would like to thank the visually impaired participants          Butterworth-Heinemann, Woburn, MA, 1998.
in our experiments for their valuable feedback.                        [14] Yanco, H. and Drury, J. A Taxonomy for Human-Robot
                                                                            Interaction. In Proceedings of the AAAI Fall Symposium on
                                                                            Human-Robot Interaction, pp. 111-119, 2002.

Weitere ähnliche Inhalte

Andere mochten auch

Bass jayne u11a1 language comparison
Bass jayne u11a1 language comparisonBass jayne u11a1 language comparison
Bass jayne u11a1 language comparisonjaynebass
 
Sociaal intranet summerschool 2012
Sociaal intranet summerschool 2012Sociaal intranet summerschool 2012
Sociaal intranet summerschool 2012Sasja Beerendonk
 
Negara yang bebas korupsi
Negara yang bebas korupsiNegara yang bebas korupsi
Negara yang bebas korupsiFelice Luce
 
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Vladimir Kulyukin
 
RFID in Robot-Assisted Indoor Navigation for the Visually Impaired
RFID in Robot-Assisted Indoor Navigation for the Visually ImpairedRFID in Robot-Assisted Indoor Navigation for the Visually Impaired
RFID in Robot-Assisted Indoor Navigation for the Visually ImpairedVladimir Kulyukin
 
Stadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekStadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekArchief 2.0
 
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingPresentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingSasja Beerendonk
 
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced Maps
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced MapsA Software Tool for Rapid Acquisition of Streetwise Geo-Referenced Maps
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced MapsVladimir Kulyukin
 
Control; the spanish eperience c.gallardo
Control; the spanish eperience c.gallardoControl; the spanish eperience c.gallardo
Control; the spanish eperience c.gallardocargallardofron
 
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudAn Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudVladimir Kulyukin
 
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceNarrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceVladimir Kulyukin
 
Op weg naar het zuidland
Op weg naar het zuidlandOp weg naar het zuidland
Op weg naar het zuidlandArchief 2.0
 
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...Vladimir Kulyukin
 
Rijksmonumenten.info
Rijksmonumenten.infoRijksmonumenten.info
Rijksmonumenten.infoArchief 2.0
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vladimir Kulyukin
 
Rotterdam Open Data
Rotterdam Open DataRotterdam Open Data
Rotterdam Open DataArchief 2.0
 

Andere mochten auch (20)

Bass jayne u11a1 language comparison
Bass jayne u11a1 language comparisonBass jayne u11a1 language comparison
Bass jayne u11a1 language comparison
 
Sociaal intranet summerschool 2012
Sociaal intranet summerschool 2012Sociaal intranet summerschool 2012
Sociaal intranet summerschool 2012
 
Bolsas femeninas
Bolsas femeninasBolsas femeninas
Bolsas femeninas
 
EUscreen
EUscreenEUscreen
EUscreen
 
Mengenal sunnah sii
Mengenal sunnah siiMengenal sunnah sii
Mengenal sunnah sii
 
Negara yang bebas korupsi
Negara yang bebas korupsiNegara yang bebas korupsi
Negara yang bebas korupsi
 
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
Robot-Assisted Shopping for the Blind: Issues in Spatial Cognition and Produc...
 
RFID in Robot-Assisted Indoor Navigation for the Visually Impaired
RFID in Robot-Assisted Indoor Navigation for the Visually ImpairedRFID in Robot-Assisted Indoor Navigation for the Visually Impaired
RFID in Robot-Assisted Indoor Navigation for the Visually Impaired
 
Stadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiekStadsarchief Ieper zkt. publiek
Stadsarchief Ieper zkt. publiek
 
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleidingPresentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
Presentatie Gastcollege Gamification voor Hogeschool Utrecht - ICT opleiding
 
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced Maps
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced MapsA Software Tool for Rapid Acquisition of Streetwise Geo-Referenced Maps
A Software Tool for Rapid Acquisition of Streetwise Geo-Referenced Maps
 
Control; the spanish eperience c.gallardo
Control; the spanish eperience c.gallardoControl; the spanish eperience c.gallardo
Control; the spanish eperience c.gallardo
 
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the CloudAn Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
An Algorithm for In-Place Vision-Based Skewed 1D Barcode Scanning in the Cloud
 
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path InferenceNarrative Map Augmentation with Automated Landmark Extraction and Path Inference
Narrative Map Augmentation with Automated Landmark Extraction and Path Inference
 
Op weg naar het zuidland
Op weg naar het zuidlandOp weg naar het zuidland
Op weg naar het zuidland
 
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...
ShopTalk: Independent Blind Shopping Through Verbal Route Directions and Barc...
 
PizzaRaptor
PizzaRaptorPizzaRaptor
PizzaRaptor
 
Rijksmonumenten.info
Rijksmonumenten.infoRijksmonumenten.info
Rijksmonumenten.info
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
 
Rotterdam Open Data
Rotterdam Open DataRotterdam Open Data
Rotterdam Open Data
 

Ähnlich wie Ergonomics-for-One in a Robotic Shopping Cart for the Blind

Soft Computing in Robotics
Soft Computing in RoboticsSoft Computing in Robotics
Soft Computing in Roboticsijtsrd
 
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...Vladimir Kulyukin
 
iWalker: Toward a Rollator-Mounted Wayfinding System for the Elderly
iWalker: Toward a Rollator-Mounted Wayfinding System for the ElderlyiWalker: Toward a Rollator-Mounted Wayfinding System for the Elderly
iWalker: Toward a Rollator-Mounted Wayfinding System for the ElderlyVladimir Kulyukin
 
A Novel Approach for Machine Learning-Based Identification of Human Activities
A Novel Approach for Machine Learning-Based Identification of Human ActivitiesA Novel Approach for Machine Learning-Based Identification of Human Activities
A Novel Approach for Machine Learning-Based Identification of Human ActivitiesIRJET Journal
 
diplomarbeit-alesiaivanova-clear
diplomarbeit-alesiaivanova-cleardiplomarbeit-alesiaivanova-clear
diplomarbeit-alesiaivanova-clearAlessya Ivanova
 
Human–Robot Interaction: Status and Challenges. Sheridan MIT
Human–Robot Interaction: Status and Challenges. Sheridan MITHuman–Robot Interaction: Status and Challenges. Sheridan MIT
Human–Robot Interaction: Status and Challenges. Sheridan MITeraser Juan José Calderón
 
Artificial Cognition for Human-robot Interaction
Artificial Cognition for Human-robot InteractionArtificial Cognition for Human-robot Interaction
Artificial Cognition for Human-robot InteractionSubmissionResearchpa
 
CIV8331 ASSIGNMENT PRESENTATION.pptx
CIV8331 ASSIGNMENT PRESENTATION.pptxCIV8331 ASSIGNMENT PRESENTATION.pptx
CIV8331 ASSIGNMENT PRESENTATION.pptxJamiluAdamuSalisu
 
IRJET- Swarm Robotics
IRJET- Swarm RoboticsIRJET- Swarm Robotics
IRJET- Swarm RoboticsIRJET Journal
 
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...ssuser50a5ec
 
Temporal Reasoning Graph for Activity Recognition
Temporal Reasoning Graph for Activity RecognitionTemporal Reasoning Graph for Activity Recognition
Temporal Reasoning Graph for Activity RecognitionIRJET Journal
 
Swarm Robotics An Overview
Swarm Robotics An OverviewSwarm Robotics An Overview
Swarm Robotics An Overviewijtsrd
 
A Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition ProceduresA Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition Proceduresijtsrd
 
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD ijcsit
 
A Path Towards Autonomous Machines
A Path Towards Autonomous MachinesA Path Towards Autonomous Machines
A Path Towards Autonomous MachinesAlejandro Franceschi
 
Distributed Robotics A Primer
Distributed Robotics A PrimerDistributed Robotics A Primer
Distributed Robotics A Primerijtsrd
 
Acoustic event characterization for service robot using convolutional networks
Acoustic event characterization for service robot using convolutional networksAcoustic event characterization for service robot using convolutional networks
Acoustic event characterization for service robot using convolutional networksIJECEIAES
 
On Fault Detection and Diagnosis in Robotic Systems.pdf
On Fault Detection and Diagnosis in Robotic Systems.pdfOn Fault Detection and Diagnosis in Robotic Systems.pdf
On Fault Detection and Diagnosis in Robotic Systems.pdfhenibelgacem
 

Ähnlich wie Ergonomics-for-One in a Robotic Shopping Cart for the Blind (20)

Soft Computing in Robotics
Soft Computing in RoboticsSoft Computing in Robotics
Soft Computing in Robotics
 
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...
Command, Goal Disambiguation, Introspection, and Instruction in Gesture-Free ...
 
iWalker: Toward a Rollator-Mounted Wayfinding System for the Elderly
iWalker: Toward a Rollator-Mounted Wayfinding System for the ElderlyiWalker: Toward a Rollator-Mounted Wayfinding System for the Elderly
iWalker: Toward a Rollator-Mounted Wayfinding System for the Elderly
 
A Novel Approach for Machine Learning-Based Identification of Human Activities
A Novel Approach for Machine Learning-Based Identification of Human ActivitiesA Novel Approach for Machine Learning-Based Identification of Human Activities
A Novel Approach for Machine Learning-Based Identification of Human Activities
 
diplomarbeit-alesiaivanova-clear
diplomarbeit-alesiaivanova-cleardiplomarbeit-alesiaivanova-clear
diplomarbeit-alesiaivanova-clear
 
Human–Robot Interaction: Status and Challenges. Sheridan MIT
Human–Robot Interaction: Status and Challenges. Sheridan MITHuman–Robot Interaction: Status and Challenges. Sheridan MIT
Human–Robot Interaction: Status and Challenges. Sheridan MIT
 
Artificial Cognition for Human-robot Interaction
Artificial Cognition for Human-robot InteractionArtificial Cognition for Human-robot Interaction
Artificial Cognition for Human-robot Interaction
 
CIV8331 ASSIGNMENT PRESENTATION.pptx
CIV8331 ASSIGNMENT PRESENTATION.pptxCIV8331 ASSIGNMENT PRESENTATION.pptx
CIV8331 ASSIGNMENT PRESENTATION.pptx
 
IRJET- Swarm Robotics
IRJET- Swarm RoboticsIRJET- Swarm Robotics
IRJET- Swarm Robotics
 
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...
Analysis_of_Navigation_Assistants_for_Blind_and_Visually_Impaired_People_A_Sy...
 
liu2019.pdf
liu2019.pdfliu2019.pdf
liu2019.pdf
 
Temporal Reasoning Graph for Activity Recognition
Temporal Reasoning Graph for Activity RecognitionTemporal Reasoning Graph for Activity Recognition
Temporal Reasoning Graph for Activity Recognition
 
Swarm Robotics An Overview
Swarm Robotics An OverviewSwarm Robotics An Overview
Swarm Robotics An Overview
 
A Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition ProceduresA Complete Analysis of Human Action Recognition Procedures
A Complete Analysis of Human Action Recognition Procedures
 
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD
CONSIDERATION OF HUMAN COMPUTER INTERACTION IN ROBOTIC FIELD
 
IEEE-ROBIO-2013-115-120
IEEE-ROBIO-2013-115-120IEEE-ROBIO-2013-115-120
IEEE-ROBIO-2013-115-120
 
A Path Towards Autonomous Machines
A Path Towards Autonomous MachinesA Path Towards Autonomous Machines
A Path Towards Autonomous Machines
 
Distributed Robotics A Primer
Distributed Robotics A PrimerDistributed Robotics A Primer
Distributed Robotics A Primer
 
Acoustic event characterization for service robot using convolutional networks
Acoustic event characterization for service robot using convolutional networksAcoustic event characterization for service robot using convolutional networks
Acoustic event characterization for service robot using convolutional networks
 
On Fault Detection and Diagnosis in Robotic Systems.pdf
On Fault Detection and Diagnosis in Robotic Systems.pdfOn Fault Detection and Diagnosis in Robotic Systems.pdf
On Fault Detection and Diagnosis in Robotic Systems.pdf
 

Mehr von Vladimir Kulyukin

Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Vladimir Kulyukin
 
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Vladimir Kulyukin
 
Generalized Hamming Distance
Generalized Hamming DistanceGeneralized Hamming Distance
Generalized Hamming DistanceVladimir Kulyukin
 
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityAdapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityVladimir Kulyukin
 
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...Vladimir Kulyukin
 
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Vladimir Kulyukin
 
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Vladimir Kulyukin
 
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsText Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsVladimir Kulyukin
 
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vladimir Kulyukin
 
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...Vladimir Kulyukin
 
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Vladimir Kulyukin
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vladimir Kulyukin
 
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Vladimir Kulyukin
 
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Vladimir Kulyukin
 
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Vladimir Kulyukin
 
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Vladimir Kulyukin
 
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsWireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsVladimir Kulyukin
 
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...Vladimir Kulyukin
 
A Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsA Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsVladimir Kulyukin
 
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...Vladimir Kulyukin
 

Mehr von Vladimir Kulyukin (20)

Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
Toward Sustainable Electronic Beehive Monitoring: Algorithms for Omnidirectio...
 
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
Digitizing Buzzing Signals into A440 Piano Note Sequences and Estimating Fora...
 
Generalized Hamming Distance
Generalized Hamming DistanceGeneralized Hamming Distance
Generalized Hamming Distance
 
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term SimilarityAdapting Measures of Clumping Strength to Assess Term-Term Similarity
Adapting Measures of Clumping Strength to Assess Term-Term Similarity
 
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
A Cloud-Based Infrastructure for Caloric Intake Estimation from Pre-Meal Vide...
 
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
Exploring Finite State Automata with Junun Robots: A Case Study in Computabil...
 
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
Image Blur Detection with 2D Haar Wavelet Transform and Its Effect on Skewed ...
 
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition LabelsText Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
Text Skew Angle Detection in Vision-Based Scanning of Nutrition Labels
 
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
Vision-Based Localization and Scanning of 1D UPC and EAN Barcodes with Relaxe...
 
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
An Algorithm for Mobile Vision-Based Localization of Skewed Nutrition Labels ...
 
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
Skip Trie Matching: A Greedy Algorithm for Real-Time OCR Error Correction on ...
 
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
Vision-Based Localization & Text Chunking of Nutrition Fact Tables on Android...
 
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
Skip Trie Matching for Real Time OCR Output Error Correction on Android Smart...
 
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
Toward Blind Travel Support through Verbal Route Directions: A Path Inference...
 
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
Eye-Free Barcode Detection on Smartphones with Niblack's Binarization and Sup...
 
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
Eyesight Sharing in Blind Grocery Shopping: Remote P2P Caregiving through Clo...
 
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support FunctionsWireless Indoor Localization with Dempster-Shafer Simple Support Functions
Wireless Indoor Localization with Dempster-Shafer Simple Support Functions
 
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
RoboCart: Toward Robot-Assisted Navigation of Grocery Stores by the Visually ...
 
A Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College StudentsA Wearable Two-Sensor O&M Device for Blind College Students
A Wearable Two-Sensor O&M Device for Blind College Students
 
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...
On the Impact of Data Collection on the Quality of Signal Strength in Wi-Fi I...
 

Kürzlich hochgeladen

The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxLoriGlavin3
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESmohitsingh558521
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxBkGupta21
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxLoriGlavin3
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfPrecisely
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brandgvaughan
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxLoriGlavin3
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 

Kürzlich hochgeladen (20)

The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptxThe Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
The Fit for Passkeys for Employee and Consumer Sign-ins: FIDO Paris Seminar.pptx
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICESSALESFORCE EDUCATION CLOUD | FEXLE SERVICES
SALESFORCE EDUCATION CLOUD | FEXLE SERVICES
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
unit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptxunit 4 immunoblotting technique complete.pptx
unit 4 immunoblotting technique complete.pptx
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptxUse of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
Use of FIDO in the Payments and Identity Landscape: FIDO Paris Seminar.pptx
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdfHyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
Hyperautomation and AI/ML: A Strategy for Digital Transformation Success.pdf
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
WordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your BrandWordPress Websites for Engineers: Elevate Your Brand
WordPress Websites for Engineers: Elevate Your Brand
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptxThe Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
The Role of FIDO in a Cyber Secure Netherlands: FIDO Paris Seminar.pptx
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 

Ergonomics-for-One in a Robotic Shopping Cart for the Blind

  • 1. Ergonomics-for-One in a Robotic Shopping Cart for the Blind Vladimir A. Kulyukin Chaitanya Gharpure Computer Science Assistive Technology Laboratory Computer Science Assistive Technology Laboratory Department of Computer Science Department of Computer Science Utah State University Utah State University vladimir.kulyukin@usu.edu cpg@cc.usu.edu ABSTRACT homes and communities as long as possible. This situation brings Assessment and design frameworks for human-robot teams a unique challenge and opportunity to assistive robotics: is it attempt to maximize generality by covering a broad range of possible to develop robotic devices that will enable older and potential applications. In this paper, we argue that, in assistive disabled individuals to maintain their independence and thereby robotics, the other side of generality is limited applicability: it is reduce the cost of institutionalized medical care? oftentimes more feasible to custom-design and evolve an Vision is a sensory modality that deteriorates with age. As of application that alleviates a specific disability than to spend now, there are 11.4 million visually impaired individuals living in resources on figuring out how to customize an existing generic the U.S. [8]. Grocery shopping is an activity that presents a framework. We present a case study that shows how we used a barrier to independence for many visually impaired people who pure bottom-up learn-through-deployment approach inspired by either do not go grocery shopping at all or rely on sighted guides, the principles of ergonomics-for-one to design, deploy and e.g., friends, spouses, and partners. Traditional navigation aids, iteratively re-design a proof-of-concept robotic shopping cart for such as guide dogs and white canes, are not adequate in such the blind. dynamic and complex environments as modern supermarkets. These aids cannot help their users with macro-navigation, which Categories and Subject Descriptors requires topological knowledge of the environment. Nor can they H.1.2 [Models and Principles]: User/Machine Systems – human assist with carrying useful payloads. factors. In summer 2004, the Computer Science Assistive Technology Laboratory (CSATL) of the Department of Computer Science General Terms (CS) of Utah State University (USU) launched a project whose Performance, Design, Experimentation, Human Factors. objective is to build a robotic shopping cart for the visually impaired. In our previous publications, we examined several technical aspects of robot-assisted navigation for the blind, such Keywords as RFID-based localization, greedy free space selection, and assistive technology, navigation and wayfinding for the blind, topological knowledge representation [6, 7]. In this paper, we assistive robotics, ergonomics-for-one. focus on how the ergonomic aspects of the system have evolved through fitting trials in two dynamic and complex environments. 1. INTRODUCTION The paper is organized as follows. In Section 2, we review Current demographic trends in the U.S. signify a demographic relevant research on human-robot interaction (HRI). In Section 3, shift from a population where most people are relatively young to we discuss the basic principles of ergonomics-for-one and present a population where most people are relatively old. In 2000, U.S. an ergonomics-for-one analysis to identify the key elements of the residents aged 65 and older constituted approximately 12 percent performance gap between the blind individual and the task of of the population. It is projected that by 2030 people aged 65 and independent grocery shopping. In Section 4, we present our initial older will make up 22 percent of the U.S. population [11]. In design aimed at bridging several elements of the performance gap. essence, older adults will make up an increasingly larger percent In Section 5, we present our navigation trials with a small sample of the population. of visually impaired participants. We describe our experiment The primary concern for aging adults is the decline in their design, analyze the collected data, and present the participants' sensory-motor abilities. Surveys show that a great number of U.S. feedback. In Section 6, we return to bridging the performance residents would like to maintain their independent status in their gap. We describe how we deployed the robotic shopping cart in a supermarket and discuss what ergonomic modifications we made in the system after several fitting trials in that environment. In Permission to make digital or hard copies of all or part of this work for Section 7, we focus on two major challenges for the future. In personal or classroom use is granted without fee provided that copies are Section 8, we give our conclusions. not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, or republish, to post on servers or to redistribute to lists, 2. RELATED WORK requires prior specific permission and/or a fee. In recent years, many researchers have asked the question of what HRI’04, March 2–3, 2006, Salt Lake City, Utah, USA. it means to design a human-robot team and to measure its Copyright 2006 ACM 1-58113-000-0/00/0004…$5.00.
  • 2. performance from the standpoint of human-robot interaction 3. ERGONOMICS-FOR-ONE (HRI). Fong et al. [1] identify common metrics for task-oriented We discovered the field of ergonomics-for-one during a literature HRI through a thorough analysis of existing HRI applications. search for conceptual frameworks to help guide the design of the Several task-specific metrics are proposed and suggested for robotic shopping cart for the blind and assess its performance with standardization. It is claimed that the metrics are applicable at any human participants. The term ergonomics-for-one was first coined point of the HRI spectrum starting at pure teleoperation and by McQuistion in 1993 [9]. In brief, ergonomics-for-one is the ending with full autonomy. Howard [5] proposes a systematic science of fitting the task to a particular individual who wants to approach for assessing performance of a human-robot team. The repeatedly accomplish the task in a given environment. approach takes into account the capabilities of both human and robotic agents and integrates the effect of cognitive stress during Although the use of the term is recent, the ideas underlying it are continuous operation. Goodrich and Olsen [4] propose several not novel: occupational therapists have been devising solutions to principles of efficient HRI based on the lessons from evaluating find better fits between individuals and their environments for neglect tolerance and interface efficiency. Each principle, e.g., decades [13]. In ergonomics-for-one, a solution to a specific manipulate the world instead of the robot, is motivated by fitting task is referred to as an accommodation system that relevant factors from cognitive information processing. Olsen and consists of five components: 1) essential task functions; 2) Goodrich [10] propose several HRI metrics for leveraging human equipment used to accomplish the task; 3) inputs and outputs; 4) attention to develop HRI interfaces that enhance the task environment in which the task is accomplished; and 5) individual effectiveness of the human-robot team. Scerri et al. [12] discuss with a disability who desires to accomplish the task. criteria to determine how to change the autonomy level of the A performance gap is identified by comparing the essential task robot to enhance the performance of the human-robot team on the functions to the disabled individual's capabilities. The assistive basis of decision costs. Yanco and Drury [14] propose qualitative device is designed to bridge the performance gap. It should be taxonomies and qualitative and quantitative metrics for human- noted that ergonomics-for-one does not define the order in which robot performance evaluation. Fong et al. [2] offer a detailed the components must be considered so long as eventually all of survey of socially interactive robots and a taxonomy of design them are taken into account. methods and system components. Given these approaches, it is natural to ask whether any of them At this point one may ask the question of what exactly can be readily applied to assess the performance or guide the ergonomics-for-one brings to the table that HRI does not. The design of a robotic shopping cart for the blind. We do not believe answer is a different research methodology. First, ergonomics-for- that, at this point in time, this question has a positive answer. one does not assume the existence a common framework that can There are several reasons to justify our belief. First, these be used to design and assess every assistive device imaginable. To approaches assume that the operator is capable of maintaining be sure, there are standard procedures that evaluate the extent of visual contact with the robot, either continuously, when the individual disabilities, e.g., standard vision or hearing tests. But operator is collocated in the same task space, or part of the time, these tests are used only as inputs to the design process. Second, when the operator is remote and interacts with the robot through ergonomics-for-one is inherently bottom-up in that it places a an interface intermediary. Second, existing assessment great deal of emphasis on fitting trials and learning through frameworks focus on interfaces, autonomy, and task efficiency deployment [13]. The objective of such trials, also known as and do not take into account the ergonomic interaction between initial usability tests or walk-throughs, is to ascertain the user's the human and the robot. Third, the scope of many frameworks is comfort, ease of use, preference, and other psychosocial elements. simply too broad. The other side of generality is inapplicability: it We used a group of five visually impaired individuals from the is oftentimes more feasible to custom-design a new approach than local visually impaired community in Logan, Utah, to help us to spend resources on figuring out how to customize an existing analyze the components of an accommodation system that could generic framework. It was this realization that prompted us to help them do grocery shopping independently. The youngest look for inspiration outside of the traditional HRI realms. individual was 13, the oldest was 47. Two participants were white cane users. The other three participants used both guide dogs and white canes. We met with the individuals in an informal setting The paper is organized as follows. In Section 2, we review and asked them about what it would take them to shop relevant research on human-robot interaction (HRI). In Section 3, independently. To minimize peer pressure, we met with each we discuss the basic principles of ergonomics-for-one and present individual separately. an ergonomics-for-one analysis to identify the key elements of the performance gap between the blind individual and the task of Essential task functions: The interviews helped us identify five independent grocery shopping. In Section 4, we present our initial essential task functions: 1) getting to a supermarket; 2) finding the design aimed at bridging several elements of the performance gap. needed grocery items; 3) getting through a cash register; 4) In Section 5, we present our navigation trials with a small sample leaving the store; and 5) getting home. None of the individuals of visually impaired participants. We describe our experiment had any problems getting to a supermarket and getting home from design, analyze the collected data, and present the participants' a supermarket. Logan has a free bus system with a network of bus feedback. In Section 6, we return to bridging the performance stops all over the city and several suburbs. All places where one gap. We describe how we deployed the robotic shopping cart in a can buy groceries have bus stops close by. supermarket and discuss what ergonomic modifications we made in the system after several fitting trials in that environment. In Function 2 was refined into three sub-functions: 1) navigating to a Section 7, we focus on two major challenges for the future. In shelf section with a needed grocery item; 2) finding the needed Section 8, we give our conclusions. item on the shelf; and 3) placing the item in a shopping basket.
  • 3. Function 3 was refined into four sub-functions: 1) navigating to a next to other Lays items or other potato chip brands. This is even cash register; 2) placing the items from the basket on the belt; 3) more of a problem with smaller items, like small bags of paying for the items; and 4) placing the bagged items back into sunflower seeds. Assuming that the individual does not have the shopping basket. Function 4 was refined into three sub- manual dexterity problems, once the item is found, the individual functions: 1) getting to the exit; 2) leaving the basket in a can place it into a shopping basket. designated place; and 3) exiting the store. Fifth, the same navigational challenges apply to the function of Equipment: None of the participants did any grocery shopping getting through a cash register. Additional challenges are knowing on their own. They either did not do any grocery shopping or used when it is time to start placing the items on the conveyor belt, sighted guides: parents, siblings, or partners. The only equipment paying for the items, and putting the bagged items back into the used by the participants were white canes, guide dogs, and basket. shopping baskets. They could not use shopping carts, because Sixth, when the shopper is ready to leave the store, she again has they could not simultaneously handle guide dogs or white canes to navigate to the exit, thus confronting the navigational and push the carts. challenges identified above, and place the basket in the properly Inputs and outputs: When asked how they would prefer to designated place. interact with an assistive grocery shopping device, if they had one, the participants suggested speech and keypad as input 4. BRIDGING THE GAP: PART I options and speech and dynamic Braille as output options. 4.1 On to a Robotic Shopping Cart Environment: The target environment was a typical supermarket. After considering the first performance gap component, There are several features that make this environment particularly independent use of a shopping cart, we concluded that the challenging. First, there is always some shopper traffic. On navigation performance of the shopping cart had to be automated. certain days, e.g. Saturday, and during certain hours, e.g., Effectively, the robotic shopping cart would act as a supermarket between 6 and 8 pm, the shopper traffic is at its highest. Second, guide for blind shoppers. This is, by no means, a novel idea as the there are indigenous processes already in place, e.g., shelf re- field of AI robotics had built robotic guides before [3]. None of stocking, cleaning, product scanning, etc., that cannot be the guides, however, were specifically built for blind shoppers in disrupted. Third, the products are periodically re-shuffled and re- supermarkets. As far as we could see, we had two options: arranged, and free open spaces are occupied with temporary building a new robotic base with a shopping cart mounted on top displays and stands. of it or mounting a shopping cart on top of an existing robotic Individual: Two participants were completely blind. Three base. We chose the second option, because the first option, after a participants had light perception, i.e., they were able to preliminary cost analysis, looked prohibitively expensive for a distinguish between light and dark. All participants were research prototype. In addition, we already had experience ambulatory, did not have any serious speech impediments, mounting equipment on our Pioneer 2DX robotic base from the hearing problems, or cognitive disabilities. ActivMedia Corporation when we experimented with our robot- assisted navigation for the blind in indoor environments [6, 7]. After the interviews, we identified the performance gap that had to be addressed by the accommodation system. First, using a guide dog and/or a white cane with a shopping cart is not feasible. Neither guide dogs nor white canes would help avoid front obstacles if the blind shopper has to push the cart in addition to handling a guide dog or using a white cane. Of course, it is possible to use a basket, but the shopper would then be restricted to buying a small number of items. Second, since the shopper cannot independently navigate, she needs to communicate her intentions to a sighted guide. This would be ordinarily done in natural language if the guide is human. Third, the visually impaired participant needs assistance with Figure 1: RoboCart in Lee's MarketPlace. navigating to shelf sections with specific grocery items. In an environment where end points of routes remain static, many guide Thus, we built a polyvinyl chloride (PVC) pipe structure, securely dog handlers and cane users can learn routes after several trials. mounted it on top of the Pioneer 2DX robotic base, and then However, this assumption does not hold in supermarkets due to placed a large shopping basket into that structure. The resulting constant re-arrangements and re-shufflings of products. design, which we called RoboCart, is shown in Figure 1. As one can easily see from Figure 2, the RoboCart design is a Fourth, even if it is assumed that the blind shopper can find her modification of RG, our indoor robotic guide for the blind that we way to the correct shelf section, she still needs to pick the right built in 2003-2004 on top of another Pioneer 2DX base. It should item. For example, suppose that the blind shopper wants to buy a be noted that this is a proof-of-concept design. The back bag of Lays Classic and finds her way to the correct shelf section directional wheel of the base is small, which results in the with Lays potato chips. There is always a chance that the shopper inherent imbalance of this design. While we have not observed will pick a wrong bag as Lays Classic bags are typically placed any accidents in which RoboCart tipped over, the future design
  • 4. will be modified to have a four-wheel base so that the device will below, there is another reason why Braille may not be a viable never tip over and injure the blind shopper. option for some users. 4.3 How Do We Navigate? When we started thinking about bridging the navigational component of the performance gap, we realized that we had little knowledge about what aspects of navigation might be important to the blind navigator. We also did not know if our communication choices described in the previous section would be ergonomically acceptable to blind individuals. Finally, we wanted to find out whether the presence or absence of the human navigator behind the robot affects the robot's navigation. To answer these questions, we decided to conduct a series of fitting trials. We had to find a suitable environment for the trials. We had started negotiations with Lee's MarketPlace, a Figure 2: RG, an indoor robotic guide for the blind. supermarket in Logan, Utah, about the possibility of testing RoboCart in their supermarket. But the negotiations were still in 4.2 How Do We Communicate? progress. We ruled out tests in our CS Department, because we Upon entering the supermarket, the shopper needs to had already tested our robotic guide in the CS Department rather communicate her wishes to RoboCart. The input options that we extensively and had achieved satisfactory results [6, 7]. considered were automatic speech recognition (ASR) and keypad. We chose to conduct fitting trials at the USU Center for Persons When using ASR, the blind shopper would wear a wireless with Disabilities (CPD). The CPD occupies an entire building on microphone coupled to an over-the-ear headphone and the North USU Campus. The building has an area of 40,000 communicate her intentions to the robot through speech. We will square feet. It has numerous offices, classrooms, laboratories, not go into details here on why we ruled out speech as an input lounges, and bathrooms. Another challenging aspect of this option, because we have described our reasons in detail in our environment that makes it similar to a supermarket is numerous previous publications [6]. In brief, our ASR experiments, both in activities occur there during its working hours. Thus, other people noisy and noise-free environments, had recognition rates of below going about their business, i.e., human traffic, are an integral part 50 percent even though all of our participants were native of the environment. speakers of American English. Our decision to rule out ASR as an input option should not be construed as a general argument against ASR as an HRI mode. Rather, we concluded that, given 5. FITTING TRIALS the current state of the art in commercial ASR and the constraints of our problem, we should explore the keypad first. 5.1 Experiment Design We used the paired differences strategy to design our pilot The input option that we chose was a small 10-key Belkin experiments. In a paired difference experiment, one is interested numeric keypad. The layout of keys on the keypad is the same as to find the mean difference between two methods of conducting the layout of keys on the cell phone. Since many visually some activity, which, in our case, is navigation. A data point is impaired people use cell phones, our thinking was that the obtained by numerically measuring the performances of two learning curve would not be steep. In addition, the number 5 key participants, say X and Y, from two different samples doing a on Belkin keypads has a small plastic protrusion that the visually designated activity and computing the difference between the two impaired can sense through touch. Once the number 5 key is measurements. When a sample of differences is obtained, one can found, it is easy to find the other keys. test two hypotheses: the null hypothesis, H 0 : µ D = 0 , against one of the three alternative hypotheses, H a : µ D > 0 , When compared to ASR, the keypad does reduce input ambiguity. However, even with the keypad the proverbial problem of shared vocabulary does not go away. The user still must know what to H a : µ D < 0 , and H a : µ D ≠ 0 , where µ D is the mean type into the robot to make the robot do what the user wants. To overcome this problem, we decided to create a Braille directory. difference. Essentially, H 0 suggests that there is no difference in The directory was to be realized as a Braille sheet with performance, whereas H a ’s suggest that there may be a instructions that map each destination to a short sequence of numbers. The semantics of each line was to be as follows: if you difference. The test statistic is a one-sample want to go to destination X, please type this numerical sequence into the keypad. t = x D / s D / n D , where x D is the sample mean The next element of the communication gap is output. The difference, s D is the standard deviation of the differences, and options that we considered were synthetic speech and dynamic n D is the number of differences. Braille displays. As we investigated dynamic Braille displays, we found out that they were expensive: the cheapest option we could We selected a total of 9 routes in the environment. Each route was find was approximately 5K USD. Originally, the cost was the more than 40 meters in length and had 3 to 5 turns. In our case, main reason why we decided on synthetic speech. As we discuss our first sample consisted of the robot. Since we focused on
  • 5. 10 11 12 13 14 15 16 17 18 3.64 3.60 -1.96 0.90 -4.07 -4.14 3.24 3.91 1.90 Table 1: T-statistics at α = 0.05 and df=4. Route,Part 0 1 2 3 4 5 10 65.79, 65.84 59.87, 60.89 62.90, 63.45 61.76, 65.08 61.94, 63.75 65.49, 67.14 11 70.83, 72.27 55.93, 57.33 56.91, 59.27 55.45, 59.16 56.42, 59.22 72.29, 73.67 12 70.94, 72.25 72.56, 73.68 75.78, 79.43 71.79, 98.53 73.96, 75.47 69.96, 72.46 13 87.88, 89.93 87.06, 87.93 89.45, 91.03 86.29, 88.78 86.29, 88.55 87.70, 90.17 14 55.76, 56.29 82.21, 83.71 84.60, 86.46 83.22, 84.89 83.61, 84.86 55.81, 57.12 15 57.35, 60.30 79.23, 80.15 78.85, 81.29 79.88, 81.85 85.91, 88.43 56.11, 64.49 16 120.74, 123.34 93.11, 97.67 95.54, 102.48 90.91, 93.09 98.70, 101.48 122.87, 129.00 17 124.72, 123.34 83.93, 103.48 87.10, 103.16 91.17, 94.08 90.25, 92.14 125.10, 126.82 18 129.11, 130.61 130.89, 139.79 97.35, 100.58 84.46, 86.63 88.14, 92.67 130.89, 139.77 Table 2: 95% confidence intervals. navigation and guidance, we used the robotic guide shown in own and the robot navigating with a visually impaired human. On Figure 2. Our second sample consisted of five visually impaired the other routes, i.e., 12, 13, and 18, there appears to be participants. To obtain the measurements, we ran the robot five insufficient evidence to reject H 0 . In other words, the presence times on each of the designated routes and recorded the time-to- completion, e.g., the amount of time it took the robot to complete of the human navigator behind the robot does not appear to affect the route. For each route, the average time-to-completion was the robot's performance. Since, in computing µ D , we subtracted computed from the five runs. the robot's time-to-completion from a participant's time-to- We then had each participant use the robot to navigate the same completion, the positive t-statistics that exceed 2.776 suggest that routes. The robot would inform the participant through synthetic the robot was slower without the navigator than with the speech about its present location. We told each participant the navigator. On the contrary, the negative t-statistics smaller than keypad codes for all destinations. The participant would type in 2.776 suggest that the robot was slower with the navigator than by the destination code through the keypad attached to a pole on the itself. back of the robot. Each route was navigated five times and the To verify the validity of these observations, we analyzed the data time-to-completion measurements were taken for each participant. through confidence intervals. We computed 95% time-to- For each participant we computed the average time-to- completion confidence intervals for each route and each completion. A sample of differences that we used to test the participant, including the robot. Table 2 gives the confidence hypothesis was obtained by computing the difference between the intervals for all routes and participants. The robot is listed as robot's average times-to-completion and the participants' times-to- participant 0. The interval table verifies the conclusions of the completion. hypothesis tests. For example, both ends of the robot's confidence We chose to test the third alternative hypothesis, interval for route 10 given in column 0 are greater than the H a : µ D ≠ 0 , at α = 0.05 corresponding ends of participants 1 through 4 and are essentially as the level of significance. The the same as the ends of participant 5. The same observations can rejection region for this hypothesis is be made on routes 11, 16, and 17. This seems to verify the test of | t |> tα / 2 = t 0.025 = 2.776 , and has 4 degrees of freedom. hypothesis conclusion that on these routes the robot without the navigator appeared to be slower than the robot with the navigator. Table 1 contains the sample t- statistics for each of the 9 routes The same technique can be applied to routes 14 and 15 on which, numbered 10 through 18. These statistics should not be viewed as according to the test of hypothesis conclusion, the robot appeared definitive. The paired differences design requires that the sample to be faster without the navigator than with the navigator. The of differences be random. This assumption may not be satisfied in robot's confidence intervals for these routes are to the left of the our case, because we did not choose the five individuals confidence intervals of participants 1 through 4 and coincide with randomly. Their names were given to us by referral. the ends of participant 5. To understand what was causing these differences, we looked at 5.2 Results the video footage of the runs. The video footage of the robot runs The results in Table 1 tell us that on routes 10, 11, 14, 15, 16, and without the navigator on routes 10, 11, 16, and 17 showed that 17, H0 is rejected, because the absolute value of the t-statistic is there was quite a bit of human traffic in the hallways. The video larger than 2.776. In other words, on these routes there appears to footage of the robot runs with the navigator on the same routes be a significant difference between the robot navigating on its showed that in the cases of participants 1 through 4, the amount of
  • 6. human traffic in the hallways declined. The exception was was on the right or left. It also told me when it was turning left or participant 5 for whom the amount of human traffic remained right. I would appreciate voice messages being spoken more essentially the same. Since the robot's speed decreases with the loudly. I understand that you cannot make it too loud without number of obstacles present in front, the robot traveled more making it obnoxious to the people around me. Perhaps, it could slowly in the presence of human traffic. be done with one over-ear headphone or a shoulder speaker so that I have my other ear available to me. The situation was reversed on routes 14 and 15. During the robot runs without the navigator the amount of human traffic was Comment 6: Overall, I felt very comfortable navigating with the minimal. However, when we ran the robot with the human robot. I felt even more comfortable after I learned on one of the navigators, human traffic picked up considerably. The exception runs that the robot can recover from situations when it gets lost again was participant 5 for whom the amount of human traffic did by finding an alternate route. Self-correction is a valuable feature not change. Our conclusion was that the amount of human traffic, of this device. i.e., the number of people on route, is a nuisance variable that may have contributed to the differences in robot performance. Comment 7: Make sure that there is no chance of the robot going off the course. Another interesting observation that we made as we watched the video footage was the effect of the occasional mismatch between the verbalized intent of the robot and the robot's actual actions. At several T-intersections the robot would tell the navigator that it was turning left and then, due to the presence of people, it started drifting to the right before actually making a left turn. When that happened, we observed that several human navigators pulled hard on the robot's handle, sometimes driving the robot to a virtual halt. We conjecture that when a communication mismatch occurs, i.e., when the robot starts doing something other than what it said it would do, the human navigators become apprehensive and try to stop the robot. Since these mismatches happened on the routes where the robot performed better without the navigator than with the navigator, we concluded that the mismatches may have contributed to the performance difference. Figure 3:RoboCart's Handle, Design 1. While watching the video footage, we also observed a different kind of communication problem that occurred several times during u-turns. The robot would inform the navigator that it had started making a u-turn after it had already started executing the maneuver. Although the robot's message was accurate, it came a bit too late and, as discussed in the next section, caused some 6. BRIDGING THE GAP: PART II In fall 2004, we received permission from Lee's MarketPlace to discomfort on the part of the participants. use their supermarket as a test site for our experiments. We asked two visually impaired individuals to participate in a series of 5.3 Participants Speak fitting trials in the store. On several occasions we ran RoboCart After the experiments, we conducted informal verbal interviews on its own. The objective was to learn through deployment what with the participants and recorded their responses. The interviews modifications in ergonomic design and navigation were required. consisted of several questions about navigation safety and user comfort. The objective was to let the participants give us feedback on their experiences. Below we give several comments verbatim. 6.1 Ergonomic Modifications As shown in Figure 2, our original design included a guide leash. Comment 1: There was some abruptness in the robot motion. However, the participants expressed a wish that the dog leash be Stops and slows down too suddenly. Sometimes it accelerates too replaced with a static handle. When asked why, the participants fast. said that the dog leash did not give them sufficient feedback as to what direction the robot was taking them. This wish was Comment 2: Sometimes the robot tells you too late when it is expressed both by the cane users and guide dog handlers. It was about to make a u-turn. This is a problem if you have a guide dog quite understandable that cane users expressed this wish because and need to tell him to get out of the robot's way. the cane is firm and does resemble a static handle. We were Comment 3: A little more user training up front would help. Let surprised, however, to hear the same complaint from the guide me touch the robot and give me some time to get comfortable with dog handlers. As we took a closer look at how the guide dogs are the keypad. handled, the explanation presented itself immediately. It turns out that guide dog handlers do not use the leash when their dogs are at Comment 4: The robot slows down at turns and then it kicks into work. They use a firm leather handle attached to a special harness high gear too abruptly. I have a back injury and so such changes on the back of the dog. The handle enables the handler to give in speed were felt a lot. directions to the animal as well as to receive immediate haptic feedback about the animal's movement. The leash is used only Comment 5: The communication was clear and helpful. The when the dog is not at work and is being treated as a pet. robot told me when I got to a destination, whether the destination
  • 7. The above lesson led to our first modification - the addition of a boxes and similar movable objects being placed into them by the static handle shown in Figure 3. The keypad hangs on the right store staff. After investigating the possibility of using Markov pole of the handle. After several trials in Lee's MarketPlace, we localization [3], we decided against it because of safety concerns. realized that the keypad's position was inconvenient for the user. Most applications of Markov localization indoors are based on It is difficult to access the keypad quickly when the robot is laser range finding. Laser range finding does not perform well in moving. To reach for the keypad requires letting go of the handle. large open spaces or environments with large glassy surfaces that Using the other hand is impossible as it is occupied with a cane or absorb laser signals. The performance of Markov localization is a leash. not predictable in dynamic environments and degrades in the presence of numerous dynamic obstacles. We considered extending our RFID-based navigation to open spaces by putting portable towers with RFID tags. We rejected this idea, too, because it called for a great deal of calibration and instrumentation and could be too disruptive to the indigenous business processes. We discussed our problem with the supermarket's owner and a senior store manager. They suggested that we put masking tape lines on the floor and use them for navigating large open spaces. In their opinion, if the system were to be deployed in their store permanently, they could easily paint such lines on the floor. As long as the paint was resistant to the floor wax, the lines were not a problem. Figure 4: RoboCart's Handle, Design 2. Figure 4 shows how we modified this design by changing the position of the keypad. We purchased the wireless version of the same keypad, attached it to a small plastic rectangle, and then attached the rectangle to the handle's bar. This position allows the navigator to quickly reach for the keypad during the navigation without letting go of the handle. We also learned that Braille may not be feasible. Of the seven visually impaired people that we informally polled about the possibility of using Braille on the robot only 2 were comfortable with the idea. As we investigated the matter further and talked Figure 6: RoboCart following a line. with the assistive technology specialists at the USU Center for RoboCart was equipped with a small LogiTech web camera. Persons with Disabilities, we learned that only a small fraction of Figure 5 shows how the camera was added to the robotic base. visually impaired people use Braille. This fraction consists mostly We put one masking tape line from the lobby and up to the aisles. of people who are blind from birth. People who lose vision later A simple vision-based line following algorithm was written and in their lives due to accident, illness, or age either never learn successfully tested on several runs. Figure 6 shows how RoboCart Braille or use it rather slowly. follows the line to reach an aisle. Once in the aisles, our original RFID-based navigation algorithm was used. One aisle has 5 shelves on both sides. An RFID tag is placed every 3 meters on the 2nd or 3rd shelf on both sides of the aisle so that the robot's RFID antenna can detect it. Thus, every aisle in which we tested RoboCart is equipped with 10 RFID tags: 5 on the left side and 5 on the right side. There is also a designated cash register where RoboCart takes the blind shopper. The cash register is equipped with two RFID tags. The first tag makes RoboCart stop and inform the blind shopper that the products can be unloaded onto the belt on the right. The second tag informs the shopper that she has to wait for the bagger to put the bags into the cart. The store management was comfortable with this instrumentation plan. Figure 5: RoboCart's Camera. 6.2 Navigation Modifications 7. A Glimpse of the Future Several important modifications were made to our navigation When we learned that Braille may not be a viable option, we algorithm. The original algorithm was designed for structured replaced Braille with a voice-based directory based on synthetic indoor environments [7], which was fine for navigating speech. Instead of reading Braille, a blind person uses the keypad supermarket aisles. The algorithm did not work in large open to scroll up and down the voice menu in which each line is spoken spaces, such as supermarket lobbies. Besides having a lot of to the user by the speech synthesis software. Modern grocery customer traffic, supermarket lobbies constantly change in terms stores carry thousands of items. One challenge that we are of their layout due to promotion displays, flower stands, product
  • 8. currently investigating is how to organize the directory for easy 10. REFERENCES browsing. [1] Fong, T., Kaber, D., Lewis, M., Scholtz, J., Schultz, A., and Since the RFID tags must be placed on both sides of each aisle, Steinfield, A. Common Metrics for Human-Robot the approximate layout of the store must be known in advance. Interaction. In citeseer.csail.mit.edu/667916.html. This layout must be maintained by the store. Thus, another [2] Fong, T., Nourbakhsh, I., and Dautenhahn, K. A Survey of challenge is how the store maintains the layout so that the robotic Socially Interactive Robots. Robotics and Autonomous shopping cart always guides the blind shopper to the appropriate Systems, 42:143-166, 2003. shelf. [3] Fox, D., Burgard, W., and Thrun, S. Markov Localization for Another ergonomic challenge is access to individual items. Mobile Robots in Dynamic Environments. Journal of AI RoboCart leads the blind person to shelf sections, not to Research, 11:391-427, 1999. individual items. For example, it will guide the person to the shelf [4] Goodrich, M. and Olsen, D. Seven Principles of Efficient section with Lays potato chips, but the person still has to pick up Human-Robot Interaction. In Proceedings of the IEEE an individual bag and put it into the robot's shopping basket. To International Conference on Systems, Man, and Cybernetics, address this problem, we have integrated a small portable barcode pp. 3943-3948. IEEE, October 2003. reader into the system. Grocery stores already use barcode reading technologies to keep track of their price inventories. The [5] Howard, A. A Methodology to Assess Performance of scenario that we are currently experimenting with is as follows: Human-Robotic Systems in Achievement of Collective RoboCart gets the blind shopper to a shelf section with a bunch of Tasks. In Proceedings of the International Conference on individual items, the shopper then uses a handheld barcode reader Intelligent Robots and Systems (IROS). IEEE/RSJ, July 2005. to read the barcodes on the shelf until the barcode of the right [6] Kulyukin, V., Gharpure, C., De Graw, N., Nicholson, J., and item is found. Under this scenario, the shopper has to find the Pavithran, S. A Robotic Wayfinding System for the Visually shelf and then slide the barcode reader along the shelf and listen Impaired. In Proceedings of the Innovative Applications of until a speech message tells the user that the proper barcode is Artificial Intelligence Conference (IAAI), pp. 864-869. read. AAAI, July 2004. [7] Kulyukin, V., Gharpure, C., Nicholson, J., and S. Pavithran. 8. CONCLUSIONS RFID in Robot-Assisted Indoor Navigation for the Visually In this paper we showed how the basic principles of ergonomics- Impaired. In Proceedings of the IEEE International for-one were applied to the design and development of a proof-of- Conference on Intelligent Robots and Systems (IROS). concept robotic shopping cart for the blind. We identified the IEEE/RSJ, October 2004. performance gap that must be overcome by an accommodation [8] LaPlante, M. and Carson, D. Disability in the United States: system that allows the blind to shop independently. We described Prevalence and Causes. U.S. Department of Education, our initial usability tests and showed how the tests shaped the Washington, DC, 2000. ergonomic modifications of the system. [9] McQuistion, L. Rehabilitation Engineering: Ergonomics for One. Ergonomics in Design, January:9-10, 1993. 9. ACKNOWLEDGMENTS The first author would like to acknowledge that this research has [10] Olsen, D and Goodrich, M. Metrics for Evaluating Human- been supported, in part, through NSF CAREER grant (IIS- Robot Interactions. In Performance Metrics for Intelligent 0346880) and two Community University Research Initiative Systems (PERMIS). NIST, September, 2003. (CURI) grants (CURI-04 and CURI-05) from the State of Utah. [11] Pollack, M. Intelligent Technology for the Aging Population. We would like to thank Mr. Lee Badger, the owner of Lee's AI Magazine, 26(2):9-24, 2005. MarketPlace, for allowing us to use his supermarket in Logan, [12] Scerri, P., Pynadath, D., and Tambe, M. Toward Adjustable Utah, as a research site. We are grateful to John Nicholson, our Autonomy for the Real World. Journal of AI Research, research colleague at USU CSATL, for helping us to conduct 17:171-228, 2002. many fitting trials. We would like to thank Ying Bing, a CS graduate student, for implementing the line following algorithm. [13] E. Berg. Ergonomics in Health Care and Rehabilitation. Finally, we would like to thank the visually impaired participants Butterworth-Heinemann, Woburn, MA, 1998. in our experiments for their valuable feedback. [14] Yanco, H. and Drury, J. A Taxonomy for Human-Robot Interaction. In Proceedings of the AAAI Fall Symposium on Human-Robot Interaction, pp. 111-119, 2002.