CELESTIAL NAVIGATION SYSTEM FOR AN AUTONOMOUS VEHICLE
20210011485 ยท 2021-01-14
Inventors
Cpc classification
G05D1/0225
PHYSICS
G06F17/00
PHYSICS
H04W88/06
ELECTRICITY
G05D1/0088
PHYSICS
A47L2201/06
HUMAN NECESSITIES
B60L53/38
PERFORMING OPERATIONS; TRANSPORTING
A47L9/2836
HUMAN NECESSITIES
Y02T10/70
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
Y10S901/01
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
A47L9/2894
HUMAN NECESSITIES
Y02T90/12
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
G05D1/0276
PHYSICS
A47L2201/04
HUMAN NECESSITIES
Y02T10/7072
GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
International classification
A47L9/28
HUMAN NECESSITIES
G06F17/00
PHYSICS
B60L53/38
PERFORMING OPERATIONS; TRANSPORTING
Abstract
A navigation control system for an autonomous vehicle comprises a transmitter and an autonomous vehicle. The transmitter comprises an emitter for emitting at least one signal, a power source for powering the emitter, a device for capturing wireless energy to charge the power source, and a printed circuit board for converting the captured wireless energy to a form for charging the power source. The autonomous vehicle operates within a working area and comprises a receiver for detecting the at least one signal emitted by the emitter, and a processor for determining a relative location of the autonomous vehicle within the working area based on the signal emitted by the emitter.
Claims
1-20. (canceled)
21. An autonomous robotic cleaning device comprising: a robot body; a drive supporting the robot body above a floor surface within a room and configured to maneuver the robot body across the floor surface; a cleaning apparatus to clean the floor surface; one or more processors configured to: navigate the autonomous robotic cleaning device based on a map, the map including an indication of areas with carpeting and areas that do not have carpeting, sense the location of the autonomous robotic cleaning device relative to the indicated areas; and adjust a cleaning behavior of the autonomous robotic cleaning device based on the sensed location relative to the marked areas; transmit data to cause a user display device to display a floor plan that includes a plurality of rooms, the displayed floor plan including an identification of the areas with carpeting and areas that do not have carpeting and an indication of doors separating multiple rooms.
22. The autonomous robotic cleaning device of claim 21, wherein the configurations to adjust cleaning behavior based on the sensed location comprise configurations to avoid areas with carpeting.
23. The autonomous robotic cleaning device of claim 21, wherein the autonomous robotic cleaning device is configured to transmit data indicating portions of the floor surface that the autonomous robotic cleaning device has traversed and portions of the floor surface that the autonomous robotic cleaning device has not traversed.
24. The autonomous robotic cleaning device of claim 21, further comprising a sensor to detect obstacles within the room, wherein the one or more processors are configured to: indicate, on the map, an entrapment area based on an output from the sensor while navigating the autonomous robotic cleaning device during a first cleaning run, and then control the autonomous robotic cleaning device using the map to avoid the entrapment area during a second cleaning run.
25. The autonomous robotic cleaning device of claim 21, wherein the cleaning apparatus is configured for floor scrubbing, sweeping and vacuuming.
26. The autonomous robotic cleaning device of claim 21, further comprising an upward-angled camera positioned on a top of the robot body and directed at least partially away from a ceiling of the room to capture visible points on wall surfaces within the room.
27. The autonomous robotic cleaning device of claim 21, wherein the one or more processors are configured to assign a name to the room and another name to another room based on signals indicative of user registrations of the name and the other name, determine an order to clean the room and the other room based on signals indicative of a user selection of the name and the other name, and control the autonomous robotic cleaning device to clean the room and the other room in accordance to the order.
28. An method comprising: navigating an autonomous robotic cleaning device based on a map, the map including an indication of areas with carpeting and areas that do not have carpeting; sensing the location of the autonomous robotic cleaning device relative to the indicated areas; and adjusting a cleaning behavior of the autonomous robotic cleaning device based on the sensed location relative to the marked areas; transmitting data to cause a user display device to display a floor plan that includes a plurality of rooms, the displayed floor plan including an identification of the areas with carpeting and areas that do not have carpeting and an indication of doors separating multiple rooms.
29. The method of claim 28, wherein adjusting the cleaning behavior based on the sensed location comprises adjusting the cleaning behavior of the autonomous robotic cleaning device to avoid areas with carpeting.
30. The method of claim 28, further comprising transmitting data indicating portions of the floor surface that the autonomous robotic cleaning device has traversed and portions of the floor surface that the autonomous robotic cleaning device has not traversed
31. The method of claim 28, further comprising: indicating, on the map, an entrapment area based on receiving an output from an obstacle detection sensor while navigating the autonomous robotic cleaning device during a first cleaning run, and then controlling the autonomous robotic cleaning device using the map to avoid the entrapment area during a second cleaning run.
32. The method of claim 28, wherein the autonomous robotic cleaning device comprises a cleaning apparatus configured for floor scrubbing, sweeping and vacuuming.
33. The method of claim 28, further comprising capturing visible points on wall surfaces within the room using an upward-angled camera positioned on a top of a robot body of the autonomous robotic cleaning device and directed at least partially away from a ceiling of the room.
34. The method of claim 28, further comprising: assigning a name to the room and another name to another room based on signals indicative of user registrations of the names, and determining an order to clean the room and the other room based on signals indicative of a user selection of the name and the other name, and controlling the autonomous robotic cleaning device to clean the room and the other room in accordance to the order.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
[0026]
[0027]
[0028]
[0029]
[0030]
[0031]
[0032]
[0033]
DESCRIPTION OF THE PRESENT TEACHINGS
[0034] Reference will now be made in detail to embodiments of the present teachings, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or like parts.
[0035] In accordance with an exemplary implementation of the present teachings,
[0036] For simplicity, this disclosure will describe vacuuming as a demonstrative task of the depicted robotic cleaning device 12. It will be apparent, though, that the navigation system disclosed herein has wide applications across a variety of autonomous systems. For example, an autonomous vehicle may be used for floor waxing and polishing, floor scrubbing, ice resurfacing, sweeping and vacuuming, unfinished floor sanding, stain/paint application, ice melting and snow removal, grass cutting, etc. Any number of task-specific components may be required for such duties, and may each be incorporated into the autonomous vehicle, as necessary.
[0037] The transmitter 20 directs at least two infrared signals 22a, 24a from emitters 24A and 24B to a surface remote from the working area 14 upon which the autonomous vehicle 12 operates. The depicted embodiment directs the infrared signals 22a, 24a to the ceiling 18, but it may also direct the signals 22a, 24a to a portion of a wall 16 or to both the walls 16 and ceiling 18. The signals 22a, 24a can be directed to a variety of points on the remote surface, but directing the signals as high as possible above the working area 14 can allow the signals 22a, 24a to be more easily detected by the autonomous vehicle 12, because the field of view of the autonomous vehicle's receiver 28 is less likely to be blocked by an obstacle (such as, for example, a high-backed chair or tall plant). In this disclosure, the regions of contact 22b, 24b of the signals 22a, 24a on the remote surface will be referred to aspoints, regardless of the size of the intersection. For example, by using a collimator in conjunction with the emitters (described below), the points of intersection 22b, 24b of the signals 22a, 24a can be a finite area with the signal strongest at approximately central points.
[0038] In certain embodiments of the transmitter 20, the signals 22a, 24a are directed toward a ceiling 18, at two points 22c, 24c, forming a line proximate and parallel to the wall 16 upon which the transmitter 20 is located. Alternatively, and as depicted in
[0039] As the autonomous vehicle 12 moves within a working area 14, it detects the signals 22a, 24a emitted by the transmitter 20 as energy bouncing or reflecting off of the diffuse ceiling surface 18. In an alternative embodiment, visible points can be used in place of infrared points. A camera onboard the autonomous vehicle can replace the infrared receiver in detecting either infrared or visible points. The autonomous vehicle's microprocessor can convert the signals 22a, 24a sensed by the receiver 28 into bearings from the robot 12 to the signals 22a, 24a. The microprocessor can then calculate representative elevation angles .sub.1, .sub.2 and azimuths .sub.1, .sub.2 of the signals to determine the location of the autonomous vehicle 12 within the working area 14. In this embodiment, the azimuths .sub.1, .sub.2 are measured using a forward direction of movement M of the autonomous vehicle 12 as a datum, but any suitable datum can be used. By calculating the elevation angle and azimuth from the autonomous vehicle 12 to the two signals 22a, 24a, the autonomous vehicle 12 can locate itself within a working area with improved accuracy.
[0040]
[0041] The transmitter 120 emits two signals 122a, 124a (depicted graphically by a plurality of arrows) Into the two rooms 136, 138, respectively. The signals 122a, 124a can be configured to not overlap each other, thus providing a distinct signal on each side of the door centerline 130. In other embodiments, an overlap of the signals 122a, 124a can be desirable. The autonomous vehicle 112 includes a receiver 128 having a field of vision 134. The emitted signals 1226, 124a can be detected by the receiver 128 when the autonomous vehicle's field of vision 134 intersects the signals 122a, 124a. Similar to the embodiment of
[0042]
[0043] In accordance with various embodiments of the present teachings, the transmitter can include a visible signal option (not shown), aligned with the emitted signals, allowing a user to direct the signals to particular locations. In accordance with the present teachings, more than one transmitter maybe used. Such a system could include communication capability between the various transmitters, for example to ensure that only one signal or a subset of signals is emitted at any given time.
[0044] A battery-powered transmitter located above a window or door frame can not only permit the autonomous vehicle to localize within a map, coordinate system, or cell grid relative to the transmitter, but can also localize the transmitter within the same map, coordinate system, or cell grid, thereby localizing the window or door frame. Localization of an autonomous vehicle within a working environment is described in detail in U.S. Patent Publication No. 2008/0294288, filed Nov. 27, 2008, the entire disclosure of which is incorporated herein by reference. In the case of a door frame, the door is ordinarily the passage by which the autonomous vehicle navigates from room to room. The transmitter illustrated in
[0045] The exemplary embodiment of a transmitter 20 illustrated in
[0046]
[0047] In accordance with various embodiments of the present teachings, more than two emitters can be utilized with collimators 22e, 24e, 122e, 124e, to distinguish different areas within a room. Such a configuration allows the autonomous vehicle to sense its relative location within a room and adjust its cleaning behavior accordingly. For example, a signal could mark an area of the room that an autonomous vehicle would likely get stuck in. The signal could allow an autonomous vehicle to recognize the area and accordingly not enter it, even though it might otherwise be able to do so unimpeded. Alternatively, or in addition, different signals could mark areas that require different cleaning behaviors (e.g., due to carpeting or wood floors, high traffic areas, etc.).
[0048] Turning back to
[0049] In various embodiments of the present teachings, each signal (regardless of the emitter's location or the number of signals) can be modulated at 10 kHz and coded with an 8-bit code serving as a unique signal identifier, preventing the autonomous vehicle from confusing one signal or point with another. Accordingly, more than two uniquely encoded signals can be employed to increase the accuracy of the autonomous vehicle's calculations regarding its location within a working area. As noted above, using only one emitter allows an autonomous vehicle to take a heading based on that signal. Using two or more signals can allow the autonomous vehicle to continue navigating if fewer than all of the signals are detected (either due to a failure of a signal transmission or the autonomous vehicle moving to a location where fewer than all of the signals are visible).
[0050] In certain embodiments, the transmitter can pulse the coded signals as follows. After an initial synchronization pulse, a first signal at 10 kHz is emitted for 100 ms. This can provide a sufficient time for the autonomous vehicle's receiver and processor to calculate azimuth and elevation angles, as discussed in detail below. So that the autonomous vehicle can determine which signal is being received, the transmitter can pulse a series of five bits, each for 10 ms. The five bits include two start bits, for example a zero and a one, followed by a unique three bit identifier to identify that particular signal or point. After a 100 ms delay, the transmitter repeats the sequence for the second signal or point. By changing the modulation frequency and/or the identifier, the second signal or point can be uniquely distinguished from the first. Any number of unique signals can be transmitted and identified in this manner. After the series of signals are transmitted, the transmitter can wait a substantially longer period of time, for example on the order of one to two seconds, before repeating the transmitting sequence, starting again with the first signal. The length of time for each transmission is merely exemplary, and may be varied based on a particular application, device, etc. As stated above, the signals can be modulated at the same or different frequencies.
[0051]
[0052] As illustrated in
[0053]
[0054] The construction of the receiver 328 can be similar to that of
[0055]
[0056] In operation, a receiver (e.g., an infrared receiver) can first measure the noise floor of the autonomous vehicle's environment, comprising the amount of energy (e.g., infrared energy) present in the autonomous vehicle's environment, which it sets as the threshold value. This value can represent an average of the values for each photodiode. Any subsequent measurement above this threshold value can trigger an event (e.g., a calculation of point azimuth and elevation). The receiver can then measure the modulation frequency again, searching for an expected increase at 10 kHz (i.e., the frequency of the initial synchronization signal transmitted by the transmitter). If a 10 kHz frequency increase is detected, the autonomous vehicle recognizes the increase as an emitted navigation signal. The autonomous vehicle can then measure the amplitude of the reflected point on all five photodiodes to determine an average value. This value can then be compared to a list of signal frequencies to determine which of the signals has been detected. Alternatively, any detected identity sequence associated with the signal can be compared to a list of transmitter codes or signal IDs stored in a lookup table in the autonomous vehicle's processor memory.
[0057] The on-board microprocessor can use the amplitude value to determine the azimuth and elevation of the received signals, which it can then use to determine its location within a working area. To determine the azimuth, the microprocessor enters the values of the two strongest readings from the four side photodiodes into an algorithm. The algorithm takes the ratio of these two readings to determine the azimuth angle. For example, if the two strongest readings from two photodiodes are equal, the algorithm recognizes that the point is located at an azimuth angle that is directly between the two photodiodes (i.e., at 45). In a similar algorithm, the amplitude value measured from the strongest side photodiode and the amplitude value measured from the top-facing photodiode value are used to determine the elevation of the signal. These values can be stored in the autonomous vehicle's memory for future reference.
[0058] After the receiver has detected at least two points, and determines the azimuth and elevation of each point, it determines its location within the working space. A triangulation algorithm based on the known ceiling height and the azimuth and elevation of the two detected points allows the processor to determine where in the working space the autonomous vehicle is located. Over time, the values of elevation and azimuth between each coded point and specific locations of the autonomous vehicle within the workspace can be stored in the autonomous vehicle's memory, creating a map of the environment in which the autonomous vehicle operates.
[0059] In various embodiments, a navigation system 200 as depicted in
[0060] In the embodiment depicted in
[0061] Of the four detectors that reside in a single plane, the values of the two strongest signals detected are used to form a ratio to determine the azimuth angle (Step 735). The ratio of second-strongest signal over the strongest signal is either compared to a look-up table or inserted into a mathematical equation to determine an azimuth angle output. Both the look-up table and the equation represent the overlap of the received sensitivity patterns of two orthogonal detectors with known sensor responses. In this embodiment, the photo detector output is modeled as a fourth-order Gaussian response to angle off of boresight, a term that generally refers to a vector that is orthogonal to the semiconductor die in the detector package.
[0062] To calculate elevation, the strongest signal from azimuth calculation (i.e., the denominator of the ratio) must first be normalized, as if it were on boresight of the respective detector (Step 740). For example, if the azimuth has been determined to be 10 off of boresight from a given detector, that 10 angle is entered into a look-up table or equation that describes the sensor response of any single photo detector. At zero degrees, the output of this look-up table/equation would be 1.00000. As the angle deviates from zero degrees, the output drops to some fraction of 1.00000 (the normalized value at boresight). For example, if a value of 10 is entered into the equation, the output of this operation can be, for example, 0.99000. The denominator of the azimuth ratio can then be divided by this fractional value in order to scale up, or normalize that value to what it would be if the azimuth were actually zero degrees. This normalized value can then be stored in memory and elevation can be determined therefrom.
[0063] To calculate elevation, the normalized output from the previous step is used to produce a new ratio with the output from the upward-looking (fifth) detector, so that the numerator is the second-strongest of the two values and the denominator is the strongest of the two values (Step 745). This ratio is then entered into the same look-up table or equation from the step above (used to calculate azimuth), thus outputting an elevation angle.
[0064] The benefits of this type of navigation system can be numerous. As the autonomous vehicle moves about a working area, measuring the azimuth and elevation of the various points detected, it can create a map of the area, thus determining its location within a given space. With this information, it can fuse data from all of its on-board sensors and improve cleaning or other task efficiency. One way it can do this is to create a map where the high-traffic areas in a house or other building are located (as indicated by readings from the dirt sensor, for example). The autonomous vehicle would then clean the areas it identified as high traffic (and therefore, often dirty) each time it passes over that area, whether directed to or not. The autonomous vehicle may also improve its cleaning function by merging the output from the wheel drop, stasis, bumper, and wall-following sensors to roughly mark areas of entrapment, or where large obstacles exist, so that those areas can potentially be avoided in future runs.
[0065] In accordance with various embodiments of the present teachings, another method of improving cleaning efficiency involves selectively programming the autonomous vehicle to clean particular areas, as detailed below. For example, a personal computer or remote control may be used to control the autonomous vehicle. Although the autonomous vehicle can operate without operator intervention, an operator can initially set up the autonomous vehicle, or can direct the autonomous vehicle to operate in particular areas or at particular times. For example, by using more then one transmitter in various rooms on one floor of a house, an operator may be able to direct the autonomous vehicle to clean specific rooms in a particular order and/or at a specific time. The operator could select, in a control program field of a computer program for example, the living room, family room, bathroom, and kitchen areas for cleaning. A remote control for use in accordance with the present teachings is described in more detail with respect to
[0066] Once commanded (either immediately or on a predefined schedule), the autonomous vehicle can be signaled to begin its cleaning cycle. The autonomous vehicle undocks from its base/charging station and begins cleaning the closest or first room on the programmed list. It can recognize this room and can differentiate it by the coded group of infrared points (e.g., on a ceiling of the room) or the coded signal emitted in the room. After the first room is cleaned, the autonomous vehicle can, for example, check its level of power and return to its charger for additional charging if needed. In accordance with certain embodiments, in order to return to the charger, the autonomous vehicle can follow the point or points on the ceiling back to the base station. Alternatively, the autonomous vehicle can use a known docking behavior. After charging is complete, the autonomous vehicle can traverse roughly back to the place it left off and resume cleaning. This sequence of events continues until all of the programmed rooms have been cleaned. Alternatively, the selection of particular areas to clean could be, for example, made by remote control or by pressing buttons on a control panel located on the base station. By using a personal computer, however, multiple transmitters could communicate with each other and with the base station via power lines using a known communication technology.
[0067] An alternative embodiment of the present teachings is depicted in
[0068] Alternatively, the autonomous vehicle 612 can emit its own coded pulse, to determine if any transmitters are in the area. This coded pulse could awaken sleeping or otherwise dormant transmitters, which would then begin their own emission cycle. Alternatively, the pulse can be an audible or visual signal such as a distinct beep, buzz, or visual strobe. Some pulses need not be within the field of view of the transmitter.
[0069] The robot 612 will continue to move toward signal 622a until one of several events happens at or near doorway 632a. In a first event, the autonomous vehicle may determine, based on readings from its photodiodes, that it is directly under the transmitter 620a. In a second event, the autonomous vehicle 612 may sense a second signal 624a, which may overlap the first detected signal 622m. This could occur if the configuration of the emitters, collimators, etc., as described in more detail above, provides overlapping signal patterns between signals 622a and 624a. In a third event, autonomous vehicle 612 can sense a signal from an entirely different transmitter, in this case signal 622b from transmitter 620b. Other events are also contemplated, as suitable for a particular application. The occurrence of an event presents the autonomous vehicle 612 with any number of behavioral, functional, or other options. For example, each coded signal may serve as a unique marker for a different working space. Upon detecting the unique marker associated with a particular working space, the autonomous vehicle may alter its cleaning function. Thus, if room A is carpeted but room B is uncarpeted, the autonomous vehicle can adjust its cleaning as it moves from room A to room B. Upon detecting a second signal (in this case, signal 622b) the autonomous vehicle can, in certain embodiments, completely disregard the first signal 622a received when its return to the base station 622 began. Using new signal 622b as a heading, it begins moving toward that signal 622b. The autonomous vehicle 612 can, in certain embodiments, check its battery level at each event, storing that value in its microprocessor. Over time, the autonomous vehicle can thereby create a table of battery levels at each event (and battery level change from event to event), and be able to accurately determine precise battery power remaining at each transmitter location.
[0070] Once the autonomous vehicle is traversing room B (shown in phantom as 617), it will eventually determine, based on battery level, time, or other factors, to follow the heading provided by signal 622b, and continue its return to its base station 622. The autonomous vehicle 612 can follow the heading until an event occurs at or near doorway 632b. Again, the event can be detecting a strength of signal 622b. Indicating that the autonomous vehicle is directly below the transmitter, detecting an overlap signal from 624b, or detecting a new signal 622c. The autonomous vehicle 612 can again perform any of the behaviors described above: check and store its battery level; change cleaning characteristics; etc.
[0071] Once in room C, the autonomous vehicle can begin following the heading provided by signal 622c. At or near the doorway 632c to room D, an event may direct the autonomous vehicle to perform any number of behaviors. Alternatively, the autonomous vehicle can move directly to charging station 622, guided by emitted signal 626 or another signal or program.
[0072] During its return to the base station, as the autonomous vehicle 612 moves from room A to room B to room C and so on, it detects and stores information about each coded signal that it detects along its route. By storing this information, the autonomous vehicle can create a map, using the coded signals as guideposts, allowing it to return to its starting location in the future. After charging, the autonomous vehicle can return to the room it was working in prior to returning to its base by comparing the detected signals and their strengths to the stored information.
[0073]
[0074]
[0075]
[0076]
[0077]
[0078]
[0079]
[0080] Lastly,
[0081] Accordingly, the navigation system can be operationally robust and adapted to compensate for variances in infrared energy. For example, if the autonomous vehicle is operating in an environment with high base infrared readings (e.g., a space with a large number of fluorescent lighting fixtures or windows that allow entry of sunlight), the autonomous vehicle can distinguish the infrared signals generated by the transmitter from the infrared noise present in the space. Similarly, the receiver can distinguish between other off-band signals, such as infrared remote controls. In such cases, establishing an initial threshold value of infrared energy and emitting a predefined, known, modulated infrared signal from the transmitter overcomes these environmental disturbances. Additionally, the transmitter can be tuned to emit a sufficiently strong infrared signal to accommodate surfaces with varied reflectivity.
[0082]
[0083] Wireless charging in accordance with the present teachings can comprise, for example, RF scavenging or magneto resonance. Wireless charging via RF scavenging can be accomplished as disclosed in U.S. Patent Publication No. 2009/0102296, the entire disclosure of which is incorporated herein by reference. The antenna 250 (e.g., a RF wireless communication antenna) can facilitate both energy harvesting and wireless communication for the transmitter 200 and, to facilitate energy harvesting, can harvest RF energy from a variety of sources including, for instance, medium frequency AM radio broadcast, very high frequency (VHF) FM radio broadcast, cellular base stations, wireless data access points, etc. The energy can be harvested from that naturally available in the environment (work area) or can be broadcast by a source such as an RF signal emitter on the autonomous vehicle or on another device such as a base station or a dedicated emitter.
[0084] Certain embodiments of the present teachings contemplate wireless charging via strongly coupled magnetic resonances, or magnetoresonance. Such wireless charging is described in detail in Kurs, et al., Wireless Power Transfer via Strongly Coupled Magnetic Resonances, Science, Vol. 317, pp. 83-86 (Jul. 6, 2008), the entire disclosure of which is incorporated herein by reference. For wireless charging via magnetoresonance, the antenna 250 can comprise, for example, a capture coil that can capture and convert magnetic energy to AC voltage or DC voltage. The magnetic energy captured by the capture coil can be supplied by a power source such as a highly resonant magnetic source. The power source can be located, for example, on the autonomous vehicle (in a scenario such as that illustrated in
[0085] One skilled in the art will appreciate that the transmitter 200 can derive its power from a source other than a battery, for example from a wall plug or by direct connection to a building's power supply. Also, the emitters can have differing locations on the transmitter, and need not be combined with a lens as illustrated. The size of the transmitter can vary in accordance with functional considerations (e.g., being large enough to house its components) as well as aesthetic considerations (e.g., minimizing size to be less obtrusive).
[0086]
[0087] One skilled in the art will appreciate that the transmitter 300 can derive its power from a source other than a battery, for example from a wall plug or by direct connection to a building's power supply. Also, the emitters can have differing locations on the transmitter, and need not be combined with a collimator and/or a lens as illustrated. The size of the transmitter can vary in accordance with functional considerations (e.g., being large enough to house its components) as well as aesthetic considerations (e.g., minimizing size to be less obtrusive).
[0088]
[0089] In embodiments of the present teachings employing more than two emitters, the signals can be utilized, e.g., with collimators or lenses, to distinguish different areas within a room. Such a configuration allows the autonomous vehicle 12 to sense its relative location within a room and adjust its cleaning behavior accordingly. For example, a signal could mark an area of the room that an autonomous vehicle would likely get stuck in. The signal could allow an autonomous vehicle to recognize the area and accordingly not enter it, even though it might otherwise be able to do so unimpeded. Alternatively or additionally, different signals could mark areas that require different cleaning behaviors (e.g., due to carpeting or wood floors, high traffic areas, etc.).
[0090] The transmitters 200, 300 as illustrated in
[0091]
[0092] As illustrated in
[0093] Another input device 330, shown in the illustrated embodiment as a toggle pad or toggle button, can allow the user to direct the autonomous vehicle to perform a number of functions. For example, the user can press a center CLEAN portion of the toggle button to direct the autonomous vehicle to begin cleaning immediately, or can select the right DOCK NOW button to direct the autonomous vehicle to begin a homing behavior and dock. A top SCHEDULE button can be pressed to allow the user to select a schedule of rooms and/or times for cleaning, an exemplary process for which is illustrated in
[0094]
[0095] In accordance with certain embodiments of the present teachings, the remote control 370, can also display a status screen such as that illustrated in
[0096] Other embodiments of the present teachings will be apparent to those skilled in the art from consideration of the specification and practice of the teachings disclosed herein. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the present teachings being indicated by the following claims.