METHODS AND SYSTEM FOR CYCLE RECOGNITION IN REPEATED ACTIVITIES BY IDENTIFYING STABLE AND REPEATABLE FEATURES
20220221482 · 2022-07-14
Inventors
Cpc classification
G01P13/00
PHYSICS
G01P1/106
PHYSICS
G06V20/52
PHYSICS
A61B5/6898
HUMAN NECESSITIES
A61B5/1123
HUMAN NECESSITIES
A61B2562/0219
HUMAN NECESSITIES
A61B5/0022
HUMAN NECESSITIES
International classification
Abstract
A system and method for monitoring performance of a repeated activity is described. The system comprises a motion sensing system and a processing system. The motion sensing system includes sensors configured to measure or track motions corresponding to a repeated activity. The processing system is configured to process motion data received from the motion sensing system to recognize and measure cycle durations in the repeated activity. In contrast to the conventional systems and methods, which may work for repeated activities having a high level of standardization, the system advantageously enables recognition and monitoring of cycle durations for a repeated activity, even when significant abnormal motions are present in each cycle. Thus, the system can be utilized in a significantly broader set of applications, compared conventional systems and methods.
Claims
1. A method for recognizing repetitions of a repeated activity, the method comprising: receiving, with a processor, first motion data corresponding to a first plurality of repetitions of a repeated activity, the first motion data including labels identifying time boundaries between each repetition in the first plurality of repetitions; identifying, with the processor, a salient segment of the first motion data corresponding to a motion of the repeated activity that occurs in all of the first plurality of repetitions; receiving, with the processor, second motion data corresponding to a second plurality of repetitions of the repeated activity from a motion sensing system; and identifying, with the processor, time boundaries between each repetition in the second plurality of repetitions by detecting segments of the second motion data that are most similar to the salient segment of the first motion data.
2. The method according to claim 1, the identifying the salient segment of the first motion data further comprising: identifying, with the processor, a plurality of candidate segments of the first motion data, each candidate segment in the plurality of candidate segments including a continuous portion of motion data starting from a different time within the first motion data; determining, with the processor, a respective score for each respective candidate segment in the plurality of candidate segments, the respective score indicating a level of similarity between the respective candidate segment and corresponding segments of each other repetition in the first plurality of repetitions; and selecting, with the processor, as the salient segment of the first motion data, the respective candidate segment of the plurality of candidate segments having the respective score that indicates a highest level of similarity of all of the plurality of candidate segments.
3. The method according to claim 2, the identifying the plurality of candidate segments further comprising: identifying, with the processor, for each respective repetition in the first plurality of repetitions, a respective plurality of candidate segments in the respective repetition.
4. The method according to claim 3 further comprising: dividing, with the processor, for each respective repetition in the first plurality of repetitions, the respective repetition into a respective plurality of frames, each frame including to a continuous portion of motion data captured during a respective interval of time.
5. The method according to claim 4, wherein, for each respective repetition in the first plurality of repetitions, each candidate segment of the respective plurality of candidate segments includes a continuous portion of motion data corresponding to a predetermined number of consecutive frames in the respective plurality of frames.
6. The method according to claim 3, the determining the respective score for each respective candidate segment further comprising: determining, with the processor, for each respective repetition in the first plurality of repetitions, a respective score for each respective candidate segment in the respective plurality of candidate segments in the respective repetition, the respective score indicating a level of similarity between the respective candidate segment and corresponding segments of each other repetition in the first plurality of repetitions.
7. The method according to claim 6, the determining the respective score for each respective candidate segment further comprising: determining, with the processor, for each respective candidate segment in the respective plurality of candidate segments in each respective repetition in the first plurality of repetitions, the corresponding segments of each other repetition in the first plurality of repetitions using a dynamic time warping algorithm.
8. The method according to claim 6, the selecting the salient segment of the first motion data further comprising: selecting, with the processor, for each respective repetition in the first plurality of repetitions, a respective salient segment of the respective repetition, the respective salient segment having the respective score that indicates a highest level of similarity of all of the respective plurality of candidate segments in the respective repetition.
9. The method according to claim 8, the selecting the salient segment of the first motion data further comprising: selecting, with the processor, as the salient segment of the first motion data, the respective salient segment having the respective score that indicates a highest level of similarity of all of the first plurality of repetitions.
10. The method according to claim 1, the receiving the second motion data further comprising: receiving, with the processor, the second motion data as a stream of measured motion data from the motion sensing system; and storing the stream of measured motion data of the second motion data in a buffer implemented on the memory.
11. The method according to claim 10, the identifying the time boundaries further comprising: detecting, with the processor, at least two segments of the second motion data in the buffer that are most similar to the salient segment of the first motion data; and determining, with the processor, at least one of the time boundaries located between at least two segments of the second motion data in the buffer.
12. The method according to claim 11, the detecting the at least two segments further comprising: identifying, with the processor, a plurality of window segments of the second motion data, each window segment in the plurality of candidate segments including a continuous portion of motion data starting from a different time within the second motion data and having a length equal to that of the salient segment of the first motion data.
13. The method according to claim 12, the detecting the at least two segments further comprising: determining, with the processor, for each respective window segment in the plurality of window segments, a respective score indicating a level of similarity between the respective window segment and the salient segment of the first motion data.
14. The method according to claim 13, the detecting the at least two segments further comprising: determining, as the at least two segments of the second motion data in the buffer, at least two window segments in the plurality of window segments having the respective scores that indicate a highest level of similarity with the salient segment of the first motion data, while also having in each case at least a threshold amount of time between one another.
15. The method according to claim 11, the determining the at least one of the time boundaries further comprising: mapping, with the processor, a continuous portion of the second motion data in the buffer that extends between two of the at least two segments of the second motion data in the buffer onto the first motion data using a dynamic time warping algorithm; and determining, with the processor, the at least one of the time boundaries located between the two of the at least two segments of the second motion data in the buffer, based on the mapping and the labels of the first motion data.
16. The method according to claim 11, the identifying the time boundaries further comprising: repeatedly performing, as further motion data of the second motion data is stored in the buffer, the (i) detecting the at least two segments of the second motion data in the buffer and (ii) the determining the at least one of the time boundaries between at least two segments of the second motion data in the buffer.
17. The method according to claim 1, further comprising: storing, in a memory, metadata of the first motion data, the metadata including the salient segment of the first motion data.
18. The method according to claim 1, further comprising at least one of: outputting, with an output device, the time boundaries between each repetition in the second plurality of repetitions.
19. A method for determining metadata of a repeated activity, the method comprising: receiving, with a processor, first motion data corresponding to a first plurality of repetitions of a repeated activity, the first motion data including labels identifying time boundaries between each repetition in the first plurality of repetitions; identifying, with the processor, a salient segment of the first motion data corresponding to a motion of the repeated activity that occurs in all of the first plurality of repetitions; and storing, in a memory, metadata of the first motion data, the metadata including the salient segment of the first motion data.
20. A method for recognizing repetitions of a repeated activity, the method comprising: storing, in a memory, metadata of first motion data corresponding to a first plurality of repetitions of a repeated activity, the metadata including a salient segment of the first motion data corresponding to a motion of the repeated activity that occurs in all of the first plurality of repetitions; receiving, with the processor, second motion data corresponding to a second plurality of repetitions of the repeated activity from a motion sensing system; identifying, with the processor, time boundaries between each repetition in the second plurality of repetitions by detecting segments of the second motion data that are most similar to the salient segment of the first motion data; and outputting, with an output device, the time boundaries between each repetition in the second plurality of repetitions.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0008] The foregoing aspects and other features of the methods and system are explained in the following description, taken in connection with the accompanying drawings.
[0009]
[0010]
[0011]
[0012]
[0013]
DETAILED DESCRIPTION
[0014] For the purposes of promoting an understanding of the principles of the disclosure, reference will now be made to the embodiments illustrated in the drawings and described in the following written specification. It is understood that no limitation to the scope of the disclosure is thereby intended. It is further understood that the present disclosure includes any alterations and modifications to the illustrated embodiments and includes further applications of the principles of the disclosure as would normally occur to one skilled in the art which this disclosure pertains.
System Overview
[0015]
[0016] In at least one embodiment, the repeated activity is a repeated human activity (also referred to herein as a “repeated task”) comprising motions performed by a human. As an example, the repeated human activity may comprise the repeated motions involved in the assembly of a product by a worker in an assembly line of a factory. As another example, the repeated human activity may comprise the repeated motions involved in certain types of physical exercise by an athlete (e.g. repetitions, steps, etc.). As a further example, the repeated human activity may comprise the repeated motions involved in scanning products for checkout at a retail store by a cashier. Finally, it will be appreciated that, in principle, the tracked motions may correspond to a repeated activity that is performed by some robot, tool, or other object, which may be directed by a human or performed autonomously.
[0017] The motion sensing system 110 comprises at least one sensor configured to track the motions that comprise the repeated activity. In at least some embodiments, the motion sensing system 110 comprises at least one inertial measurement unit (IMU) 112. The IMU 112 includes one or more gyroscope sensors and one or more accelerometers configured to provide motion data in the form of acceleration and orientation measurements. In one embodiment, the IMU 112 comprises an integrated 6-axis inertial sensor that provides both triaxial acceleration measurements and triaxial gyroscopic/orientation measurements. In at least one embodiment, the IMU 112 is worn on the body of a human and may, for example, take the form of a wrist-worn watch or a hand-worn glove having the IMU 112 integrated therewith. In other embodiments, the IMU 112 may be integrated with an object that is carried by the human, such as a smartphone or a tool used in the repeated task.
[0018] In further embodiments, motion sensing system 110 may alternatively include other types of sensors than the IMU 112, such as an RGB-D camera, infra-red sensors, ultrasonic sensors, pressure sensors, or any other sensor configured to measure data characterizes a motion. Additionally, in some embodiments, the motion sensing system 110 may include multiple different types of sensors that provide multi-modal motion data that is processed in a multi-channel manner by the processing system 120, or which is fused using one or more sensor data-fusion techniques by the processing system 120 or other component of the system 100.
[0019] The processing system 120 is configured to process motion data captured by the motion sensing system 110 to recognize and measure cycle durations in the repeated activity. As used herein, a “cycle” refers to an individual repetition of a repeated activity. Advantageously, the processing system 120 is trained to measure cycle durations in a repeated activity based on only a limited set of motion data that has been manually labeled with cycle boundaries (i.e., at least start and end times for each individual cycle). Based on this limited set of labeled motion data, the processing system 120 determines the most salient features of the repeated activity and generates metadata of labeled motion data that is used to determine cycle durations in new unlabeled motion data.
[0020] In the illustrated exemplary embodiment, the processing system 120 comprises at least one processor 122, at least one memory 124, a communication module 126, a display screen 128, and a user interface 130. However, it will be appreciated that the components of the processing system 120 shown and described are merely exemplary and that the processing system 120 may comprise any alternative configuration. Particularly, the processing system 120 may comprise any computing device such as a desktop computer, a laptop, a smart phone, a tablet, or another electronic device. Thus, the processing system 120 may comprise any hardware components conventionally included in such computing devices.
[0021] The memory 124 is configured to store data and program instructions that, when executed by the at least one processor 122, enable the processing system 120 to perform various operations described herein. The memory 124 may be of any type of device capable of storing information accessible by the at least one processor 122, such as a memory card, ROM, RAM, hard drives, discs, flash memory, or any of various other computer-readable medium serving as data storage devices, as will be recognized by those of ordinary skill in the art. Additionally, it will be recognized by those of ordinary skill in the art that a “processor” includes any hardware system, hardware mechanism or hardware component that processes data, signals or other information. Thus, the at least one processor 122 may include a central processing unit, graphics processing units, multiple processing units, dedicated circuitry for achieving functionality, programmable logic, or other processing systems. Additionally, it will be appreciated that, although the processing system 120 is illustrated as single device, the processing system 120 may comprise several distinct processing systems 120 that work in concert to achieve the functionality described herein.
[0022] The communication module 126 may comprise one or more transceivers, modems, processors, memories, oscillators, antennas, or other hardware conventionally included in a communications module to enable communications with various other devices. In at least some embodiments, the communication module 126 includes a Wi-Fi module configured to enable communication with a Wi-Fi network and/or Wi-Fi router (not shown). In further embodiments, the communications module 126 may further include a Bluetooth® module, an Ethernet adapter and communications devices configured to communicate with wireless telephony networks.
[0023] The display screen 128 may comprise any of various known types of displays, such as LCD or OLED screens. In some embodiments, the display screen 128 may comprise a touch screen configured to receive touch inputs from a user. The user interface 130 may suitably include a variety of devices configured to enable local operation of the processing system 120 by a user, such as a mouse, trackpad, or other pointing device, a keyboard or other keypad, speakers, and a microphone, as will be recognized by those of ordinary skill in the art. Alternatively, in some embodiments, a user may operate the processing system 120 remotely from another computing device which is in communication therewith via the communication module 126 and has an analogous user interface.
[0024] The program instructions stored on the memory 124 include a repeated activity monitoring program 132. As discussed in further detail below, the processor 122 is configured to execute the repeated activity monitoring program 132 to process labeled motion data captured by the motion sensing system 110 to derive metadata describing the most salient features of a repeated activity. Additionally, the processor 122 is configured to execute the repeated activity monitoring program 132 to process unlabeled motion data captured by the motion sensing system 110 to recognize and measure cycle durations in the repeated activity.
Methods for Cycle Duration Recognition
[0025]
[0026] In summary, the method 200 has two major components: an offline preprocessing phase and an online processing phase. In the offline preprocessing phase, labeled motion data corresponding to a plurality of individual cycles of a repeated activity are provided as input. The labeled motion data includes labels that identify the time boundaries for each cycle of the repeated activity. A plurality of segments of the motion data are identified as feature candidates and each evaluated for their salience across all of the individual labeled cycles. For each labeled cycle, the feature candidate that is evaluated as most consistent across all of the individual labeled cycles is selected as the cycle feature, and most consistent one of the selected cycle features is selected as the main feature of the repeated activity. Finally, metadata describing the cycle features, the main feature, and other properties of the labeled cycles are determined and provided as an output.
[0027] Next, in the online processing phase, unlabeled motion data corresponding to a plurality of individual cycles of the repeated activity are provided as input. Cycle features are identified in the unlabeled motion data and cycle boundaries in unlabeled motion data are determined based on the identified cycle features. Based on the identified cycle boundaries in the unlabeled motion data, cycle durations are determined and provided as an output.
[0028] In greater detail and with continued reference to
[0029] The method 200 continues, in the offline preprocessing phase, with identifying a plurality of feature candidates in the plurality of individual cycles (block 220). Particularly, the processor 122 identifies a plurality of feature candidate segments (also referred to herein as simply “feature candidates”) of the labeled motion data. Each feature candidate is a continuous portion of the labeled motion data in a particular cycle of the plurality of individual cycles of the repeated activity. As will be described below, these feature candidates are evaluated for their salience and one or more are selected as being the most salient segment(s) and are used to identify cycles of the repeated activity in new unlabeled motion data.
[0030] In at least one embodiment, the processor 122 is configured to divide the labeled motion data into a plurality of frame segments (also referred to herein as simply “frames”). More particularly, the processor 122 divides each particular cycle of the labeled motion data into a respective plurality of frames. Each frame of each cycle of the labeled motion data corresponds to a discrete interval of time and, thus, comprises a continuous portion of the labeled motion data that was captured during the discrete interval of time. In at least one embodiment, each frame corresponds to a fixed duration of time (e.g., 2 seconds). Alternatively, in at least one embodiment, each frame corresponds to a duration of time that is a fixed percentage of the total duration of cycle (e.g., 10% of the cycle duration). In at least one embodiment, each frame is defined such that there is a predetermined amount of overlap with the previous frame and with the subsequent frame (e.g., 50% overlap).
[0031] As noted above, the labeled motion data corresponds to a plurality of individual cycles of the repeated activity. Each individual cycle of the labeled motion data is denoted C.sub.i, where 0≤i<N is the cycle index and N is the total number of cycles. Each cycle C.sub.i is divided into a plurality of frames f.sub.i.sup.j, where 0≤j<L.sub.i is the frame index and L.sub.i is the total number of frames in the cycle C.sub.i or, in other words, L.sub.i is the length of cycle C.sub.i.
[0032] Next, the processor 122 is configured to group sets of consecutive frames together to form the plurality of feature candidate, denoted F.sub.i.sup.j. More particularly, for each individual cycle C.sub.i of the labeled motion data, the processor 122 groups sets of W.sub.k consecutive frames f.sub.i.sup.j together to form a respective plurality of feature candidates F.sub.i.sup.j, where W.sub.k the total number of frames in each feature candidates F.sub.i.sup.j or, in other words, W.sub.k is the length of each feature candidate F.sub.i.sup.j. Each feature candidate F.sub.i.sup.j thus comprises the motion data of consecutive frames f.sub.i.sup.j, . . . , f.sub.i.sup.j+W.sup.
[0033] In at least one embodiment, for each individual cycle C.sub.i, given a value for W.sub.k, the processor 122 is configured to determine the plurality of feature candidates F.sub.i.sup.j as including each possible combination of W.sub.k consecutive frames f.sub.i.sup.j in the cycle individual cycle C.sub.i. In other words, the processor 122 is configured to determine the plurality of feature candidates F.sub.i.sup.j where 0≤j≤M.sub.i is the starting frame index of the feature candidates F.sub.i.sup.j and M.sub.i=L.sub.i−W.sub.k is the starting frame index of that last in time feature candidate F.sub.i.sup.j. With continued reference to
[0034] In one embodiment, at least initially, the processor 122 determines the plurality of feature candidates F.sub.i.sup.j for each cycle C.sub.i only for a predetermined starting/minimum feature length W.sub.k (e.g., for W.sub.k=3). However, in some embodiments, the processor 122 determines the plurality of feature candidates F.sub.i.sup.j for each cycle C.sub.i for each of a range of feature lengths W.sub.k (e.g., for W.sub.k=3, 4, 5).
[0035] Returning to
[0036] In at least some embodiments, the processor 122 selects a best or most salient feature in each individual cycle C.sub.i of the labeled motion data, which are referred to herein as cycle features. Particularly, the processor 122 determines the cycle feature for each individual cycle C.sub.i, denoted F.sub.i.sup.best, as the feature candidate in the plurality of feature candidates F.sub.i.sup.j that has a highest degree of similarity across corresponding regions of all of the labeled cycles {C.sub.i}.sub.i=0.sup.N−1, and which can also be uniquely identified within each individual cycle C.sub.i (i.e., it doesn't occur multiple times within individual cycles). Thus, the processor 122 determines a set of N cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1, where each cycle feature F.sub.i.sup.best is the most salient feature in the respective cycle C.sub.i.
[0037] In at least some embodiments, after the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1 is determined, the processor 122 selects one cycle feature from the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1 as the main feature, denoted F.sup.main, for the repeated activity. Particularly, the processor 122 determines which cycle feature from the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1 has the highest degree of similarity across corresponding regions of all of the labeled cycles {C.sub.i}.sub.i=0.sup.N−1.
[0038] As noted above, the cycle features F.sub.i.sup.best and the main feature F.sup.main are selected from the feature candidates F.sub.i.sup.j based on their degree of similarity across corresponding regions of all of the individual cycles {C.sub.i}.sub.i=0.sup.N−1. Notably, since the different individual cycles {C.sub.i}.sub.i=0.sup.N−1 may have different lengths and may include different motions, the corresponding regions of each other cycle C.sub.c, where c≠i, may have a different duration and timing compared to the feature candidate F.sub.i.sup.j. In at least one embodiment, for each feature candidate F.sub.i.sup.j in each cycle C.sub.i, the processor 122 determines the corresponding regions of each other cycle C.sub.c, where c≠i, using a dynamic time warping (DTW) algorithm. In this way, the corresponding regions describe the same or mostly similar motion, but may have different duration and timing.
[0039]
[0040] In at least some embodiments, the processor 122 determines a respective evaluation score S.sub.i.sup.j for each the plurality of feature candidates F.sub.i.sup.j for each individual cycle C.sub.i. The evaluation score S.sub.i.sup.j indicates a degree of similarity between the feature candidate F.sub.i.sup.j and the corresponding regions of each other cycle C.sub.c, where c≠i, in the plurality of individual cycles {C.sub.i}.sub.i=0.sup.N−1. In at least one embodiment, the processor 122 calculates the evaluation score S.sub.i.sup.j of a feature candidate F.sub.i.sup.j as the average geometric distance/difference between the motion data of the feature candidate F.sub.i.sup.j and the motion data of the corresponding regions of each other cycles C.sub.c. Notably, as a result of the dynamic time warping, the evaluation score S.sub.i.sup.j ignores temporal distance/difference between the motion data of the feature candidate F.sub.i.sup.j and the motions of the corresponding regions of each other cycles C.sub.c
[0041] Once all of the evaluation scores S.sub.i.sup.j for the plurality of feature candidates F.sub.i.sup.j for each individual cycle C.sub.i are calculated, the processor 122 selects the feature candidate F.sub.i.sup.j having the best score in each individual cycle C.sub.i, as the respective cycle feature F.sub.i.sup.best for each individual cycle C.sub.i, thus deriving the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1. It should be appreciated that, in the example in which the evaluation score S.sub.i.sup.j an average distance/difference between corresponding regions across all cycles, the best score indicating the highest degree of similarity is the lowest average distance/difference. Finally, the processor 122 selects the cycle feature F.sub.i.sup.best in the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1 having the best score as the main feature F.sup.main for the repeated activity.
[0042] As noted above, the cycle features F.sub.i.sup.best and the main feature F.sup.main for the repeated activity should be uniquely identifiable within each individual cycle C.sub.i. In some embodiments, the processor 122 evaluates the uniqueness of the cycle features F.sub.i.sup.best (or the main feature F.sup.main) by determining a uniqueness score as an average geometric distance/difference between the motion data of the cycle feature F.sub.i.sup.best and the motions data of each other feature candidate F.sub.i.sup.j in the same cycle C.sub.i. In this case a higher average geometric distance/difference indicates a more unique feature. In some embodiments, the feature length W.sub.k is be gradually increased so that the cycle features F.sub.i.sup.best and the main feature F.sup.main are significantly different from other feature candidates within their respective cycles. In one embodiment, the processor 122 increases the feature length W.sub.k until a threshold uniqueness score is achieved.
[0043] The method 200 continues, in the offline preprocessing phase, with determining and storing metadata of labeled motion data including at least the main feature (block 240). Particularly, the processor 122 writes to the memory 124 metadata of the labeled motion data, which at least includes the main feature F.sup.main for the repeated activity. In some embodiments, the stored metadata of the labeled motion data further includes the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1. As detailed below, the stored metadata will be used to identify cycles of the repeated activity in new unlabeled motion data.
[0044] In some embodiments, the stored metadata of the labeled motion data further includes a previously determined evaluation score S.sup.main for the main feature F.sup.main. In some embodiments, the stored metadata of the labeled motion data further includes a previously determined set of evaluation scores {S.sub.i.sup.best}.sub.i=0.sup.N−1 for the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1.
[0045] In some embodiments, the stored metadata of the labeled motion data further includes the plurality of individual cycles {C.sub.i}.sub.i=0.sup.N−1 (i.e., the original labeled motion data itself). In one embodiment, the stored metadata of the labeled motion data further includes a plurality of re-aligned cycles D.sub.i that start and end with regions corresponding to the main feature F.sup.main in the plurality of individual cycles {C.sub.i}.sub.i=0.sup.N−1.
[0046] Returning to
[0047] The processor 122 is configured to accumulate unlabeled motion data in the buffer, without further processing, until a threshold amount of unlabeled motion data is accumulated. In at least one embodiment, the processor 122 is configured to divide the unlabeled motion data into a plurality of frame, preferably having the same length as the frames of the unlabeled motion data (e.g., 2 seconds). In at least one embodiment, each frame is defined such that there is a predetermined amount of overlap with the previous frame and with the subsequent frame (e.g., 50% overlap). In one embodiment, the threshold amount of unlabeled motion data is twice the average of the cycle lengths L.sub.i of the plurality of individual cycles {C.sub.i}.sub.i=0.sup.N−1 of the labeled motion data (e.g., 20 frames).
[0048] The method 200 continues, in the online processing phase, with identifying regions within the unlabeled motion data that correspond to the main feature (block 260). Particularly, once the buffer is filled with the threshold amount of unlabeled motion data, processor 122 identifies regions within the buffered unlabeled motion data that correspond to the same or essentially similar motion as the main feature F.sup.main for the repeated activity, which was stored in metadata. In the exemplary case that the threshold amount of unlabeled motion data is twice the average of the cycle length, then the processor 122 should identify at least two regions corresponding to the main feature F.sup.main.
[0049] In at least some embodiments, the processor 122 is configured to utilize a sliding window approach to detect the regions corresponding to the main feature F.sup.main Particularly, a sliding window having a window length equaling the feature length W.sub.k of the main feature F.sup.main is slid across the frames in the unlabeled motion data in the buffer from beginning to end. Each particular position of the sliding window corresponds to a candidate region and is evaluated to determine its similarity with the main feature F.sup.main The processor 122 selects the candidate regions that are most similar to the main feature F.sup.main while also being sufficiently far apart from one another, as the regions corresponding to the main feature F.sup.main.
[0050] In at least some embodiments, the processor 122 determines a respective evaluation score S.sub.m, where m is the frame index within the buffer, for each position of the sliding window in the buffer or, in other words, each candidate region in the buffer. Each evaluation score S.sub.m indicates a degree of similarity between the motion data within the sliding window at the position m within the buffer and the motion data of the main feature F.sup.main. In at least one embodiment, the processor 122 calculates the evaluation score S.sub.m for the sliding window at each position m as the average geometric distance/difference between motion data within the sliding window at the position m within the buffer and the motion data of the main feature F.sup.main. Thus, for a given buffer of unlabeled motion data, the processor determines a set of evaluation scores {S.sub.m}.sub.m=0.sup.M−1, where M is the total number of frames in the buffer.
[0051] Next, based on the set of evaluation scores {S.sub.m}.sub.m=0.sup.M−1, the processor 122 is configured to identify which positions of the sliding window or, in other words, which candidate regions within the buffer, correspond to the main feature F.sup.main with each unlabeled cycle of the unlabeled motion data. It should be appreciated that, in the case that the evaluation score S.sub.m is the average geometric distance/difference between motion data within the sliding window at the position m within the buffer and the motion data of the main feature F.sup.main then the lowest scores are the best scores indicating the highest degree of similarity. Thus, in one embodiment, the processor 122 selects those sliding window positions having the lowest evaluation score S.sub.m and which are sufficiently spaced apart from one another (e.g., at least the minimum cycle length L.sub.i from the labeled motion data) as being the regions corresponding to the main feature F.sup.main.
[0052] In one embodiment, the processor 122 is configured to determine the sliding window positions that are the regions corresponding to the main feature F.sup.main according to the following process. First, the processor eliminates evaluation scores S.sub.m that are not either (i) a local minimum in the set of evaluation scores {S.sub.m}.sub.m=0.sup.M−1 or (2) less that a predetermined threshold score (e.g., the maximum of the set of evaluation scores {S.sub.i.sup.best}.sub.i=0.sup.N−1 for the set of cycle features {F.sub.i.sup.best}.sub.i=0.sup.N−1). In other words, only evaluation scores S.sub.m that are less than the predetermined threshold score or a local minimum are kept for consideration. Next, the processor 122 sorts the remaining evaluation scores S.sub.m from highest to lowest (i.e., worst to best). Finally, the processor 122 checks each remaining evaluation scores S.sub.m, from highest to lowest. If the start position m of the corresponding window is within a threshold distance (e.g., the minimum cycle length L.sub.i) the start position of an adjacent window, then the evaluation score S.sub.m is eliminated. The processor 122 performs this check iteratively from the highest to lowest remaining evaluation score S.sub.m until the only remaining evaluation scores S.sub.m are sufficiently far apart from one another (e.g., at least the minimum cycle length L.sub.i). The processor 122 identifies these remaining sliding window positions as being the regions corresponding to the main feature F.sup.main.
[0053] The method 200 continues, in the online processing phase, with identifying time boundaries between the plurality of individual cycles in the unlabeled motion data (block 270). Particularly, the processor 122 determines the time boundaries between each individual cycle of the repeated activity in the unlabeled motion data stored in the buffer. As noted above, in the exemplary case that the buffer includes unlabeled motion data that totals twice the average of the cycle length, the processor 122 identifies at least two regions corresponding to the main feature F.sup.main. Accordingly, it should be appreciated that a time boundary between the at least two individual cycle will be located between the first region corresponding to the main feature F.sup.main and the second region corresponding to the main feature F.sup.main.
[0054] The region starting from a region corresponding to the main feature F.sup.main and the next region corresponding to the main feature F.sup.main in the unlabeled motion data is denoted as a miss-aligned cycle d.sub.i. The processor 122 maps the miss-aligned cycle d.sub.i to one of the re-aligned cycles D.sub.1 in the set of re-aligned cycles {D.sub.i}.sub.i=0.sup.N−2 using a mapping algorithm, such as a dynamic time warping algorithm. Since the time boundaries of the cycles C.sub.i have known time boundaries within the re-aligned cycles D.sub.i, the processor 122 projects this known time boundary back into to the miss-aligned cycle d.sub.i of the unlabeled motion data to determine the estimated time boundary between cycles of the unlabeled motion data.
[0055] The processor 122 performs the processes of blocks 260 and 270 iteratively for each new buffer of unlabeled motion data. In this way, the processor 122 determines a plurality of time boundaries between a plurality of cycles in the unlabeled motion data, thus labeling the unlabeled motion data. From the plurality of time boundaries, the processor determines a plurality of cycle durations by calculating the time difference between consecutive time boundaries.
[0056] The method 200 continues, in the online processing phase, with outputting labels indicating the time boundaries or cycle durations for the unlabeled motion data (block 280). Particularly, the processor 122 outputs, with an output device, the time boundaries between each cycle in the plurality of cycles of the unlabeled motion data. For example, in some embodiments, the processor 122 writes the previously determined plurality of time boundaries and/or plurality of time boundaries to the memory 124. In some embodiments, the processor 122 operates the display 128 to display the previously determined plurality of time boundaries and/or plurality of time boundaries. In some embodiments, the processor operates the communication module 126 to transmit the previously determined plurality of time boundaries and/or plurality of time boundaries to another device.
[0057] Embodiments within the scope of the disclosure may also include non-transitory computer-readable storage media or machine-readable medium for carrying or having computer-executable instructions (also referred to as program instructions) or data structures stored thereon. Such non-transitory computer-readable storage media or machine-readable medium may be any available media that can be accessed by a general purpose or special purpose computer. By way of example, and not limitation, such non-transitory computer-readable storage media or machine-readable medium can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures. Combinations of the above should also be included within the scope of the non-transitory computer-readable storage media or machine-readable medium.
[0058] Computer-executable instructions include, for example, instructions and data which cause a general-purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Computer-executable instructions also include program modules that are executed by computers in stand-alone or network environments. Generally, program modules include routines, programs, objects, components, and data structures, etc. that perform particular tasks or implement particular abstract data types. Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represents examples of corresponding acts for implementing the functions described in such steps.
[0059] While the disclosure has been illustrated and described in detail in the drawings and foregoing description, the same should be considered as illustrative and not restrictive in character. It is understood that only the preferred embodiments have been presented and that all changes, modifications and further applications that come within the spirit of the disclosure are desired to be protected.