MEASURING TOOL FOR ASSESSING GOLF BALL ALIGNMENT
20260021365 ยท 2026-01-22
Assignee
Inventors
- Charles Hightower (Hyde Park, MA, US)
- Paul Furze (Tiverton, RI, US)
- Jordan R. Tate (Providence, RI, US)
- Donald A. Serino (Plymouth, MA, US)
- Michael J. Toupin (Fall River, MA, US)
- Ian W. Connolly (Trumbull, CT, US)
Cpc classification
A63B69/3688
HUMAN NECESSITIES
A63B2220/05
HUMAN NECESSITIES
A63B2225/02
HUMAN NECESSITIES
International classification
Abstract
An alignment measuring tool has a computing device having at least one processor and at least one memory, a camera configured to capture images of a surface within a field of view, and a support structure configured to support the computing device and the camera in a stationary position. The support structure has a plurality of support walls forming a perimeter of an open bottom side of the support structure, wherein the plurality of support walls are configured to rest on the surface to enclose the field of view of the camera to an area of the surface. The support structure also has a calibration feature configured to provide feedback on a position of the camera relative to an area outside of the field of view of the camera.
Claims
1. An alignment measuring tool, comprising: a computing device comprising at least one processor and at least one memory; a camera configured to capture images of a surface within a field of view; a support structure configured to support the computing device and the camera in a stationary position, the support structure comprising: a plurality of support walls forming a perimeter of an open bottom side of the support structure, wherein the plurality of support walls are configured to rest on the surface to enclose the field of view of the camera to an area of the surface, and a calibration feature configured to provide feedback on a position of the camera relative to an area outside of the field of view of the camera.
2. The alignment measuring tool of claim 1, wherein the camera is integrated with the computing device.
3. The alignment measuring tool of claim 2, wherein the computing device is a mobile device comprising at least one of a tablet or mobile phone.
4. The alignment measuring tool of claim 3, wherein the computing device and camera are removable from the support structure.
5. The alignment measuring tool of claim 4, wherein the support structure further comprises an open top side, wherein the computing device is configured to rest at the open top side with the camera facing toward the bottom side.
6. The alignment measuring tool of claim 5, wherein the support structure further comprises a support cradle for supporting the mobile device.
7. The alignment measuring tool of claim 6, wherein the support cradle comprises a perimeter ledge and the computing device is configured to rest on the perimeter ledge and close the open top side of the support structure.
8. The alignment measuring tool of claim 5, further comprising a user interface integrated with the computing device, the user interface facing away from the support structure when the computing device is resting on the support structure.
9. The alignment measuring tool of claim 8, wherein the computing device comprises a display configured to present the user interface.
10. The alignment measuring tool of claim 1, wherein the camera is configured with a vertical field of view of 2-4 inches.
11. The alignment measuring tool of claim 1, wherein the calibration feature comprises one or more positional markers connected to the plurality of support walls.
12. The alignment measuring tool of claim 11, wherein the one or more positional markers each comprise a first component configured to be securely connected to the surface and a second component connected to the support structure.
13. The alignment measuring tool of claim 12, wherein the first component comprises a connector configured to mate with a corresponding connector of the second component.
14. The alignment measuring tool of claim 13, wherein the first component is configured to be embedded flush with the surface.
15. The alignment measuring tool of claim 1, wherein the support structure comprises handles for picking up the alignment measuring tool.
16. The alignment measuring tool of claim 1, further comprising a flash inside the support structure configured to control light exposure of the camera.
17. An alignment measuring tool, comprising: a computing device comprising at least one processor and at least one memory; a camera configured to capture images of a surface within a field of view, a support structure configured to support the computing device and the camera in a stationary position, the support structure being spaced from the field of view of the camera to allow a golfer to place a golf ball on the surface within the field of view without hindrance by the support structure.
18. The alignment measuring tool of claim 17, wherein the camera is configured with a vertical field of view of 2-4 inches.
19. The alignment measuring tool of claim 17, further comprising an enclosure configured to be placed on the surface spaced from the camera and the support structure, the enclosure comprising an opening providing an aperture for the camera to capture images of the surface within the enclosure.
20. The alignment measuring tool of claim 17, further comprising a calibration line configured to connect a point within the field of view of the camera to a target point outside of the field of view of the camera.
Description
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] The foregoing and other aspects of the present invention are best understood from the following detailed description when read in connection with the accompanying drawings. For the purpose of illustrating the invention, there are shown in the drawings embodiments that are presently preferred, it being understood, however, that the invention is not limited to the specific instrumentalities disclosed. Included in the drawings are the following Figures:
[0007]
[0008]
[0009]
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
[0016]
[0017]
[0018]
[0019]
[0020]
[0021]
[0022]
[0023]
[0024]
[0025]
[0026]
DETAILED DESCRIPTION OF THE INVENTION
[0027] In golf, especially in putting, the goal is to direct the golf ball from the place that it lies toward an intended target in order to make the ball in the hole in the fewest number of strokes. When on a putting green, a golfer can mark the position of the golf ball, pick up the golf ball, and place it back down in an orientation such that an alignment aid points toward the intended target. This practice allows the alignment aid to be used as a visual cue to help direct the golf ball as intended with the ensuing putt. For example, for a straight, flat putt, a golfer would ideally putt the ball from its position on the green toward a center of the hole. The golfer might place their golf ball in such a way that the alignment aid is pointing directly at the center of the hole. In other situations (e.g., for breaking putts), the golfer might choose a point on the green or in the distance as a target point. The disclosed embodiments include systems for measuring the degree to which a golf ball has been placed with the alignment aid aimed directly at a target point.
[0028] As used herein, alignment aid generally refers to a visual element that is on or integral with an object, such as a golf ball, wherein the visual element has an identifiable direction for being oriented toward a target point. In general, any indicia on a golf ball can be considered an alignment aid if the indicia defines a predominant axis for being aimed at a target point. Various examples of golf balls having different alignment aids are shown in
[0029] Each alignment aid in
[0030]
[0031] As shown in
[0032]
[0033] Ideally, a golfer would place their golf ball such that the aimed direction is the same as the target direction. This would represent perfect alignment of the golf ball to assist the golfer with directing the golf ball toward the target. However, as a golfer may not actually place the golf ball with the alignment aid perfectly aligned with the target direction, there may be an offset angle between the target direction and the aimed direction, as shown in
[0034] There are several challenges associated with creating a device that can reliably determine an aimed direction and compare the aimed direction to a target direction. Initially when placing a golf ball for an alignment measurement, the golfer should be given enough space to simulate a real golf scenario. That is, an alignment measurement would be most representative of a golfer's actual alignment ability if there were no cues or lines present within the test space that would give the golfer an advantage when orienting the alignment aid. The present embodiments therefore contemplate a device that can be moved away from and/or is spaced from the placement position to thereby enable a golfer to place a golf ball as they typically would during a real golf scenario. Further, because it is not difficult to place an alignment aid such that it is aimed in the general direction of a target (even if perfect alignment is difficult), the measurement capabilities of the disclosed embodiments are such that tolerances and margins of error must be very small. These objectives are further complicated as the distance between a placement position and a target point increases (especially once the placement position and target point are out of a single field of view of a measuring device such as a camera). Disclosed embodiments include features to achieve the objectives of providing a simulated experience for the golfer to place the golf ball while maintaining repeatably high accuracy and very low margin of error.
[0035]
[0036] In order to produce a simulation of a real golf scenario, the alignment measuring tool should allow for placement of a golf ball without any visual cues other than the location of the target point. As a result, the golf ball should not be placed on the placement position while any calibration lines are visible within the area. This makes an option in which a target direction and aimed direction are determinable from a same captured image more difficult. While some embodiments of the present disclosure may include comparison of an aimed direction and a target direction from a single image, other embodiments allow the target direction to be stored, the aimed direction to be determined, and an alignment value to be determined based on a comparison of the two directions from different images. Another embodiment includes an alignment measuring tool that measures its own relationship to the target point and uses the result to calculate an alignment value. For instance, the target direction may be extrapolated from the orientation of a field of view of an image. For example, the direction of a vertical field of view may be defined as the target direction. The alignment measuring tool may include features that ensure that the vertical field of view of the camera is aligned with the target direction prior to an image being captured.
[0037] The alignment measuring tool 100 may include, for example, a computing system 110, a camera 120, a support structure 130, and a user interface 140. These features and/or other components of the alignment measuring tool 100 may act in combination to perform one or more disclosed methods for assessing golf ball alignment. The computing system 110 may include a processor and/or memory device configured to execute software instructions. The camera 120 may be configured to capture an image of an area within a field of view of the camera 120. The support structure 130 may be a physical device configured to be positioned relative to a golf ball in order to support the camera 120. The support structure 130 may include positional markers that allow the computing system 110 to determine a frame of reference to compare the target direction with the aimed direction. The user interface 140 may be a software and/or hardware component configured to enable the alignment measuring tool 100 to provide output that can be viewed by a user.
[0038]
[0039] In an exemplary embodiment, the computing system 110 is a computer such as laptop or mobile device. In a particular embodiment, the computing system 110 is a tablet or phone. These devices include processing units, memory and other storage, and input/output devices such as touch screens, data ports, wireless connections (e.g., Wi-Fi, Bluetooth, etc.), and the like. In other embodiments, the computing system 110 may include multiple computing devices, such as a plurality of computers, one or more servers, multiple mobile devices, etc., connected via a wired or wireless network. Disclosed embodiments contemplate a computing system 110 that enables measurement of a golf ball's alignment in one location with the output of information via a user interface at another location.
[0040] An exemplary computing system 110 may include a plurality of modules that may be hardware and/or software-based components that enable the computing system 110 to complete one or more steps of a disclosed process. The disclosed modules are presented separately for ease of explanation, but it should be understood that one or more of the modules may be combined into a single hardware or software feature, such as an application or program. For example, the computing system 110 may store an alignment measuring software application that is, for example, a Single Page Application (SPA) which captures and processes ball images which contain markings. The computing system 110 may include a component, such as a database and/or server, that stores images, analysis data, and user information that can be accessed via an application programming interface (API). The SPA may be programmed using coding languages known in the art such as Java, C, C++, C #, etc. In an exemplary embodiment, the SPA may be programmed using Blazor web framework available from Microsoft. In an exemplary embodiment, the computing system 110 includes one or more of an imaging module 160, a calibration module 162, and an analysis module 164.
[0041] In an exemplary embodiment, the imaging module 160 may be configured to communicate with the camera 120 in order to receive image data (e.g., data associated with captured images). In some embodiments, the imaging module 160 may be configured to transmit image data to other components of the alignment measuring tool 100. For example, the imaging module 160 may provide image data to the user interface 140 in order to display a captured image to a user.
[0042] In some embodiments, the imaging module 160 may use image processing and/or recognition techniques to extract information from image data associated with one or more images. For example, the imaging module 160 may be configured to process image data to identify objects or markers within the image, such as a golf ball, alignment aid, or calibration line. In an exemplary embodiment, the imaging module 160 may be configured to perform a best-fit analysis to associate a line or direction with an alignment aid or calibration line found in image data. For example, the imaging module 160 may extract lines from image data by finding similar pixels that have the same characteristics and form a line or contour within the image. An example of a software application that may be used to perform image processing to extract lines from image data is the HALCON vision application available from MVTEC Software.
[0043] In an exemplary embodiment, the imaging module 160 is configured to use image data to determine one or more parameters, such as a target direction and aimed direction. The imaging module 160 may process calibration images to determine and store a target direction in the memory 152. Additionally, the imaging module 160 may process image data to determine an aimed direction of an alignment aid on a golf ball. For example, the imaging module 160 may process image data to identify an alignment aid, determine a predominant axis of the alignment aid, and determine an aimed direction of the alignment aid based on the predominant axis.
[0044] In an exemplary embodiment, the imaging module 160 (or other component of the computing system 110) may perform image processing and extract lines from image(s) by detecting pixels within the image data whose intensity have the same line direction. For example, pixels that connect and have the same intensity and direction may exhibit the characteristic of having a local maxima (i.e., their second derivative is 0). All of the points found this way should fall on the same line (i.e., they have the same slope). The imaging module 160 may collect these points to form a contour. Next, the imaging module 160 may perform a linear regression on detected contours to produce a least-squares estimate of the line. For example, two or more points along the best fit line estimate may be identified as the target direction. These points may be used later to determine the angle between the target direction and the aimed direction. For example, the imaging module 160 may use the two farthest points on the best-fit line as a representation of the target direction (the line connecting these points follows the path on the target direction).
[0045] In another example, the imaging module 160 is configured to extract an alignment aid from image data to determine an aimed direction. For example, the imaging module 160 may use pattern matching to detect an alignment aide within image data. Next, the imaging module 160 may perform a linear regression on the contours of the alignment aid to produce a least-squares estimate of the predominant axis of the alignment aid (e.g., the line that intersects the center-line of the alignment aide). Two or more points along the best fit line may be stored as the aimed direction. For example, the imaging module 160 may use the two farthest points on the best-fit line as a representation of the aimed direction (the line connecting these points follows the path on the aimed direction).
[0046] In some embodiments, the imaging module 160 may receive image data associated with a single image that includes information to determine both a target direction and an aimed direction. The imaging module 160 may be configured to compare the target direction and the aimed direction in the image data and determine an alignment value, such as an offset angle between the two directions. For example, the imaging module 160 may perform image processing to determine an offset angle between the two directions found in the single image. In another example, the imaging module 160 may associate an orientation of the image with the target direction (e.g., the vertical field of view direction is the target direction). In an exemplary embodiment, the imaging module 160 is configured to use vector algebra to calculate the angle between the target direction (e.g., the saved points along the best fit line associated with the target direction) and the aimed direction (e.g., the saved points along the best fit line associated with the aimed direction).
[0047] In other embodiments, the imaging module 160 may extract a target direction and an aimed direction from different images. In these embodiments, the imaging module 160 may work with the calibration module 162 to maintain positional consistency across the images. For example, the imaging module 160 may first determine and store a target direction and await information about an aimed direction. The calibration module 162 may be configured to receive an aimed direction, recall the stored target direction, and compare the two directions to determine an alignment value, such as an offset angle. The imaging module 160 and/or calibration module 162 may determine an angle through which the aimed direction is rotated from the target direction to determine the alignment value. In an exemplary embodiment, the imaging module 160 is configured to use vector algebra to calculate the angle between the target direction (e.g., the saved points along the best fit line associated with the target direction) and the aimed direction (e.g., the saved points along the best fit line associated with the aimed direction). The calibration module 162 may use the camera field of view as a common reference and overlay the aimed direction onto the target direction. The imaging module 160 and/or the calibration module 162 may be configured to send information, such as alignment values, to the user interface 140 for display to a user.
[0048] The analysis module 164 may be configured to perform one or more additional processes to enhance the functionality of the alignment measuring tool 100. For example, the analysis module 164 may receive an alignment value related to the placement of a golf ball and determine the maximum length of putt that would be made if the ball were to be directed along the aimed direction. In another example, the analysis module 164 may be configured to compare results from multiple measurements in order to provide feedback regarding which alignments aids provide the best results. For example, the analysis module 164 may compare average alignment values across multiple tests to identify alignment aids that are more often or more closely aligned with the target direction. In particular, in some embodiments, the analysis module is configured to calculate the dispersion at various distances to the target point using the alignment value (e.g., the offset angle) to predict whether the putt will be successful.
[0049] The computing system 110 may be configured to perform other processes via the same or different modules than those described herein. For example, the computing system 110 may be configured to perform pattern matching to identify an alignment aid on a golf ball in at least one image and match the alignment aid to a known or stored alignment aid in a memory or database. The computing system 110 may associate a subsequent measurement with the stored alignment aid to enable the user interface 140 to display measurements associated with that particular alignment aid. Further, the computing system 110 may use stored information about alignment aids in determining the predominant axis of the alignment aid and/or for determining the aimed direction. For example, the stored information may identify how the predominant axis relates to the alignment aid.
[0050] In another example, the computing system 110 may be configured to perform one or more machine learning processes to augment and/or substitute one or more image processing and recognition processes. For example, the computing system 110 may be trained on images of alignment aids with a ground truth defining the predominant axis and/or aimed direction associated with a particular alignment aid. The computing system 110 may use the machine learning tools to help accurately determine aimed and/or target directions from images.
[0051]
[0052] In an exemplary embodiment, the support structure 130 includes a camera link 170, a light control 172, and a power supply 174. The camera link 170 may be, for example, a support cradle for the camera 120 to direct the aperture of the camera 120 at a target area. In another example, the camera link 170 may be an opening for providing a view of a placement position to a camera 120 located separate from the support structure 130. The support structure 130 may include the light control 172 and/or power supply 174 to help enhance the quality of images captured by the camera 120. The light control 172 may be, for example, an enclosure for blocking out sunlight that could flood images captured by the camera 120. In some embodiments, the light control 172 may include a light source, such as a light bulb, flash, or LED. The light control 172 may be powered by the power supply 174 to sufficiently illuminate the field of view of the camera 120. In some embodiments, the power supply 174 may be a battery pack.
[0053] For embodiments in which the target direction and aimed direction are determined from different images, the alignment measuring tool 100 may require a frame of reference to be established across images. Accordingly, the support structure 130 may include a calibration element 176 that provides consistency to the placement of the support structure 130 and/or camera 120. For example, the calibration element 176 may include positional markers that maintain the location of the support structure even if it is removed and replaced. In another example, the calibration element 176 may include a feature that detects a position and/or orientation of the alignment measuring tool itself, in order to allow for adjustments based on position. In some embodiments, the support structure 130 may include components that enable the user to efficiently and easily place and replace the support structure 130, such as telescoping handles to easily pick up and replace the support structure 130.
[0054] A disclosed alignment measuring tool may be used in a process to assess the alignment of an alignment aid placed by a golfer on a placement position. In general, a setup for the measurement may include a flat portion of a putting green with a typical hole/cup in the green selected as the target point (i.e., the center of the hole being the target) and a placement position some distance from the hole (e.g., 5 ft, 10 ft, 15 ft, etc.). The intention of the measurement may be to assess how a golfer places a golf ball on the placement position such that the alignment aid is aimed directly at the target point. A placement position may be marked on the green and the golfer asked to place a golf ball with the alignment aid pointed toward the target point. The alignment measuring tool 100 may thereafter capture one or more images of the placed golf ball and determine an alignment value associated with the placement. The process may be repeated for multiple alignment aids and multiple placement positions in order to collect multiple data points to compare and analyze the golfer's ability to aim different alignment aids from various distances.
[0055] In determining a target direction for a particular measurement, the alignment measuring tool 100 may be configured to capture an image of a calibration line. The calibration line may be a visual object, marking, or other visual indicator that is placed in the field of view of the camera 120 in order to represent a direct path from the placement position to the target point for that measurement. For instance, a linear marking such as a rod, string, or tape may be placed on a line from the target point to the placement position. In a particular example, a string may be tied to a flag stick (which is placed at the center of the cup) and pulled to the placement position to represent the path therebetween in an image to be captured by the camera 120. In another example, a laser line may be used. For instance, a single laser line may be projected between the target point and the placement position. In another embodiment, one or more calibration lines (e.g., laser lines) may be positioned along a line or lines parallel to the line connecting the target point and placement position. In this way, the calibration line(s) can be spaced from a placed golf ball such that both the calibration line(s) and the alignment aid can be compared within the same field of view of the camera 120.
[0056]
[0057] In the example of
[0058] In an exemplary embodiment, the calibration lines 202 may be selectively removed from the field of view 208 so that they are not present when the golfer places the golf ball 206. For instance, the calibration lines 202 may be produced by a pair of laser devices that can be turned off while the golfer is placing the golf ball and turned back on prior to the measurement. For instance, the calibration lines 202 may be a pair of laser devices that are symmetrically attached to opposing sides of a flag stick and used to project the calibration lines 202 back toward the placement position 212.
[0059]
[0060] After the golf ball has been placed and one or more calibration lines are in position (e.g., as shown in the field of view 208 of
[0061] The computing system 110 may then perform a calculation to determine an alignment value associated with the placed golf ball (Step 228). For example, the computing system 110 may compare the aimed direction and the target direction to determine an offset angle between the two directions. The offset angle represents the degree to which the alignment aid is or is not aimed at the target point. For example, the computing system 110 may set the target direction to be at an angle of 0 and use geometry to determine the angle of the aimed direction in relation to the target direction. The computing system 110 may use negative and positive angle values to represent offset angles that are rotated in clockwise and counter-clockwise directions, respectively. For example, an offset angle of 0.3 may represent that the alignment aid is directed 0.3 off to the right of the target point. An offset angle of 0.3 would represent that the alignment aid is directed 0.3 off to the left of the target point.
[0062] The alignment measuring tool 100 may provide the calculated alignment value to the user interface 140 for display to a user (Step 229), thereby providing feedback to the user regarding the placement of the golf ball and alignment aid. In some embodiments, the computing system 110 may analyze calculated alignment values to provide additional information to the user. For example, the computing system 110 may associated alignment values with particular alignment aids and different distances from the target point and provide a suggestion to the user of the alignment aid pattern that was most accurately placed with respect to the target point. In another example, the computing system 110 may determine the further distance that a putt would be made based on the calculated offset angle.
[0063] In a second embodiment of a measuring process, the alignment measuring tool 100 may use a common frame of reference to compare information extracted from multiple captured images. In this embodiment, the alignment measuring tool 100 may store a target direction and separately determine an aimed direction to be compared to the target direction.
[0064]
[0065] In an exemplary embodiment, the computing system 110 may determine a target direction for a given setup of a placement position and target point from a first image (e.g., a calibration image) (Step 252). The target direction represents a straight line from the placement position to the target point. A golfer may place a golf ball on the placement position with the intention of directing the alignment aid toward the target point (e.g., toward the center of the hole). When the golf ball has been placed, the camera 120 may capture an image of the placed golf ball (Step 254) The camera 120 may provide image data associated with the captured image to the computing system 110.
[0066] The computing system 110 may use the image data associated with the captured image to calculate an aimed direction of the alignment aid on the placed golf ball (Step 256). For example, the imaging module 160 may use a best fit analysis to determine the aimed direction. The computing system 110, having received the target direction and aimed direction, may thereafter determine an alignment value (Step 258). For example, the computing system 110 may compare the target direction to the aimed direction to determine an offset angle between the two directions. In some embodiments, the computing system 110 may perform a calibration correction to correct any differences between the orientation and/or field of view of the multiple images when determining the alignment value. After the alignment value has been determined, it may be provided to a user via the user interface 140 (Step 259).
[0067]
[0068] The computing system 110 may receive image data associated with the calibration line and perform image processing to determine the target direction. For example, the computing system 110 (e.g., the imaging module 160) may perform image processing to find a best fit line within an image that matches the calibration line (Step 308). For example, the computing system 110 may detect pixels within the image data whose intensity have the same line direction, thereby matching the calibration line in the image. The computing system 110 may identify a contour within the image based on the matching-intensity pixels and select two or more points on the contour as a representation of the best fit line. For example, the computing system 110 may select two points on the contour that are a furthest distance from each other as end points of the calibration line in the image. In another embodiment, the computing system 110 may select at least three points on the contour (e.g., the two further points and a point in-between). A line connecting the selected points thus would match the calibration line. In this way, the two or more selected points represent a determination of the target direction.
[0069] The computing system 110 may store the identified best fit line as the target direction (Step 310). The computing system 110 may associate the stored target direction with a particular measurement, frame of reference, field of view, etc. in order to enable the stored target direction to be appropriately recalled and compared to an aimed direction. In some embodiments, the computing system 110 may use the target direction to compare to an aimed direction determined from the same image.
[0070]
[0071] The computing system 110 may determine a predominant axis associated with the alignment marking (Step 326). For example, the computing system 110 may perform a best fit analysis to identify a line that best matches the predominant axis of the alignment marking as it appears in the captured image.
[0072] The computing system 110 may determine a best fit line for the aimed direction based on an alignment aid in a manner similar to that described in relation to determining the target direction based on a calibration line. For example, the computing system 110 may detect pixels in the image within the boundary of the golf ball whose intensity have the same line direction, thereby matching the alignment aid in the image. The computing system 110 may identify a contour within the image based on the matching-intensity pixels and select two or more points on the contour as a representation of the best fit line. For example, the computing system 110 may select two points on the contour that are a furthest distance from each other as end points of the alignment aid in the image. In another embodiment, the computing system 110 may select at least three points on the contour (e.g., the two further points and a point in-between). A line connecting the selected points thus would match the direction of the alignment aid. In this way, the two or more selected points represent a determination of the aimed direction. The computing system 110 may store the aimed direction as the direction of the predominant axis within the field of view of the capture image (Step 328).
[0073]
[0074] The alignment measuring tool 350 includes a computing device 372, a support structure 374, and a camera 376 operably connected to the computing device 372. The computing device 372 may be a computing system as described in the present disclosure. The support structure 374 is configured to support the computing device 372 and the camera 376. The camera 376 is configured to capture images of a portion of the surface 355 within a field of view of the camera 376. For example, the camera 376 may be configured to capture images of a portion of the surface 355 directly beneath the camera 376.
[0075] The camera 376 may be configured to capture high quality images within the field of view. In an exemplary embodiment, the camera 376 may be a Basler ACA1600-20 gm. The camera 376 may be positioned relative to the support structure 374 and the surface 355 such that the vertical field of view of the camera is 2-4 inches. This size field of view helps ensure a large number of pixels per inch relative to the golf ball being imaged while allowing enough visible space to make it easy to find the golf ball in the camera's field of view when placing the alignment measuring tool 350 on the surface 355. The camera 376 may use a lens and mounting distance to provide the desired size field of view. For example, the camera 376 may use a 35 mm lens and be positioned to be 14 to 27 inches from the surface 355. In another embodiment, the camera 376 may have a 50 mm lens and be mounted 20 to 39 inches from the surface 355. In another embodiment, the camera 376 may have a 16 mm lens and be mounted 6.5 to 12 inches from the surface 355.
[0076] In other embodiments, the computing device 372 may include an integrated camera 376. For example, the computing device 372 may be a mobile device such as a tablet and/or mobile phone. In these embodiments, the integrated camera 376 may also provide a similar field of view (e.g., 2-4 inch vertical field of view) to ensure high quality image capture of the golf ball 362 and alignment aid 364.
[0077] The computing device 372 may be connected to a user interface. For example, the computing device 372 may include an integrated I/O device (e.g., a display, touchscreen, etc.) for presenting a user interface. In some embodiments, the computing device 372 is configured to connect to another computing device (e.g., a network connection to a mobile device) that has a user interface. In some embodiments, the computing device 372 may be removable from the support structure 374.
[0078] The support structure 374 may include a plurality of support walls 378 the create an enclosure for surrounding a portion of the surface 355. The alignment measuring tool 350 may further include a flash 380 configured to provide a light source within the enclosure of the support structure 374 when the camera 376 captures an image. The support structure 374 and the flash 380 thereby provide light control for images captured by the camera 376 to help ensure sufficient quality images for assessing orientation of an alignment aid.
[0079] In some embodiments, the alignment measuring tool 350 may further include a calibration feature configured to provide feedback on a position of the camera 376 relative to an area outside of the field of view of the camera 376. For example, the calibration feature may align the camera 376 relative to the target point 360. The calibration feature may provide feedback on the position of the camera 376 indirectly by providing information about the position of the support structure 374 or other component of the alignment measuring tool 350, relative to the target point 360.
[0080] In one example, the calibration feature comprises a laser line 381 projecting from the alignment measuring tool 350 toward the target point 360. The laser line 381 may be pointed toward an object at the target point 360, such as a flag stick. The laser line 381 can thus be used to control a position and orientation of the alignment measuring tool 350 relative to the target point 360 by providing feedback about how the camera 376 is positioned relative to the target point 360.
[0081] In another embodiment, the calibration feature includes a second camera that has a field of view outside of the support structure 374. For example, the second camera may be used to detect the target point. The calibration feature may further include a motor configured to rotate the first camera 376 to align it with the target point detected by the target point.
[0082] In some embodiments, the alignment measuring tool 350 may include positional markers 382. The positional markers 382 may include a first component 384 configured to be securely connected to the surface 355 and a second component 386 connected to the support structure 374. The first component 384 may be configured to embed flush with the surface 355 and includes a connector configured to mate with a corresponding connector of the second component 386. For example, the first component 384 and the second component 386 may include pin and socket, magnetic, snap, or any other type of connection configured to all the positional markers to repeatedly detach and reattach in the same location.
[0083] The positional markers 382 help to maintain a position and orientation of the camera 376 across multiple images even when the alignment measuring tool 350 is moved away from the placement position 366 in between images. In other words, the positional markers 382 allow the support structure 374 to be picked up off of the surface 355 and replaced in the same position and orientation. As such, the alignment measuring tool 350 may be calibrated to a particular location (e.g., target direction determined), the calibration lines may be removed, and a golfer allowed to place a golf ball with the alignment aid pointing toward a target point. The support structure 374 may thereafter be replaced on the positional markers 382 to ensure consistency of field of view across images captured by the camera 376.
[0084]
[0085] The computing device 372 may be positioned on top of the support structure such that a user interface (e.g., a touch screen) 392 is facing outward toward a user, thereby enabling the user to interact with the computing device 372 while it is on the support structure 374. A rearward facing camera of the computing device 372 points toward the inside of the support structure 374 and thus is configured to capture an image of a surface enclosed by the support structure 374 and any objects (e.g., calibration lines, golf balls) placed on the surface in that area. The computing device 372 may be configured to output a preview image of the field of view of the camera on the user interface 392. In this way, a user can position the support structure 374 over a golf ball and view a preview of an image to be captured. The user can provide input to the computing device 372 to cause the camera to capture an image.
[0086] As shown in
[0087]
[0088] The alignment measuring tool 350 may be particularly well suited to help simulate a real golf experience when a golfer is placing a golf ball for an alignment assessment. The alignment measuring tool 350 can be initially placed over a calibration line and an image captured to determine a target direction. The alignment measuring tool 350 and the calibration line can then be removed from the space, allowing a golfer to step up and place a golf ball on the placement position with the goal of aligning the alignment aid with the target point. The alignment measuring tool 350 can then be replaced at the same position (e.g., via the positional markers 382) over the placed golf ball. The alignment measuring tool 350 can thereafter capture a high quality image of the placed golf ball due to the enclosure of the support structure 374 and the interior light control elements (e.g., flash). The process can be repeated for multiple alignment aids and various distances from the target point to collect robust data to characterize a user experience with aligning golf balls to a target.
[0089]
[0090] The support structure 424 maintains the camera 426 in a single position so that the field of view of the camera 426 does not change across captured images. For example, the support structure 424 may be securely supported on and/or attached to the surface 405 to inhibit movement of the camera 426. In some embodiments, an enclosure 428 may be placed over the golf ball 415 to help control light exposure for the camera 426. An opening 432 in the enclosure 428 provides an aperture for the camera 426 to capture images. A flash or light source may be connected to the support structure 424, camera 426 and/or enclosure 428 to further enable the capture of high quality images by the camera 426.
[0091] In accordance with the disclosure, the alignment measuring tool 400 may be configured to capture images within the field of view of the camera 426, identify an alignment aid within the image(s), and compare an aimed direction of the alignment aid to a target direction. In some embodiments, the alignment measuring tool 400 may compare a target direction extracted from a first image to an aimed direction extracted from a second image, as has been described in the present disclosure. In other embodiments, the alignment measuring tool 400 may compare a target direction and aimed direction extracted from the same image, including situations in which the target direction is extrapolated from the orientation of the field of view. For example,
[0092] The calibration device 430 may be configured to be turned on to project the laser line 435 onto the surface 405 within the field of view of the camera 426. The alignment measuring tool 400 is configured to capture an image of a portion of the laser line 435 and use the image to determine a target direction for the measurement. In some embodiments, the calibration device 430 may be initially turned off, a golfer allowed to place the golf ball 415 on the placement position 420, and then the calibration device 430 turned on so that the laser line 435 and golf ball 415 are present simultaneously in the field of view of the camera 426. The camera 426 may thereafter capture an image of the golf ball 415 and the laser line 435 and perform image processing to determine an alignment value related to the alignment aid on the golf ball 415.
[0093] Disclosed embodiments include alignment measuring tools configured to capture images of a surface, such as a putting green, to determine an aimed direction of an alignment aid of a golf ball and calculate an alignment value associated with the aimed direction's relationship to a target direction. In some embodiments, the alignment measuring tool calculates a target direction as a reference to which the aimed direction may be compared. In some embodiments, the alignment measuring tool may include a calibration element configured to ensure that the alignment measuring tool is consistently in the same orientation.
[0094] In some embodiments, the alignment measuring tool may be positioned such that an aspect of a captured image may be reliably associated with the target direction. For example, the alignment measuring tool may include a calibration element that aligns the vertical field of view direction of the camera with the target direction. The alignment measuring tool may thereafter use the orientation of the image to compare an aimed direction to a target direction and determine an alignment value. In other embodiments, the alignment measuring tool may include a calibration element configured to determine an orientation and/or position in space to correct for any deviations or movements, such that the field of view of the camera does not need to be aligned with the target direction as long as the relationship between the two is known.
[0095]
[0096] In an exemplary embodiment, the alignment measuring tool 450 is configured to use the laser lines 460 to determine an orientation of the alignment measuring tool 450 relative to the target direction. For example, in
[0097] While it is apparent that the illustrative embodiments of the invention disclosed herein fulfill the objectives stated above, it is appreciated that numerous modifications and other embodiments may be devised by those skilled in the art. Therefore, it will be understood that the appended claims are intended to cover all modifications and embodiments which would come within the spirit and scope of the present invention.