System and method for providing and/or collecting information relating to objects

11568725 · 2023-01-31

Assignee

Inventors

Cpc classification

International classification

Abstract

A system for providing and/or collecting information relating to at least one object is provided. The system comprises a radio-frequency identification, RFID, tag that is provided proximate to the at least one object, the RFID tag being configured to enable detection of a position and/or a movement of the at least one object; a server device configured to receive position and/or movement data about the at least one object; and a mobile device configured to: identify the at least one object; send a request to the server device for object information about the identified at least one object; wherein: the server device is further configured to, in response to the request, provide to the mobile device targeted object information about the identified at least one object, the targeted object information being at least partially based on the position and/or movement data; and the mobile device is further configured to display the targeted object information.

Claims

1. A system comprising: a processor; and a memory coupled to the processor, the memory encoding one or more programs to cause the processor to: detect a position and/or a movement of at least one object of a plurality of objects located in an environment based on a radio-frequency identification tag placed proximate to each of the at least one object of a plurality of objects that enables detection of the position and/or the movement of the at least one object; receive position and/or movement data about the detected at least one object; process the position and/or movement data of the at least one object to obtain values of aggregate indicators relative to a predefined duration of time, the position and/or movement data of the at least one object being detected over the predefined duration of time and combined to obtain the values of the aggregate indicators; receive a request for object information about the identified at least one object; and provide, in response to the request, targeted object information about the identified at least one object to the mobile device at least partially based on the position and/or movement data and data about at least one secondary object other than the identified at least one object.

2. The system of claim 1, wherein the least one secondary object is related to the identified at least one object by a complementary use.

3. The system of claim 1, wherein the data about the at least one secondary object includes a location of the secondary object in the environment.

4. The system of claim 1, wherein the data about the at least one secondary object is not based on the position and/or movement data.

5. The system of claim 1, wherein the position and/or movement data comprise an indication that the at least one object is located in a check-out area.

6. The system of claim 5, wherein at least one of the targeted object information about the identified at least one object and the data about the at least one secondary object includes at least one indication of a check-out action.

7. The system of claim 1, wherein the radio-frequency identification tag further provides a theft prevention function and the memory encodes one or more programs to cause the processor to disable, when payment of the at least one object has been made, the theft prevention function of the radio-frequency identification tag.

8. The system of claim 1, wherein the targeted object information comprises a route from the at least one object to the secondary object.

9. A computer-implemented method for providing and/or collecting information relating to at least one object, the method comprising: detecting a position and/or a movement of at least one object of a plurality of objects located in an environment based on a radio-frequency identification tag placed proximate to each of the at least one object of a plurality of objects that enables detection of the position and/or the movement of the at least one object; receiving position and/or movement data about the detected at least one object; processing the position and/or movement data of the at least one object to obtain values of aggregate indicators relative to a predefined duration of time, the position and/or movement data of the at least one object being detected over the predefined duration of time and combined to obtain the values of the aggregate indicators; receiving a request for object information about the identified at least one object; and providing, in response to the request, targeted object information about the identified at least one object to the mobile device at least partially based on the position and/or movement data and data about at least one secondary object other than the identified at least one object.

10. The method of claim 9, wherein the least one secondary object is related to the identified at least one object by a complementary use.

11. The method of claim 9, wherein the data about the at least one secondary object includes a location of the secondary object in the environment.

12. The method of claim 9, wherein the data about the at least one secondary object is not based on the position and/or movement data.

13. The method of claim 9, wherein the position and/or movement data comprise an indication that the at least one object is located in a check-out area.

14. The method of claim 13, wherein at least one of the targeted object information about the identified at least one object and the data about the at least one secondary object includes at least one indication of a check-out action.

15. The method of claim 9, wherein the radio-frequency identification tag further provides a theft prevention function and the memory encodes one or more programs to cause the processor to disable, when payment of the at least one object has been made, the theft prevention function of the radio-frequency identification tag.

16. The method of claim 9, wherein the targeted object information comprises a route from the at least one object to the secondary object.

17. A non-transitory computer-readable medium storing program instructions that when executed cause a processor to perform a method comprising: detecting a position and/or a movement of at least one object of a plurality of objects located in an environment based on a radio-frequency identification tag placed proximate to each of the at least one object of a plurality of objects that enables detection of the position and/or the movement of the at least one object; receiving position and/or movement data about the detected at least one object; processing the position and/or movement data of the at least one object to obtain values of aggregate indicators relative to a predefined duration of time, the position and/or movement data of the at least one object being detected over the predefined duration of time and combined to obtain the values of the aggregate indicators; receiving a request for object information about the identified at least one object; and providing, in response to the request, targeted object information about the identified at least one object to the mobile device at least partially based on the position and/or movement data and data about at least one secondary object other than the identified at least one object.

18. The medium of claim 17, wherein the least one secondary object is related to the identified at least one object by a complementary use.

19. The medium of claim 17, wherein the data about the at least one secondary object includes a location of the secondary object in the environment.

20. The medium of claim 17, wherein the data about the at least one secondary object is not based on the position and/or movement data.

Description

BRIEF DESCRIPTION OF THE DRAWINGS

(1) Details of exemplary embodiments are set forth below with reference to the exemplary drawings. Other features will be apparent from the description, the drawings, and from the claims. It should be understood, however, that even though embodiments are separately described, single features of different embodiments may be combined to further embodiments.

(2) FIG. 1 shows an exemplary setup of a system according to the present disclosure.

(3) FIG. 2 shows an example of the system according to the present disclosure.

(4) FIG. 3 shows an exemplary functional block diagram of the system according to the present disclosure.

(5) FIG. 4 shows a flowchart of an exemplary processing for providing and/or collecting information relating to one or more objects.

(6) FIG. 5 shows a flowchart of another exemplary processing for providing and/or collecting information relating to one or more objects.

(7) FIG. 6 shows a flowchart of yet another exemplary processing for providing and/or collecting information relating to one or more objects.

(8) FIG. 7 shows an exemplary functional block diagram of the software layers according to the present disclosure.

DETAILED DESCRIPTION

(9) In the following, a detailed description of examples will be given with reference to the drawings. It should be understood that various modifications to the examples may be made. Unless explicitly indicated otherwise, elements of one example may be combined and used in other examples to form new examples.

(10) FIG. 1 shows an exemplary setup of a system according to the present disclosure. In the exemplary setup shown in FIG. 1, objects O1, O2, . . . (hereinafter, collectively referred to as “objects O” or, in case of referring to one of the objects, as “object O”) are provided at different physical locations in a certain environment. In this example, the environment may be at least part of a shop or store or showroom where certain products are sold. Further, in this example, the objects O may be the products sold in the environment.

(11) In the example of FIG. 1, tags T1, T2, . . . (hereinafter, also collectively referred to as “tags T” or, in case of referring to one of the tags, as “tag T”) are provided proximate to the objects O. For instance, at least one of the tags T may be attached and/or linked in any other way (e.g. with fastening means) to the corresponding object O.

(12) In some examples, the tags T may be RFID (radio-frequency identification) tags. In some cases, the RFID tags may be passive RFID tags that are activated by radio energy transmitted by an RFID reader. The tags T can enable detection of a position and/or a movement of the object O to which the tag T is affixed.

(13) The RF technology of the RFID tag can be used for position detection, but alternatively or additionally a Bluetooth beacon such as Puck.js may be integrated in the tag T. If triangulation and/or trilateration is used to determine the position of the tag T and, consequently, of the object O to which the tag T is affixed, at least three detection devices 30 may be part of the system. Depending on the size of the environment and the range of the technology used for position detection, a plurality of triplets of detection devices 30 may be used, wherein triplets may have common elements. The detection devices 30 may be RFID readers and/or Bluetooth devices that can receive Bluetooth signals, e.g. Bluetooth Low Energy (BLE) signals.

(14) In order to allow for movement detection, the tag T may comprise a magnetometer, which is also part of the Bluetooth beacon Puck.js, and/or a gyroscope and/or an accelerometer.

(15) It should be noted that the number of the objects O and the corresponding tags T provided in the environment may be more than or less than that shown in FIG. 1. Further, the locations of the objects O and the optional detection devices 30 as shown in FIG. 1 serve only illustrative purposes and do not necessarily correspond to real locations.

(16) In the example of FIG. 1, the user U (e.g., a customer) may carry a mobile device 10. The mobile device 10 may be configured to identify at least one object O. In case the user U is interested in an object O, the user U may instruct the mobile device 10 to obtain information corresponding to the identified object O. The mobile device 10 may send a request for information to a server device 20 via a wireless network (not shown).

(17) The server device 20 may be provided within the environment or at a different location outside the environment, e.g. remotely from the environment. The server device 20 may comprise one or more server applications may be configured to receive, process and/or provide the data concerning the one or more objects O. The server device 20 may be in communication with an object information database 40 storing object information. The server device 20 may further communicate via a wireless network with the tags T and/or the detection devices 30, in order to receive position and/or movement data. The server device 20 may store and/or analyse the position and/or movement data.

(18) The server device 20 may retrieve targeted object information from the object information database about an object O in response to the request of the mobile device 10, wherein the targeted object information is chosen at least partially based on the position and/or movement data. Alternatively or additionally, the server device 20 may generate targeted object information. The server device 20 may then transmit the data constituting the targeted object information to the mobile device 10.

(19) Exemplarily, the server device 20 may be run via a cloud provider e.g. by using a platform like Kyma.

(20) The mobile device 10 may display the targeted object information received from the server device 20, for instance by using augmented reality.

(21) FIG. 2 shows an example of the system according to the present disclosure. As also stated above with reference to FIG. 1, the system shown in FIG. 2 may comprise the tags T, the mobile device 10 and the server device 20. The mobile device 10, the tags T and the server device 20 may be connected via the network 60. The network 60 may include the Internet and/or one or more intranets. Further, at least part of the network 60 is implemented by a wireless network (e.g. wireless local area network (WLAN), cellular network, etc.).

(22) The exemplary system shown in FIG. 2 may further comprise detection devices 30, an object information database (DB) 40 and an additional mobile device 50, which may be connected to the network 60. The details of these devices will be described later below with reference to FIG. 3.

(23) FIG. 3 shows an exemplary functional block diagram of the system according to the present disclosure. As shown in FIG. 3, the mobile device 10 may comprise a identification unit 100, a communication unit 102, input unit 104, output unit 106 and a control unit 108.

(24) The communication unit 102 may be an interface for communicating with other devices such as the server device 20. The communication unit 102 may be implemented by, for example, a network interface that enables communication via a wireless network, such as Bluetooth or Wi-Fi.

(25) The identification unit 100 may be configured to identify an object O. The identification unit 100 may be implemented by a device that is suitable for identifying an object O, depending on the technology used for the identification. For example, the identification unit 100 may be a camera configured to capture an image of the object O. In other examples, the identification may be based on the tag T affixed to the object O. For instance, the tag T may comprise computer-readable code (such as barcode or QR code etc.) and/or human-readable code, so that the identification unit 100 may be a camera or another reader device (e.g. QR code reader or optical character recognition (OCR) reader). In a further example, if the tag T has Bluetooth functionalities, the identification unit 100 may be part of the communication unit 102. The actual identification may further require specific software.

(26) The input unit 104 may be configured to receive inputs from the user. The input unit 104 may comprise, for example, one or more switches (e.g., buttons) corresponding to one or more possible user instructions, which may be implemented by a touch panel displaying the switches. For instance, one of the switches may correspond to a user instruction to request object information, to purchase an object, to start navigation towards an object and other actions, some of which will be further described below. In particular, the input unit 104 may be the touch screen of a smartphone.

(27) The output unit 106 may be configured to provide outputs. The output unit 106 may comprise, for example, a haptic output device, an audio output device and/or a display device. In some examples, the output unit 106 may provide an output presenting the received targeted object information, e.g. displaying the information on a display device. For instance, if the mobile device 10 provides real-time guidance along a route, in case the output unit 106 comprises a haptic output device (e.g., vibration motor), the output unit 106 may output different haptic effects to lead the user. Further, in case the output unit 106 comprises an audio output device, the output unit 106 may output different audio outputs providing indications about the route (e.g. ‘turn left’).

(28) The control unit 108 may be configured to control the mobile device 10. The control unit 108 may be implemented by a processor and a memory, for example. The control unit 108 may be configured to control the identification unit 100, the communication unit 102 and the output unit 106 of the mobile device 10 to perform processing as described herein.

(29) In some examples, the mobile device 10 may be implemented by a smartphone or a tablet computer, with one or more software applications that cause the device to perform the processes as described herein.

(30) The additional mobile device 50 may include any of the elements described with reference to the mobile device 10.

(31) The object information DB 40 may store object information for each object. The object information may be information relating to an object and may include but is not limited to, a default location of the object, rest in stock of the object, possible delivery dates of the object, promotions relating to the object, offerings relating to the object, cross sells relating to the object, bundles of the objects, etc.

(32) The server device 20 may be implemented by a server computer. The server device 20 may be configured to provide, in response to a data request by the mobile device 10 and/or the additional mobile device 50, the information relating to the one or more objects of interest.

(33) In some examples, the server device 20 may provide further application(s) which uses the information relating to the object(s) of interest selected by the user. For example, the server device 20 may provide an online shopping website from which the user may place purchase orders of the object(s) of interest selected by the user.

(34) The tags T may be conventional security tags for products in a store, such as RFID tags, enhanced with additional functionalities that turn the products into self-managed products. These additional functionalities may include, but are not limited to, position and/or movement detection and communication of position and/or movement data.

(35) The detection devices 30 may be devices that interact with the tags T e.g. in order to determine a position of the tags T and, consequently, of the objects O. The detection devices 30 may also interact with the mobile device 10, in particular with the communication unit 102, in order to determine a position of the mobile device 10 and, consequently, of a user U holding the mobile device 10.

(36) It should be noted that the system described above with reference to FIGS. 1 to 3 provides a mere example of the system according to the present disclosure and the system may have different configurations than the one shown in FIGS. 1 to 3.

(37) FIG. 4 shows a flowchart of an exemplary processing for providing and/or collecting information relating to one or more objects.

(38) An example of an object O may be a camera that is provided at 410 with a tag T that comprises a motion sensor that can sense whether the object O and the tag T are moved, e.g. lifted from a shelf in a shop. The motion sensor may be a magnetometer and/or an accelerometer and/or a gyroscope. Further, the tag T may comprise a communication unit to transmit movement data. Exemplarily, the tag T may transmit movement data when a movement is detected.

(39) A user U, such as a customer, may take the camera O off the shelf, thus triggering the detection of the movement (step 420) by the motion sensor in the tag T and the transmission of the movement data to the server device 20, which receives the data at 430. The server device 20 may store, keep track and analyze the movement data (e.g. using a software like Qualtrics), therefore recognizing that the camera has been lifted for the first time in a long time, e.g. no movement events have been detected over a predefined amount of time, such as 24 hours or seven days.

(40) The customer U may be interested in learning more about the camera O and may use a mobile app on their mobile device 10 to check the details of the camera. For example, the customer U may capture a picture of the camera or its box with the mobile device 10 and a picture-recognition process may be performed, leading to the identification of the camera O at 440. The mobile device 10 may then send a request for information about the camera to the server device 20 at 450.

(41) Once the server device 20 receives the request of the mobile device 10, based on the fact that the camera has been lifted for a first time in a long time, it may decide to offer a discount for the camera to the customer and so it generates a targeted piece of information about a 10% discount on the camera price, providing it to the mobile device 10 at 460.

(42) The mobile device 10 may display a notification about the discount to the customer U at 470, wherein the notification may comprise an interactive feature, such as a button, that enables the customer to accept the discount.

(43) Additionally or alternatively, at 460 the server device 20 may retrieve from the object information database 40 a list of accessories related to the camera, such as an SD card and a camera bag. The customer U may browse through the list displayed at 470 and select e.g. the SD card, upon which the server device 20 may inform the customer through the mobile device 10 at 470 that the SD card is currently not available in the store, and offer the customer U at 470 the opportunity of buying it online or ordering it and picking it up at the store in the next days. The customer U may select one of the options via a GUI on the mobile device 10.

(44) The customer U may additionally or alternatively select the camera bag and the targeted object information provided at 460 and displayed at 470 may be a route to the camera bag, with an option for guiding the customer U to the location of the camera bag in the store. The customer may accept the offer and real-time guidance along a route to the shelf where the bag is located may be provided. For example, the mobile device 10 may capture pictures/video of the environment, i.e. the shop, while the customer U is moving among the shelves and use augmented reality to show the route e.g. as a line with arrows changing direction according the user movements.

(45) Further, the server device 20 may provide at 460, while the route is being displayed at 470, additional information about objects along the route, such as special offers, which are also displayed (at 470).

(46) FIG. 5 shows a flowchart of another exemplary processing for providing and/or collecting information relating to one or more objects. The processing shown in FIG. 5 may be combined with the processing shown in FIG. 4 or may be independent from it.

(47) An object O, such as a camera, may be provided at 510 with a tag T enabling detection of position, e.g. an RFID tag T. Further, the RFID tag may be used as an anti-theft measure.

(48) A customer U may have considered buying the camera O and used their mobile device 10 to receive more information about the camera O, identifying the camera O at 520 e.g. by capturing an image, and sending a request for information about the camera O, such as technical specs, to the server device 20 at 530. The information may have been provided and displayed as discussed with reference to FIG. 4 above or according to other examples.

(49) The customer U may then take the camera O, with the intention of purchasing it, and makes their way to the check-out counters. At least one RFID reader 30 may be located in proximity of a check-out area and may detect at 540 that the camera O is in the check-out area. The RFID reader 30 may communicate the position data relative to the camera O to the server device 20, which receives it at 550.

(50) Based on the fact that the camera O is in the check-out area, the server device 20 may offer the customer U the possibility to check out by paying via the mobile device 10. Accordingly, at 560 the server device 20 may provide information relative to the check out and at 570 the mobile device 10 may display such information, which may include e.g. a link to an app for paying.

(51) Once the customer U has paid, the mobile device 10 may disable the RFID tag so that the customer U can leave the store without triggering any theft alarms. For example, the RFID tag may be a RapiTag.

(52) In the exemplary cases discussed above, information can be effectively collected and/or provided in the context of a real-world shopping experience. In particular, the collection of position and/or movement data is made effective thanks to the use of “smart” tags affixed to the products. Further, the provision of information about products is made effective by leveraging the collected position and/or movement data. The interplay between the mobile device 10, the server device 20 and the tags T makes the system efficient in terms of resource consumption and time. In particular, the real-time guidance to complementary items (e.g. batteries) is a guided human-machine interaction process that assists the user U in assembling a functioning product. The self-check-out prompted by a position of the items in a check-out area makes the check-out procedure more efficient and secure.

(53) FIG. 6 shows a flowchart of yet another exemplary processing for providing and/or collecting information relating to one or more objects.

(54) Retailers may not know what customers are looking at in their shops or showrooms. The retailers do not always have insights in how customers are experiencing the product assortment and what forms the interest of the customers. Further, the retailers might not have means to obtain real-time analytics data in a physical shop, store or showroom. Accordingly, a huge gap may exist between physical and online shopping experience.

(55) The plurality of objects, e.g. products, O on display in a store may be provided with respective tags T, i.e. one tag T for each product O, at 610. The tags T may be such that they enable detection of position and/or movement, e.g. tags with Bluetooth beacons and accelerometers. The tags T themselves and/or in association with one or more detection devices 30 may detect the position and/or movement of the objects and the server device 20 may receive the position and/or movement data at 630.

(56) The detection and reception may take place over time, at fixed intervals or based on given triggers. The server device 20 may store at 640 the position and/or movement data relating to the plurality of objects over the course of time. The stored data may be processed at 650 in order to obtain values of aggregate indicators relative to a predefined amount of time. For example, an aggregate indicator may indicate how many times an object has been lifted over 24 hours.

(57) A user U, such as a store owner, may be interested in learning which products are more interesting for customers and may request this information to the server device 20 via their mobile device 50. The mobile device 50 may display, e.g. thanks to an app, the values of the aggregate indicators superimposed on a spatial representation of the shop, e.g. a map of the shop (at 660).

(58) In particular, the mobile device 50 may display the values of the aggregate indicator using different shades of red, e.g. going from white to dark red, changing the ratio of the white/red components. The values of the aggregate indicator for the plurality of objects may be divided into intervals, and each interval may be associated to a specific shade of red. A pixel on the map of the shop corresponding to the standard location of an object may be displayed with a color corresponding to the value of the aggregate indicator for that object. Each pixel on the map may be accordingly color-coded, so that the layout of the shop is displayed in color and the shop owner U can see which areas of the shop and which objects are most sought after (e.g. smartphones 4 dark-red areas, objects lifted more than 100 times in 24 hours), as well as which ones attract less interest (e.g. DVDs .fwdarw.white areas, lifted less than 10 times in 24 hours).

(59) The displayed informative map may allow for zooming in, so that a more refined understanding of the customer experience may be achieved.

(60) Therefore, the position and/or movement data may enable real-time analytics for the physical retail space. This may lead to an improved customer support in-store due to better knowledge of the customers, to a more effective placement of the products in the store and to an optimized selection of which/how many items to provide in the store.

(61) Further, the examples of the system described above may be implemented in a store, shop or showroom with minimal invasive installation.

(62) FIG. 7 shows an exemplary functional block diagram of the software layers according to the present disclosure.

(63) A user such as a customer C may operate a mobile device 10 on which one or more application programs 80 run to perform the functionalities described heretofore, in particular with reference to FIGS. 4 and 5. A user such as a shop owner O may operate a mobile device 50 on which one or more application programs 80 run to perform the functionalities described heretofore, in particular with reference to FIG. 6.

(64) The application program(s) 80 may communicate with an application server 82, which may be part of the server device 20. An application server 82 may be deployed on premises or in a cloud environment. The tags T may also communicate with the application server 82.

(65) The application server 82 may provide an interface between the tags T and mobile devices 10, 50 on one side and a database management system 84 on the other side, wherein the database management system 84 may also be part of the server device 20. The database management system 84 may manage one or more databases, such as the object information database 40. The database management system 84 may be implemented as described in “The SAP HANA Database—An Architecture Overview”, Franz Färber, et al., 2012.

(66) The examples of the system described above with reference to FIGS. 1 to 7 may provide the above-stated advantages for any process that involves information of a plurality of objects and requires support for the customer as well as the retailer or any other party involved in the process.