METHOD OF MANUFACTURING AN AUTOMOTIVE VEHICLE USING AUTONOMOUS CONVEYANCE OF A VEHICLE CHASSIS
20200156722 ยท 2020-05-21
Assignee
Inventors
- Adam Keith WATKINS (Hixson, TN, US)
- Douglas Herbert Bartow (Signal Mountain, TN, US)
- Simon Heinrich STARK (Chattanooga, TN, US)
- Harold WALTERS (Hixson, TN, US)
- Antonio PINTO (Chattanooga, TN, US)
Cpc classification
B62D65/022
PERFORMING OPERATIONS; TRANSPORTING
G06Q10/087
PHYSICS
G06Q10/08
PHYSICS
B62D15/0265
PERFORMING OPERATIONS; TRANSPORTING
B62D65/18
PERFORMING OPERATIONS; TRANSPORTING
B62D65/005
PERFORMING OPERATIONS; TRANSPORTING
International classification
B62D65/18
PERFORMING OPERATIONS; TRANSPORTING
B62D15/02
PERFORMING OPERATIONS; TRANSPORTING
B62D65/02
PERFORMING OPERATIONS; TRANSPORTING
Abstract
A system and method for manufacturing a motor vehicle using autonomous conveyance of a vehicle chassis and a control system to execute functions including steering, acceleration/deceleration, environment monitoring and obstacle avoidance.
Claims
1. A vehicle manufacturing method comprising: assembling together a chassis, accelerator system, steering system, braking system, computer, and wheels to create a partially fabricating a vehicle; providing a control system in communication with the computer of the partially fabricated vehicle wherein the control system comprises: one or more sensors; and a gateway configured to interface inputs from the one or more sensors with the vehicle; sensing surroundings of the vehicle by the control system and accessing by the control system the steering system of the vehicle to steer the vehicle based on the sensed surroundings; and communicating by the control system information pertaining to further vehicle assembly steps for use in manufacturing substation.
2. The method of claim 1 further comprising: completing fabrication steps to create a completed vehicle; and self-guiding the completed vehicle through one or more delivery steps.
3. The method of claim 1 further comprising: identifying flawed vehicle builds; taking the flawed vehicle builds out of the assembly line; and permitting other vehicles in the assembly line to continue along the assembly line to one or more of the plurality of additional assembly line fabrication steps without assembly line stoppages.
4. The method of claim 1 wherein at least one of the one or more sensors is a visual management guide sensor configured to follow a pre-determined path along the assembly line.
5. The method of claim 1 wherein the control system is integrated with a vehicle tracking system.
6. The method of claim 1 wherein at least one of the one or more sensors is a pedestrian safety scanner.
7. The method of claim 1 further comprising self-driving the vehicle to a parts warehouse for part installation.
8. A vehicle manufacturing control system for use in a vehicle manufacturing process comprising: one or more sensors; a gateway configured as an interface between the one or more sensors and the vehicle to receive data from the one or more sensors and output commands based on the received data to the vehicle to operate the vehicle; an electronic control unit wherein; the electronic control unit is configured to autonomously execute one or more vehicle functions; the electronic control unit is configured to sense surroundings of the vehicle; the electronic control unit is integrated with a vehicle tracking system; and the electronic control unit is configured to communicate with manufacturing substations.
9. The control system of claim 8 wherein the one or more sensors comprises a visual management guide sensor.
10. The control system of claim 9 wherein the visual management guide sensor is configured to: provide visual or magnetic strip guidance to follow a pre-determined path; and output data to the gateway to ensure the vehicle is moving on a correct path within an allowed space.
11. The control system of claim 8 wherein the one or more sensors comprises a pedestrian safety scanner.
12. The control system of claim 8 wherein the one or more vehicle functions are selected from steering, acceleration/deceleration, environment monitoring, and dynamic driving task fallback strategies.
13. The control system of claim 8 wherein the vehicle tracking systems is configured to track location and identification of vehicles on the assembly line and parts needed by the vehicle for vehicle fabrication completion.
Description
DESCRIPTION OF THE DRAWINGS
[0006] The detailed description particularly refers to the accompanying figures, which depict illustrative embodiments, and in which:
[0007]
[0008]
[0009]
[0010]
[0011]
[0012]
[0013]
[0014]
[0015]
DETAILED DESCRIPTION OF ILLUSTRATIVE EMBODIMENTS
[0016] All embodiments and components described and depicted herein are illustrative examples only.
[0017]
[0018] The term vehicle as used herein means, if not specifically stated, either a fully or partially fabricated vehicle, depending on the context in which the term is used. The term assembly line or line includes any series of manufacturing steps, and does not necessarily mean an actual line or uninterrupted series of stations as in the traditional manufacturing sense.
[0019] In an illustrative embodiment, control system 100 includes gateway 102, radar sensor 104, visual management guide sensor 106, secondary pedestrian safety scanner 108 and user interface 110. Additional or other combinations of sensors, interfaces and gateways may be included in control system 100. Illustrative sensors include: radar sensors, mono and stereo cameras, near, medium and high range cameras, LIDAR systems and ultra sound sensors. Some of these systems may be implemented in a redundant or overlapping manner. For simplicity, the term camera may include any type of image capture device, which may capture still or video images that would be compatible with the vehicle and manufacturing process.
[0020] Any autonomous or driver assist technology that can facilitate a vehicle progressing through the assembly line may be incorporated into control system 100, including, for example, technology to guide, propel and stop the vehicle along the assembly line. More specifically, control system 100 may autonomously execute one or more of the following functions: steering, acceleration and deceleration; monitoring of the environment; and dynamic driving task fallback strategies. These capabilities may be incorporated into subsystems on control system 100, generally known as park assist, blind spot monitoring, lane-keeping assist or lane centering, collision avoidance, adaptive cruise control, cross-traffic monitoring, brake assist, distance control and emergency braking, for example, or other autonomous vehicle functionality that can facilitate a driverless vehicle progressing through an assembly line. Subsystems may provide assistance or avoidance functions and/or generate warning signals such as audio or visual alerts.
[0021] By way of example, a combination of cameras radar or other sensors is used to analyze a vehicle's speed and its distance to other vehicles or objects. The positions and movement of objects can also be obtained by this combination of cameras sensors, which may be provided as input to vehicle control algorithms. Other illustrative input parameters are LIDAR or ultrasound-generated signals that represent objects or humans in the vicinity of the vehicle. Vehicles may be equipped to recognize and act on signage information. Signage may be fixed or provided as a changeable display that can be altered according to changes in desired assembly line manufacturing steps.
[0022] Examples of technology/components/subsystems to be include:
[0023] Audi Traffic Jam Assist
[0024] Adaptive Cruise Control
[0025] PLA
[0026] Electronic Power Steering
[0027] Anti-Crash
[0028] Loading Cells
[0029] Remote guidance
[0030] Magnetic Guidance
[0031] External Control Module
[0032] Interface with the vehicle
[0033] Programmable route
[0034] Path Guidance Device (Vision)
[0035] Object Detection System
[0036] GPS Monitoring
[0037] Human Machine Interface
[0038] Visual Guidance System
[0039] Personnel Detection System
[0040] Emergency Stop
[0041] Control system 100 may use a centralized architecture having a central control unit that processes parameters input from a plurality of sensors. Alternatively, more than one control system may be used in a distributed arrangement, wherein each control system governs different functions of an autonomous system. Control system 100 may decode and analyze images, and other input parameters obtained from various sensors.
[0042] Gateway 102 serves as an interface between inputs from the sensors and the vehicle. Gateway 102 transforms data from radar sensor 104 and visual management guide sensor 106 into commands to operate the vehicle, such as for example, speed up, slow down, go left, go right, stop, reverse, forward, etc. The data is transformed based on parameters ascertained by sensors 104, 106 or other sensors employed in the system. The data may undergo an initial processing before serving as the basis of a command. For example, parameters may be analyzed to determine if they are within particular ranges or meet designated thresholds.
[0043] Visual management guide sensor 106 follows a pre-determined path (magnetic strip or visual guidance) and feeds data back to gateway 102 to ensure that the vehicle is moving on the correct path within the confines of the allowed space along an X-axis (left to right path). For example, the assembly line can be mapped out and the vehicle guided along the mapped path. Accordingly, if the vehicle encounters obstacles, it may arrest movement to the vehicle or guide the vehicle along its mapped path after the vehicle navigates around the obstacle.
[0044] Radar sensor 104 ensures correct speed and distance from a Y-axis perspective, wherein the Y-axis is defined as the line of forward or backward movement of the vehicle. For example, radar sensor 104 may sense the distance between a vehicle and a station. The distance information is fed back through gateway 102 to control system 100, which provides a signal to the vehicle to cause it to speed up, slow down, stop, or start as needed, and position itself appropriately with respect to the station. It is noted that the control system can also be programmed so the X-axis and Y-axis are referenced to the plant coordinates.
[0045] A pedestrian safety scanner 108, such as a laser scanner, detects interferences in the travel path of the vehicle, and signals the vehicle to either stop or go based on clearance. Clearances are fed back through gateway 102 to control system 100, which provides a signal to the vehicle to cause it to speed up, slow down, stop, or start as needed. An illustrative scanner 108 is a spinning LIDAR unit that uses laser beams that reflect back to the LIDAR unit to create a 360 degree image of the vehicle's surrounding.
[0046] A user interface 110 allows an operator to input commands to start and stop the assembly line system. User interface 110 may be either graphical, for example a display screen, or non-graphical, for example buttons or switches. Gateway 102, radar sensor 104, visual management guide sensor 106 and secondary pedestrian safety scanner 108 may be incorporated into a single control system 100 that can be functionally attached (mechanically and electrically) to a vehicle that has been assembled to the point of being able to be self-powered and self-guided. Once activated, control system 100 will allow the vehicle to progress along the assembly line without, or with little, human intervention or use of conveyors or lifts. User interface 110 allows or disallows gateway 102 to communicate with the vehicle. It also serves as a manual override for the system so the system can be shut own, such as in an emergency situation.
[0047]
[0048] Illustrative sensors 202 include LIDAR, camera or other optical image capture device, radar, geo-positioning system (GPS) and wheel speed sensor.
[0049] Electronic control unit 204 receives input from sensors 202, which is processed to produce signals representative of decision making results. Electronic control unit 204 may include machine readable code, which when executed, guides and controls a vehicle through an assembly line where it undergoes various manufacturing steps. The machine readable code includes algorithms, such as those to implement various autonomous vehicle functionality. Electronic control unit 204 may also include a user interface, which, by way of example, may be a display screen or non-graphical user interface.
[0050] Actuators 206 receive signals from electronic control unit 204, which direct the actuators to turn on or off, or implement actions as dictated by the received signals. Actuators 206 may, for example, control acceleration, braking and steering.
[0051]
[0052] A partially assembled vehicle with a vehicle control system 100 can progress through an assembly line autonomously. In conventional vehicle manufacturing assembly processes, however, the vehicle would not be equipped to operate autonomously until near the end of the assembly process. For example, a drive train is typically installed as one of the last steps, which marries the engine chassis and body. Following this step on traditional assembly lines, brake lines, fuel pipes, air-line components, and essential connections for electric vehicles. Fuel, brake fluid and coolant are then filled into the vehicle and essential connections for electric vehicles are made. Finally, wheels are mounted and the steering wheel is installed. Some or all of these steps would be carried out during a first phase of manufacturing according to the described embodiments. Therefore, the traditional assembly line process would not generate a vehicle that could be self-powered and self-guided until the manufacturing process was nearly complete. Accordingly, any of the later process steps
[0053] After the step of installing the wheels noted above, the invention includes having the partially made vehicle self power and drive itself through the rest of the manufacturing process, including alignment and calibration. The self-guided and self-driven vehicle will arrive at each successive next station required for further fabrication of the vehicle.
[0054] It is only after the vehicle can be autonomously driven that additional assembly steps can be executed. Illustrative assembly line steps through which the vehicle may progress autonomously include: cockpit assembly, sunroof installation, door installation, interior and exterior mirror and light installation and seat placement and connection. The vehicle may also self-power to stations for systems calibration and vehicle testing.
[0055] Through the disclosed embodiments, the following equipment, personnel and steps may be eliminated, resulting in potential cost savings and process improvement: devices for conveying vehicle bodies, lifts, persons responsible for loading and unloading supplier parts to go from supplier warehouse to OEM factories, truck drivers and space (both at supplier warehouse and OEM factories) for storing parts.
[0056] Illustrative production facilities may be configured such that a vehicle can be taken out of line and the remaining vehicles can continue cycling through production. This can be accomplished manually or control system 100 can be configured to recognize events that warrant removal of a vehicle from the assembly line and also to autonomously guide the vehicle away from the line. The control system may be configured so vehicles removed from the assembly line may be guided to a single location or to a location specific to the event that triggered the vehicle's removal. This will prevent production line down time.
[0057] The specialized control device and components are configured so they will not interfere with the remainder of the manufacturing process. In an illustrative embodiment, control system 100 is configured to be hung on the front of the vehicle. Using its sensors, it can communicate with the vehicle on which it is hung and on other vehicles either directly or through components already installed in the vehicle, allowing all vehicles to be automated within the assembly process. Generally, control system 100 will be an external control module that can be removed once the vehicle is fully or partially assembled.
[0058] Embodiments of the autonomous assembly system may incorporate use of a traditional manifest that is placed on a vehicle and provides information on specific parts, features or other options that should be included with the specific vehicle. Typically, option codes are associated with options to be included on a vehicle. The option codes are used to determine what parts are needed for the vehicle. These codes can be stored in a database that is accessed pursuant to algorithms integrated with the autonomous conveyance and manufacturing system to automatically provide required parts to stations. Manifests can be coded and scanned or read by workers at stations and elsewhere in the manufacturing environment. Vehicles can be tracked throughout the manufacturing and delivery process, so a vehicle's position is always known to the system and can be identified by logistics or manufacturing personnel. Conventional digitized logistics systems for vehicle manufacturing can be integrated with control system 100 to further automate the manufacturing and delivery processes.
[0059]
[0060] Vehicle 406 then progresses to a test and finish area 412. Vehicle 206 can be tested and evaluated for any manufacturing flaws or defects. Various finishing steps may be employed here. Vehicle 406 then advances to a yard 414, where it awaits return to the test and finish area 412 or advancement toward shipping and delivery. Depending on the nature of flaws or defects found, vehicle 406 may be returned to the test and finish area or to assembly line 404. Testing may include, for example, digital roll & alignment and digital test track.
[0061] Once vehicle 406 successfully completes testing and meets designated standards, it proceeds to preparation and vehicle distribution area 416. In vehicle preparation and distribution area 416, protective measures may be applied to vehicle 416 for transport purposes and other actions taken to facilitate safe and efficient transport and delivery of the vehicle.
[0062] Delivery buffering area 418 is the next stop for vehicle 406. Delivery buffering area 418 may serve to house vehicles awaiting shipment to dealers, for example. Buffering area 418 may also serve to maintain enough supply of vehicles to keep operations and delivery running on schedule. For example, delivery buffering area 418 may keep buffer inventory on hand to compensate for fluctuations in supply off of the manufacturing line.
[0063] Vehicle 406 leaves delivery buffer area 418 for either a train staging area 420 or a truck 422. Vehicle 405 may be loaded onto a train in area 424 where it will be delivered further down the distribution line. From wheel installation station 410 through delivery buffer area, and to trucks or train, vehicle 406 can be operated in autonomous mode all or some of the time.
[0064]
[0065]
[0066] After supplier content is installed, vehicle 606 self-drives to a test and finish area 618, where testing and finishing as described above are performed. Once vehicle 606 is designated as having passed all tests and any finishing steps have been completed, vehicle 606 may move autonomously to an external sales location 620. From here, vehicle 606 can drive to a train or truck at location 622 or to a customer at location 624. It will be understood that although vehicle 606 is enabled as a self-driving vehicle during the manufacturing and before the testing and distribution processes, vehicle 606 may be driven by a human during any segment of the process as chosen or advanced for a manufacturing segment by other traditional means, such as conveyors. This is also the case with the illustrative production layouts of
[0067]
[0068]
[0069]
[0070] Electronic tracking systems for vehicle advancing through manufacturing, testing and delivery processes may be employed to coordinate the autonomous vehicle features with the manufacturing process. Vehicle positions and parts information can be monitored and relayed and/or displayed to users in real time. Data exchange, for example using Bluetooth technology, can also be employed. Conventional body tracking can also be coordinated with the autonomous vehicle manufacturing controllers. Parts inventory, location and distribution to stations where needed can be coordinated with the control system.
[0071] Various embodiments have been described, each having a different combination of elements. The invention is not limited to the specific embodiments disclosed, and may include different combinations of the elements disclosed, omission of some elements or the replacement of elements by the equivalents of such structures and devices.
[0072] While illustrative embodiments have been described, additional advantages and modifications will occur to those skilled in the art. Therefore, the invention in its broader aspects is not limited to specific details shown and described herein. Modifications, for example, self-driving features, sequence of steps and incorporation of equivalent components, may be made without departing from the spirit and scope of the invention. Accordingly, it is intended that the invention not be limited to the specific illustrative embodiments, but be interpreted within the full spirit and scope of the appended claims and their equivalents.