MONITORING VEHICLE MOVEMENT FOR TRAFFIC RISK MITIGATION
A monitoring system detects and mitigates traffic risks among a group of vehicles. The group of vehicles includes a ground-based vehicle (GBV), e.g. an automotive vehicle, and an air-based vehicle (ABV), e.g. a drone, which is operated to track a ground-based object (GBO), e.g. an unprotected road user or an animal. The monitoring system performs a method comprising: obtaining (301) predicted navigation data for the ground-based vehicle and the air-based vehicle, processing (302) the predicted navigation data to obtain one or more future locations of the ground based-object and to detect an upcoming spatial proximity between the ground-based object and the ground-based vehicle, and causing (305), upon detection of the upcoming spatial proximity, an alert signal to be provided to at least one of the ground-based object and the ground-based vehicle.
This application claims the benefit of Swedish Patent Application No. 1950651-8, filed Jun. 3, 2019, the disclosure of which is incorporated herein by reference in its entirety.
TECHNICAL FIELDThe present disclosure relates generally to mitigating traffic risks and, more particularly, to techniques for monitoring movement of vehicles for mitigation of traffic risks in relation to a moveable ground-based object.
BACKGROUNDAutomotive vehicles are increasingly provided with sensors capable of obtaining information about the surroundings of the vehicle. The information may be provided to safety functions in the vehicle, which may warn the driver of a potentially dangerous situation or take automatic action to mitigate such a traffic risk, for example by activating the brakes, reducing the speed of the vehicle, or steering the vehicle. Autonomous vehicles (AVs) are in development and are foreseen to include advanced control systems capable of interpreting sensory information as well as inter-vehicle communication to identify appropriate navigation paths, obstacles and relevant signage, and thereby control the movement of the vehicle with little or no human input.
Irrespective of the level of sophistication of the safety functions and the control systems, there is still a risk that the vehicle sensors are unable to properly detect a dangerous situation, for example with respect to unprotected road users, livestock and wildlife, in particular when any such object is hidden from view of the driver and/or vehicle sensors.
SUMMARYIt is an objective to at least partly overcome one or more limitations of the prior art.
A further objective is to provide a technique for monitoring movement of vehicles for mitigation of traffic risks in relation to moveable ground-based objects such as unprotected road users, livestock and wildlife.
A yet further objective is to provide such a technique which is resource-efficient and automated.
One or more of these objectives, as well as further objectives that may appear from the description below, are at least partly achieved by a method of mitigating traffic risks, a computer-readable medium, and a monitoring system according to the independent claims, embodiments thereof being defined by the dependent claims.
Still other objectives, as well as features, aspects and technical effects will appear from the following detailed description, the attached claims and the drawings.
Embodiments will now be described in more detail with reference to the accompanying schematic drawings.
Embodiments will now be described more fully hereinafter with reference to the accompanying drawings, in which some, but not all, embodiments are shown. Indeed, the subject of the present disclosure may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure may satisfy applicable legal requirements. Like numbers refer to like elements throughout.
Also, it will be understood that, where possible, any of the advantages, features, functions, devices, and/or operational aspects of any of the embodiments described and/or contemplated herein may be included in any of the other embodiments described and/or contemplated herein, and/or vice versa. In addition, where possible, any terms expressed in the singular form herein are meant to also include the plural form and/or vice versa, unless explicitly stated otherwise. As used herein, “at least one” shall mean “one or more” and these phrases are intended to be interchangeable. Accordingly, the terms “a” and/or “an” shall mean “at least one” or “one or more”, even though the phrase “one or more” or “at least one” is also used herein. As used herein, except where the context requires otherwise owing to express language or necessary implication, the word “comprise” or variations such as “comprises” or “comprising” is used in an inclusive sense, that is, to specify the presence of the stated features but not to preclude the presence or addition of further features in various embodiments. The term “compute”, and derivatives thereof, is used in its conventional meaning and may be seen to involve performing a calculation involving one or more mathematical operations to produce or determine a result, for example by use of a computer.
As used herein, a “group of vehicles” is intended to imply a provision of two or more vehicles. Likewise, the terms “multiple”, “plural” and “plurality” are intended to imply provision of two or more vehicles.
As used herein, the attribute “predicted” implies an event or value that is planned, scheduled or estimated at one or more future time points. Thus, a predicted trajectory of an object refers to a future movement of the object. Depending on the type of object, the predicted trajectory may be known from a predefined schedule or be estimated based on previous movement and/or other input data, for example a vision-based guidance system.
As used herein, “air-based vehicle” is intended to imply any vehicle that is propelled for movement or levitation above the surface of the Earth and is operable to track a ground-based object. Thus, air-based vehicles include aircrafts such as unmanned aerial vehicles (UAVs), also known as drones, and helicopters, as well as artificial satellites including both geostationary and non-geostationary satellites. The air-based vehicles may be controlled by an onboard automated control system, an onboard human pilot, a ground-based control system or a ground-based human pilot.
As used herein, “ground-based vehicle” is intended to imply any self-propelled vehicle for movement in contact with the surface of the Earth, including but not limited to automotive cars, lorries, motorbikes, buses, etc. The ground-based vehicle may be controlled by an automated control system and/or by a human driver.
As used herein, “ground-based object” is intended to imply any moving or moveable object that is in contact with the surface of the Earth and may be tracked by an air-based vehicle. The ground-based object may be any animate or inanimate object, including unprotected human individuals, as well as livestock, pets, wild animals, etc. The unprotected human individuals may, for example, be on foot, running or operating a human-powered vehicle such as a bicycle. For the avoidance of doubt, a ground-based object could be one or more human individuals, such as a group of runners or a group of cyclists.
A used herein, “vision sensor” is intended to imply any sensor or combination of sensors that provides a two- or three-dimensional representation within a field of view. The vision sensor may be or include any of one or more light sensors or cameras (for example in the visible spectrum, the infrared spectrum, etc.), one or more Light Detection and Ranging (LIDAR) systems, one or more Radio Detection and Ranging (RADAR) systems, one or more ultrasonic sensors, etc.
It will furthermore be understood that, although the terms first, second, etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element, and, similarly, a second element could be termed a first element, without departing the scope of the present disclosure. As used herein, the term “and/or” between elements includes any and all combinations of one or more of the elements.
Embodiments are related to detection and mitigation of traffic risks, specifically traffic risks that involve a potential collision between a ground-based vehicle (denoted GBV in the following) and a ground-based object (denoted GBO in the following). The traffic risks may be detected by use of air-based vehicles (denoted ABV in the following), which are configured and/or operated to follow or “track” a respective GBO.
The following description will refer to ground-based objects (GBOs), ground-based vehicles (GBVs), and air-based vehicles (ABVs). The individual 10 is an example of a GBO, the car 20 is an example of a GBV, and the drone 30 is an example of an ABV.
The method 300 is performed repeatedly at consecutive time points while the ABV 30 is flying. Each such consecutive time point may be regarded as a current time point. As will be exemplified in further detail below with reference to
In one embodiment, the method 300 involves an initial grouping step (not shown) of defining the group of vehicles, for example by use of a clustering algorithm. For example, the initial step may determine the group of vehicles among a larger plurality of vehicles by space-time cluster analysis of respective locations of the larger plurality of vehicles at the current time point and/or at one or more future time points. The group of vehicles may be given by a cluster that includes a selected ABV 30. The initial grouping step may perform the clustering based on the respective locations and associated time points, or perform an initial partitioning of locations by time points followed by a location-based clustering.
Step 301 obtains predicted navigation data (PND) for GBV 20 and ABV 30. The PND estimates the future movement of GBV 20 and ABV 30 in relation to the current time point. Such future movement also includes the option that the GBV/ABV will not move. The PND may be determined in any suitable way, for example based on the historic movement of GBV 20 and ABV 30 up to the current time point, the local topology (location of the road, location of slopes and bends, road condition, etc.), the current traffic situation (road works, amount of traffic, traffic jams, speed limits, traffic signs, traffic lights, etc.), a predefined movement plan such as a predefined planned route, or any combination thereof. The historic movement may be determined based position data from the position sensor in the respective local control device 26, 36 (
Step 302 processes the PND for detection of an upcoming spatial proximity between GBV 20 and GBO 10, according to any suitable criterion (see below). It is realized that step 302 also involves estimating one or more future locations of GBO 10 based on the PND of ABV 30, in order to allow for the detection of the upcoming spatial proximity between GBV 20 and GBO 10. In a simple and non-limiting example, ABV 30 is controlled to fly directly above GBO 10, causing a predicted trajectory of GBO 10 to be equal to the predicted trajectory of ABV 30 as projected onto the ground. More advanced and generic algorithms for movement estimation of GBO 10 will be described below with reference to
If no spatial proximity is detected by step 302, step 303 directs the method 300 to execute step 301 at a subsequent time point. If step 302 detects an upcoming spatial proximity, step 303 may direct the method to step 305, which causes an alert signal to be provided to at least one of GBO 10 and GBV 20. Subsequent to step 305, the method 300 may proceed to execute step 301 at a subsequent time point.
Reverting to the example in
The proximity detection of
In one embodiment, at least one of the spatial and temporal limits are set as a function of a velocity vector of the GBV 20 and/or a velocity vector of the GBO 10. The velocity vector may be an actual velocity vector at the current time or a predicted velocity vector at a future time point. For example, if the method determines that the first and second predicted trajectories intersect, the spatial limit and/or the temporal limit may be set to scale with the velocity of the GBV 20 and/or the GBO 10 towards the point of intersection. This will effectively increase the size of the region that is deemed to involve a traffic risk if occupied by the GBO 10 and the GBV 20 at the same time (cf. 100 in
Reverting to
As indicated by dashed lines in
One advantage of the occlusion analysis is that the GBV/GBO may be specifically alerted of a traffic risk that is likely to be unexpected. The occlusion analysis may also be implemented to restrict the number of alert signals, thereby reducing the cognitive load on the individual 10 and/or the driver of the GBV 20, or the computational load on the vehicle control system in the GBV 20. Reverting to
If implemented, step 304 may be performed when the method 300 has detected an upcoming spatial proximity. This improves the processing-efficiency of the method 300, by limiting the occlusion analysis of step 304 to situations that are deemed to pose a traffic risk based on the proximity detection of step 302.
In the following, embodiments of the occlusion analysis will be described with reference to
The first occlusion analysis, OA1, is exemplified in
Based on the detection of intervening objects in step 323, step 324 evaluates the LOS between the reference point and the predicted location is blocked by the intervening object(s). In this evaluation, step 324 may also account for the extent of the target (GBO/GBV) around the predicted location. If the LOS is found to be blocked, step 324 indicates an occluded LOS (step 325), otherwise a non-occluded LOS is indicated (step 326).
In a variant, OA1 also involves predicting the LOS of the origin at one or more subsequent time points based on the sensor data, e.g. an image, by determining a predicted location of the origin in the sensor data and evaluating if the LOS between the predicted location of the origin and the corresponding predicted location of the target is blocked by any object included in the sensor data.
OA1 is further exemplified in
PND for one or more further GBVs in the vicinity of the GBV 20 and the GBO 10. Step 332 determines the LOS between the GBV 20 and the GBO 10 at the current time point and/or one or more future time points. Step 333 maps the location of each further GBV, given by the PND obtained in step 331, to the LOS at the respective time point. Step 334 evaluates the mapping of step 333 and indicates an occluded LOS when at least one further GBV blocks the LOS at one or more time points (step 335). Otherwise, step 334 may indicate a non-occluded LOS (step 336). Step 334 may account for the extent of the further GBV in the evaluation.
Step 332 may use the PND obtained in step 301 (
OA2 is further exemplified in
In
In the situations of
In one embodiment (not shown), the method 300 may comprise distributing sensor data between any combination of the GBV, the ABV and the GBO. Such distributing may be performed upon spatial proximity detection (steps 302-303) and/or occluded LOS detection (step 304). For example, the ABV 30 may transmit sensor data captured by its vision sensor 32 to the GBV 10, for example for display to the driver by the feedback device 27 or for analysis by the vehicle control system. Correspondingly, the GBV 20 may transmit sensor data captured by its vision sensor 22 to the GBO 10 via the ABV 30, for example for display on the electronic device 11 and/or the eyewear 12. This embodiment will give the recipient of the sensor data a better understanding of the impending event.
In one embodiment (not shown), the method 300 may comprise the ABV 30 measuring and reporting characterizing data of the GBO 10 that it is tracking. The characterizing data may include one or more size parameters, such as width, length, area, shape, number of individuals, etc. For example, a drone tracking a group of runners or cyclists may report a size parameter to automotive vehicles within a relevant range, for example the group of vehicles given by the above-mentioned clustering. In one embodiment, the ABV 30 autonomously reports the characterizing data to all GBVs 20 in the group of vehicles. In another embodiment, the ABV 30 is caused to selectively, upon the spatial proximity detection and/or the occluded LOS detection, report the characterizing data to all GBVs in the group of vehicles or to the GBV 20 in spatial proximity. In either embodiment, the reporting may be performed repeatedly to provide updated characterizing data. One or more GBVs 20 will thus be fed relevant real-time information about the GBO 10, allowing the respective driver and/or vehicle control system of the respective GBV 20 to take appropriate action for risk mitigation.
As noted above, the method 300 of
It may be noted that some functions of the described methods may be implemented in hardware, which may be invoked by the executing instructions 4 to produce a specific type of output from a specific type of input. The instructions 4 may be supplied to the LCD on a computer-readable medium 80, which may be a tangible (non-transitory) product (for example magnetic medium, optical disk, read-only memory, flash memory, etc.) or a propagating signal.
In the example of
While the subject of the present disclosure has been described in connection with what is presently considered to be the most practical embodiments, it is to be understood that the subject of the present disclosure is not to be limited to the disclosed embodiments, but on the contrary, is intended to cover various modifications and equivalent arrangements included within the spirit and the scope of the appended claims.
Further, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, parallel processing may be advantageous.
In the following, items are recited to summarize some aspects and embodiments as disclosed in the foregoing.
Item 1: A method of mitigating traffic risks among a group of vehicles comprising an air-based vehicle (30) and a ground-based vehicle (20), the air-based vehicle (30) being operated to track a ground-based object (10), said method comprising, at a current time point:
obtaining (301) predicted navigation data for the ground-based vehicle (20) and the air-based vehicle (30),
processing (302) the predicted navigation data to obtain one or more future locations of the ground based-object (10) and to detect an upcoming spatial proximity between the ground-based object (10) and the ground-based vehicle (20), and
causing (305), upon detection of the upcoming spatial proximity, an alert signal (AS) to be provided to at least one of the ground-based object (10) and the ground-based vehicle (20).
Item 2: The method of item 1, wherein the predicted navigation data is indicative of a first predicted trajectory (T20) for the ground-based vehicle (20), and a second predicted trajectory (T10) for the ground-based object (10) that is tracked by the air-based vehicle (30).
Item 3: The method of item 2, wherein the predicted navigation data comprises a predicted trajectory for the air-based vehicle (30), said processing (302) comprising: determining (311) location data of the ground-based object (10) in relation to the air-based vehicle (30), and converting (312) the predicted trajectory for the air-based vehicle (30) into the second predicted trajectory (T10) for the ground-based object (10) based on the location data.
Item 4: The method of item 2 or 3, wherein the first predicted trajectory (T20) comprises a sequence of first locations and associated first time points in relation to a reference time point, and wherein the second predicted trajectory (T10) comprises a sequence of second locations and associated second time points in relation to the reference time point.
Item 5. The method of item 4, wherein said processing (302) comprises: mapping (313) the first locations and the associated first time points of the first predicted trajectory (T20) to the second locations and the associated second time points of the second predicted trajectory (T10), and detecting (314) the upcoming spatial proximity based on the mapping.
Item 6. The method of item 5, wherein the upcoming spatial proximity is detected when a distance (D3, D4, D5) between a first location and a second location is below a spatial limit and when a time difference between a first time point associated with the first location and a second time point associated with the second location is below a temporal limit.
Item 7: The method of item 5 or 6, wherein at least one of the spatial and temporal limits are set as a function of a velocity vector of the ground-based vehicle (20) and/or a velocity vector of the ground-based object (10).
Item 8: The method of any preceding item, further comprising: analysing (304), upon the detection of the upcoming spatial proximity, the predicted navigation data and/or sensor data for determination of an occluded line-of-sight (LOS) between the ground-based object (10) and the ground-based vehicle (20), wherein said sensor data is obtained by a vision sensor (22; 12; 32) on at least one of the ground-based vehicle (20), the ground-based object (10) and the air-based vehicle (30).
Item 9: The method of item 8, wherein the group of vehicles comprises at least one further ground-based vehicle (20′), wherein said analysing (304) comprises: determining (332) an LOS between the ground-based vehicle (20) and the ground-based object (10) at one or more time points, mapping (333) a location of the at least one further ground-based vehicle (20′) to the LOS (L1, L2, L3) at said one or more time points, and detecting (334) the occluded LOS when the at least one further ground-based vehicle (20′) blocks the LOS (L1, L2, L3) at at least one of said one or more time points.
Item 10: The method of item 9, wherein said one or more time points are future in relation to the current time point.
Item 11: The method of item 9 or 10, wherein said determining (332) the LOS comprises: determining an LOS vector (LOSV1) between the air-based vehicle (30) and the ground based-vehicle (20), obtaining a location vector (LV) between the air-based vehicle (30) and the ground based-object (10), and computing the LOS between the ground-based object (10) and the ground-based vehicle (20) as a function of the LOS vector (LOSV1) and location vector (LV).
Item 12: The method of any one of items 8-11, wherein said analysing (304) comprises: determining (321) a predicted location of the ground-based object (10) or the ground-based vehicle (20), and operating (322) the vision sensor (22; 12; 32) of the ground-based vehicle (20), the ground-based object (10) or the air-based vehicle (30) to capture the sensor data with the predicted location positioned within a field of view (24; 34) of the vision sensor (22; 12; 32).
Item 13: The method of item 12, wherein said analysing (304) further comprises: processing (323) the sensor data for detection of an occluding object (40) between a reference point and the predicted location, the reference point being located on the ground-based vehicle (20) or on the ground-based object (10), respectively.
Item 14: The method of item 12 or 13, wherein said determining (321) the predicted location and said operating (322) the vision sensor (22; 12; 32) are performed at or subsequent to the current time point.
Item 15: The method of any preceding item, wherein the alert signal (AS) is provided to the ground-based object (10) via the air-based vehicle (30).
Item 16: The method of any preceding item, further comprising: determining the group of vehicles among a larger plurality of vehicles by clustering based on respective locations of the larger plurality of vehicles at one or more time points.
Item 17. A computer-readable medium comprising computer instructions (4) which, when executed by a processor (2), cause the processor (2) to perform the method of any one of items 1-16.
Item 18: A monitoring system for mitigating traffic risks among a group of vehicles comprising an air-based vehicle (30) and a ground-based vehicle (20), the air-based vehicle (30) being operable to track a ground-based object (10), said monitoring system being configured to, at a current time point:
obtain predicted navigation data for the ground-based vehicle (20) and the air-based vehicle (30),
process the predicted navigation data to determine one or more future locations of the ground based-object (10) and to detect an upcoming spatial proximity between the ground-based object (10) and the ground-based vehicle (20), and
cause, upon detection of the upcoming spatial proximity, an alert signal (AS) to be provided to at least one of the ground-based object (10) and the ground-based vehicle (20).
Item 19: The monitoring system of item 18, which is located on one vehicle in the group of vehicles and configured to communicate wirelessly with a respective local control device (26, 36) on other vehicles in the group of vehicles.
Item 20: The monitoring system of item 18, which is separate from the group of vehicles and configured to communicate wirelessly with local control devices (26, 36) on vehicles in the group of vehicles.
Item 21: A vehicle comprising a monitoring system in accordance with any one of items 18-20.
Claims
1. A method of mitigating traffic risks among a group of vehicles comprising an air-based vehicle and a ground-based vehicle, the air-based vehicle being operated to track a ground-based object, said method comprising, at a current time point:
- obtaining predicted navigation data for the ground-based vehicle and the air-based vehicle;
- processing the predicted navigation data to obtain one or more future locations of the ground based-object and to detect an upcoming spatial proximity between the ground-based object and the ground-based vehicle; and
- causing, upon detection of the upcoming spatial proximity, an alert signal to be provided to at least one of the ground-based object and the ground-based vehicle.
2. The method of claim 1, wherein the predicted navigation data is indicative of a first predicted trajectory for the ground-based vehicle, and a second predicted trajectory for the ground-based object that is tracked by the air-based vehicle.
3. The method of claim 2, wherein the predicted navigation data comprises a predicted trajectory for the air-based vehicle, said processing comprising: determining location data of the ground-based object in relation to the air-based vehicle, and converting the predicted trajectory for the air-based vehicle into the second predicted trajectory for the ground-based object based on the location data.
4. The method of claim 2, wherein the first predicted trajectory comprises a sequence of first locations and associated first time points in relation to a reference time point, and the second predicted trajectory comprises a sequence of second locations and associated second time points in relation to the reference time point.
5. The method of claim 4, wherein said processing comprises: mapping the first locations and the associated first time points of the first predicted trajectory to the second locations and the associated second time points of the second predicted trajectory, and detecting the upcoming spatial proximity based on the mapping.
6. The method of claim 5, wherein the upcoming spatial proximity is detected when a distance between a first location and a second location is below a spatial limit and when a time difference between a first time point associated with the first location and a second time point associated with the second location is below a temporal limit.
7. The method of claim 5, wherein at least one of the spatial and temporal limits are set as a function of a velocity vector of the ground-based vehicle and/or a velocity vector of the ground-based object.
8. The method of claim 1, further comprising: analysing, upon the detection of the upcoming spatial proximity, the predicted navigation data and/or sensor data for determination of an occluded line-of-sight between the ground-based object and the ground-based vehicle, wherein said sensor data is obtained by a vision sensor on at least one of the ground-based vehicle, the ground-based object and the air-based vehicle.
9. The method of claim 8, wherein the group of vehicles comprises at least one further ground-based vehicle, wherein said analysing comprises: determining an LOS between the ground-based vehicle and the ground-based object at one or more time points, mapping a location of the at least one further ground-based vehicle to the LOS at said one or more time points, and detecting the occluded LOS when the at least one further ground-based vehicle blocks the LOS at at least one of said one or more time points.
10. The method of claim 9, wherein said one or more time points are future in relation to the current time point.
11. The method of claim 9, wherein said determining the LOS comprises:
- determining an LOS vector between the air-based vehicle and the ground based-vehicle, obtaining a location vector between the air-based vehicle and the ground based-object, and computing the LOS between the ground-based object and the ground-based vehicle as a function of the LOS vector and location vector.
12. The method of claim 8, wherein said analysing comprises: determining a predicted location of the ground-based object or the ground-based vehicle, and operating the vision sensor of the ground-based vehicle, the ground-based object or the air-based vehicle to capture the sensor data with the predicted location positioned within a field of view of the vision sensor.
13. The method of claim 12, wherein said analysing further comprises: processing the sensor data for detection of an occluding object between a reference point and the predicted location, the reference point being located on the ground-based vehicle or on the ground-based object, respectively.
14. The method of claim 12, wherein said determining the predicted location and said operating the vision sensor are performed at or subsequent to the current time point.
15. The method of claim 1, wherein the alert signal is provided to the ground-based object via the air-based vehicle.
16. The method of claim 1, further comprising: determining the group of vehicles among a larger plurality of vehicles by clustering based on respective locations of the larger plurality of vehicles at one or more time points.
17. A non-transitory computer-readable medium comprising computer instructions which, when executed by a processor, cause the processor to perform the method of claim 1.
18. A monitoring system for mitigating traffic risks among a group of vehicles comprising an air-based vehicle and a ground-based vehicle, the air-based vehicle being operable to track a ground-based object, said monitoring system being configured to, at a current time point:
- obtain predicted navigation data for the ground-based vehicle and the air-based vehicle;
- process the predicted navigation data to determine one or more future locations of the ground based-object and to detect an upcoming spatial proximity between the ground-based object and the ground-based vehicle; and
- cause, upon detection of the upcoming spatial proximity, an alert signal to be provided to at least one of the ground-based object and the ground-based vehicle.
19. The monitoring system of claim 18, which is located on one vehicle in the group of vehicles and configured to communicate wirelessly with a respective local control device on other vehicles in the group of vehicles.
20. The monitoring system of claim 18, which is separate from the group of vehicles and configured to communicate wirelessly with local control devices on vehicles in the group of vehicles.
Type: Application
Filed: Apr 13, 2020
Publication Date: Dec 3, 2020
Patent Grant number: 11820400
Inventors: Ola THÖRN (Limhamn), Peter EXNER (Malmö), Shaun LEE (Winchester)
Application Number: 16/846,738