GB2566462A - Mobile terminal - Google Patents
Mobile terminal Download PDFInfo
- Publication number
- GB2566462A GB2566462A GB1714716.6A GB201714716A GB2566462A GB 2566462 A GB2566462 A GB 2566462A GB 201714716 A GB201714716 A GB 201714716A GB 2566462 A GB2566462 A GB 2566462A
- Authority
- GB
- United Kingdom
- Prior art keywords
- image
- determining
- mobile terminal
- distance
- distal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/593—Depth or shape recovery from multiple images from stereo images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
A method and apparatus for a mobile device 10 to determine distance, comprising first 12a-1 and second 12a-2 distal lenses a predetermined distance apart on the same surface of the device. The distal lenses are each connected to a camera by a respective connecting means, e.g. an optical fibre. The camera is used to capture a first image from the first lens 12a-1 and a second image from the second lens 12a-2. The first and second images are compared and the distance of an object in the images determined using the predetermined lens separation distance. Also claimed is a mobile device with two cameras spaced a predetermined distance apart to determine object distance.
Description
MOBILE TERMINAL
The present invention relates to a mobile terminal, such as a smartphone. In particular, the present invention relates to a method of determining a distance using a mobile terminal.
BACKGROUND TO THE PRESENT INVENTION
In the present disclosure, mobile terminals may include mobile telephones, smartphones, tablets, laptop computers, smartwatches, etc.
Mobile terminals such as smartphones are capable of determining distance using GPS technology. This method is limited by the accuracy of commercial GPS and the availability of GPS mapping of the terrain. Accuracies are higher for military applications and can be enhanced by multi-frequency systems and augmentation technologies, but these are generally not available for the public.
Distances can be accurately determined using laser in a laser distance meter. However, such systems are expensive and powerful lasers require special precautions, and therefore not suitable for general use.
Stadiametric rangefinding is a technique of measuring distances with a telescopic instrument. The method relies on either knowing the range to an object to determine the dimension of the object, or knowing the dimension of the object to determine the range. If neither the range to the object or the dimension of the object is known, this method cannot be used.
Other active rangefinding methodologies exist including sonar, ultrasound, radar, etc., with varying degrees of accuracy, which are unsuitable for general use.
In view of the foregoing, it is desirable to provide an improved method of determining a distance using a mobile terminal and an improved mobile terminal.
SUMMARY OF THE PRESENT INVENTION
Embodiments of the present invention provide a method of determining a distance using a mobile terminal. The mobile terminal comprises first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal.
The first and second distal lenses are each connected to a camera by a respective connecting means, e.g. an optical fibre. The method comprises capturing, by the camera,
-2a first image of an object from the first distal lens; capturing, by the camera, a second image of the object from the second distal lens; comparing the first image with the second image; and determining a distance in relation to the object based on a result of comparing the first image with the second image using the predetermined distance.
In another aspect, embodiments of the present invention provide a mobile terminal comprising a camera configured to capture an image; first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal; connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light from each distal lens to the camera; and processing means configured to compare an image of an object captured via the first distal lens with an image of the object captured via the second distal lens, and determine a distance in relation to the object based on the comparison using the predetermined distance between the two distal lenses.
According to preferred embodiments, two distal lenses separated by a predetermined, known, distance are provided on the same surface of the mobile terminal. Each distal lens is connected to a single camera such that an image can be taken by the mobile terminal from two different viewpoints using only one camera. Each image captured from each distal lens correspond to a different line of sight from the mobile terminal to an object that is the focus of the image. The object therefore would appear to have shifted with respect to the second image when compared to the first image. Comparing the first image with the second image provides information of the difference in the two lines of sight, and with the known predetermined distance between the two distal lenses, thus the distance separating the two lines of sight, it is possible to determine a distance in relation to the object, for example a range between the mobile terminal (thus the user) to the object, and/or a dimension of the object. Preferred embodiments therefore enables a user to determine a distance in relation to an object that does not rely on the availability of a third-party service such as GPS, in a way that is inexpensive and easily portable.
In preferred embodiments of the method, comparing the first image with the second image comprises determining an angle of convergence between the first image and the second image. The angle of convergence is an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens. The angle of convergence may be determined using any suitable method.
In preferred embodiments, the method further comprises selecting a reference point on the object in the first image; and identifying the reference point on the object in the second image. The step of comparing the first image with the second image may then comprise detecting a difference between a first position of the reference point within the first image and a second position of the reference point within the second image. In an embodiment
- 3of the method, the step of comparing the first image with the second image further comprises determining an angle of convergence using the detected difference between the first position and the second position. According to the embodiments, a reference point on the object in the first image may be automatically selected by a processing means of the mobile terminal, or it may be selected by the user. Similarly, the same or corresponding reference point on the object in the second image may be identified by the processing means or manually by the user.
In preferred embodiments of the method, determining a distance in relation to the object comprises determining a range from the mobile terminal to the object. The range may be determined using any suitable method as desired. In an embodiment, the range is determined by trigonometry using the known predetermined distance between the first and second distal lenses and the determined angle of convergence.
In preferred embodiments of the method, a final range from the mobile terminal to the object is determined by averaging two or more results. In particular, the final range is determined by determining a second (and subsequent) range from the mobile terminal to the object using the predetermined distance and a second (and subsequent) angle of convergence determined by comparing a third (and subsequent) image of an object captured by the camera from the first distal lens with a fourth (and subsequent) image of the object captured by the camera from the second distal lens, and averaging the first range and the second (and subsequent) range. By averaging two or more results, it is possible to improve the accuracy of the final range determined by the method.
In preferred embodiments of the method, the step of determining a distance in relation to the object further comprises determining a dimension of the object based on the determined range from the mobile terminal to the object.
In an embodiment, determining the dimension of the object comprises selecting, in one of the first image or the second image, a first point on the object and a second point on the object spaced from the first point; determining a difference between the first point and the second point within the first or second image; and determining the dimension of the object based on the determined difference between the first point and the second point. The selection of a first point and a second point on the object in either the first or the second image may be performed by the processing means of the mobile terminal, or it may be performed manually by the user, as desired. The first point and the second point may be any points on the object spaced apart from each other in any direction, for example, the first point may be a top corner of an object and the second point may be a bottom corner of the object. The determining of a difference between the first point and the second point within the image may be a difference between their positions within the image, for example, measured by the number of pixels between the two points and the pixel pitch.
-4In an embodiment, a second dimension of the object is determined based on the determined range from the mobile terminal to the object. The second dimension may be orthogonal to the previously determined dimension or at other angles to the previously determined dimension, as desired.
In an embodiment, the second dimension of the object is determined by selecting, in one of the first image or the second image, a third point on the object and a fourth point on the object spaced from the third point; determining a difference between the third point and the fourth point within the first or second image; and determining the second dimension of the object based on the determined difference between the third point and the fourth point.
In an embodiment, the method further comprises generating a scale grid based on the determined dimension of the object. The scale grid may be generated using one or more determined dimension(s) of the object. The scale grid may be used in many different ways, for example, for quickly reading off the size of different parts of the object, for comparing with the size of another object, etc.
In an embodiment, the generated scale grid may be embedded in the first image or the second image.
In an embodiment, the first image or the second image may be stored with the embedded scale grid in a predetermined format. The stored image may, for example, be shared with another party with the scale grid embedded, and/or be rescaled for other purposes whereby the grid may be scaled in direct proportion to the image of the object.
In an embodiment, the generated scale grid may be superimposed on the first image or the second image and displayed on a screen of the mobile terminal.
Further embodiments provides a mobile terminal comprising first and second cameras. Each camera is configured to capture an image, the first and second cameras being disposed spaced apart by a predetermined distance on the same surface of the mobile terminal. The mobile terminal further comprises processing means configured to compare an image of an object captured by the first camera with an image of the object captured by the second camera, and determine a distance in relation to the object based on the comparison using the predetermined distance.
It will be clear to a skilled person that other optional or alternative features are possible. For example, more than two distal lenses and/or cameras may be provided to a mobile terminal. Embodiments have been contemplated in which three, four or more distal lenses are provided on the same surface of a mobile terminal, each connected to a single camera within the body of the mobile terminal by respective optical fibre. Embodiments have also been contemplated in which multiple distal lenses are provided on multiple surfaces of a
- 5mobile terminal. Further embodiments have been contemplated in which a combination of multiple distal lenses and multiple cameras are provided to a mobile terminal.
BRIEF DESCRIPTION OF THE DRAWINGS
Various embodiments of the present invention will now be described, by way of example only, and with reference to the accompanying drawings in which:
Fig. 1 shows two alternative examples of a mobile terminal according to embodiments of the present disclosure;
Fig. 2 shows a flow diagram of an example method of determining a distance according to an embodiment; and
Fig. 3 shows a flow diagram of an example method of determining a dimension according to an embodiment.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENT
Embodiments of the present invention provide a method of determining a distance using a mobile terminal. The mobile terminal comprises first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal. The first and second distal lenses are each connected to a camera by a respective connecting means, e.g. an optical fibre. The method comprises capturing, by the camera, a first image of an object from the first distal lens; capturing, by the camera, a second image of the object from the second distal lens; comparing the first image with the second image; and determining a distance in relation to the object based on a result of comparing the first image with the second image using the predetermined distance between the first and second distal lenses.
In another aspect, embodiments of the present invention provide a mobile terminal comprising a camera configured to capture an image; first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal; connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light from each distal lens to the camera; and processing means configured to compare an image of an object captured via the first distal lens with an image of the object captured via the second distal lens, and determine a distance in relation to the object based on the comparison using the predetermined distance.
According to preferred embodiments, two distal lenses separated by a predetermined, known, distance are provided on the same surface of the mobile terminal. Each distal lens is connected to a single camera such that an image can be taken by the mobile terminal from two different viewpoints using only one camera. Each image captured from each distal lens correspond to a different line of sight from the mobile terminal to an object that is
-6the focus of the image. The object therefore would appear to have shifted with respect to the second image when compared to the first image. Comparing the first image with the second image provides information of the difference in the two lines of sight, and with the known predetermined distance between the two distal lenses, thus the distance separating the two lines of sight, it is possible to determine a distance in relation to the object, for example a range between the mobile terminal (thus the user) to the object, and/or a dimension of the object. Preferred embodiments therefore enable a user to determine a distance in relation to an object that does not rely on the availability of a third-party service such as GPS, in a way that is inexpensive and easily portable.
Fig. 1 shows an example of a mobile terminal according to an embodiment. In the embodiment, the mobile terminal is a smartphone 10A. However, the embodiment is also applicable to other mobile terminals such as a tablet computer, a laptop computer, a smart watch, or any other suitable mobile devices.
The terminal 10A comprises a screen 11 on the front surface. The screen 11 may be any conventional and commercially available screen suitable for use with a mobile terminal. A first distal lens 12a-1 is disposed on the front surface at the top left corner of the terminal 10A, and a second distal lens 12a-2 is disposed on the front surface at the top right corner of the terminal 10A. The first and second distal lenses 12a-1 and 12a-2 is mounted to the front surface of the terminal 10A. Preferably, the first and second distal lenses 12a-1 and 12a-2 are fixedly mounted to the front surface. However, in some embodiments, it may be desirable to mount the first and/or second distal lenses 12a-1 and 12a-2 through a respective carriage (not shown) that allows controlled movement of the first and second distal lenses 12a-1 and 12a-2 and allows the distal lenses 12a-1 and 12a-2 to be locked in place such that the lines of sight of both lenses 12a-1 and 12a-2 are parallel (both lenses point in the same direction. Where the first and second distal lenses 12a-1 and 12a-2 are movably mounted, a control 13 may be provided on one side of the terminal 10A for controlling the movement of the distal lenses 12a-1 and 12a-2. Flash 15a is provided on the front surface of the terminal 10A between the first and second distal lenses 12a-1 and 12a-2 to provide additional light for image capturing in low light conditions.
Another embodiment is also shown in Fig. 1, where a terminal 10B comprises a first distal lens 12b-1 and second distal lens 12b-2 disposed on the rear surface at diagonal corners of the terminal 10B. The first and second distal lenses 12b-1 and 12b-2 are similarly fixedly mounted to the rear surface of the terminal 10B. Similarly, flash 15b is provided on the rear surface of the terminal 10B between the first and second distal lenses 12b-1 and 12b-2.
- 7Other embodiments have been contemplated in which more than two distal lenses are provided on a surface of a mobile terminal, or multiple surfaces, e.g. both front and rear, of the mobile terminal. Further embodiments have been contemplated in which multiple cameras are provided on a surface (or multiple surfaces) of a mobile terminal. Combinations of multiple distal lenses and multiple cameras on one or more surfaces of a mobile terminal are also contemplated.
Each of the distal lenses 12a-1, 12a-2, 12b-1 and 12b-2 is connected to a single camera (not shown) via a connecting means such as an optical fibre. In other embodiments, connecting means may be other types of connection such as an electrical or digital connection. Using optical fibre cables has the advantage that they do not require any electrical components, such that the cables do not suffer wear and tear of electrical connections through stress as a result of frequent movement of the distal lens. The use of optical fibre cables may also provide flexibility for the movement of the distal lens if desired. The cable may be articulated or jointed to support the flexibility for movement, or be made of flexible material such as natural or synthetic rubber. In the present embodiment, the connecting means may comprise one or more small diameter, high bandwidth optical fibre cables (e.g. Corning’s Clearcurve ®).
According to the embodiments, where two or more distal lenses are used, there is no requirement for the single digital camera, to which the two or more distal lenses are connected via a respective optical fibre cable, to be placed at any specific position or orientation on the terminal. The camera may therefore be placed within the terminal if desired. In particular, the camera may be mounted perpendicular to the plane of the terminal within the terminal. This allows additional devices and accessories to be disposed within the terminal to provide additional functions.
In further embodiments, additional devices and accessories may be provided within the terminal and may include various lenses and filters (e.g. to achieve different resolutions and/or effects). The additional devices and accessories may be disposed adjacent the camera or remotely from the camera, and may be connected to one or more distal lenses and the camera by means of one or more optical fibres. According to the further embodiments, a user may select one or more different lenses and/or filters to achieve the desired result. For example, the user may achieve a desired resolution through selecting a suitable lens, or the user may achieve a desired result through combining a plurality of
- 8images each obtained by directing light from a distal lens through a different lens by means of a respective electrically switchable reflective device.
Further embodiments have been contemplated in which one or each of the optical fibre cables may be unbundled and divided into individual, or multiple groups of, component optical fibres. Each group of the unbundled optical fibres that makes up the cable may be fed to a respective component of an image sensor array. According to the present embodiments, light received at a distal lens is transmitted along each group of unbundled optical fibres to the respective component of the image sensor array, and the final composite digital image may be formed by combining the input received at each component of the image sensor array. Through the present embodiments, it is possible to configure an image sensor array by arranging component image sensors in configurations other than the conventional square or rectangular arrangements, e.g. in a narrow strip, thus enabling greater flexibility for the location of the image sensor array within the mobile terminal and imposes less limitation on the size and shape of the terminal.
Fig. 2 shows a flow diagram of a method of determining a range of an object according to an embodiment. In the embodiment, a mobile terminal, such as the mobile terminal 10A or 10B, comprising a first distal lens and a second distal lens connected to a single camera via a respective optical fibre cable is used. The first and second distal lenses are provided on the same surface of the mobile terminal, separated by a base length Y.
At block S201, a user adjusts the camera setting to focus the camera on an object through the first distal lens. At block S202, the user selects or identifies one or more points of interest on the object in an image captured through the first distal lens (first image) for use as one or more reference points. Optionally, at block S203, an astigmatiser lens is preferably provided to the mobile terminal, which is used to convert the or each reference point into a line. This then allows distance measurements with respect to the reference point to be made more accurately.
The method proceeds to block S204, when (the processing means of) the mobile terminal detects or identifies the one or more reference points from the first image obtained through the first distal lens.
In preferred embodiments, an electrically switchable reflective device may be provided to the mobile terminal, which is used to switch the input to the camera between the first distal lens and the second distal lens. At block S205, the electrically switchable reflective device
- 9is used to select the second distal lens to obtain an image through the second distal lens (second image). The processing means of the mobile terminal then detects or identifies the one or more reference points selected at block S202 in the second image at block
5206. Optionally, the one or more reference points detected or identified in the second image may similarly be converted into respective lines using the astigmatiser. At block
5207, the thus identified one or more reference points in the second image are crossmatched with the corresponding one or more reference points in the first image.
The processing means then, at block S208, computes the distance (or shift) between each pair of corresponding reference points on the first and second images. The method may optionally proceed to block 209, where blocks S204 to S208 may be repeated one or more times to obtain multiple results for averaging in order to reduce the effect of systematic errors and/or movement of the mobile terminal. Blocks S204 to S208 may optionally be repeated for multiple different reference points on the first and second images. Moreover, further images may be captured via the first and second distal lenses and corresponding one or more reference points in the further images may be used for repeating blocks S204 to S208. In embodiments where more than two distal lenses are provided on the same surface, further images may be captured via each of the plurality of distal lenses and corresponding one or more reference points in the captured images may be used for repeating blocks S204 to S208. The method then proceeds to block S210 when an average distance (or shift) between corresponding reference points on the first and second images is calculated.
The average distance calculated at block S210 is used to determine an angle of convergence Θ at block S211, which is the angle between a first line of sight from the first distal lens to the object and a second line of sight from the second distal lens to the object. The angle Θ is then used, at block S212, to compute a range of the object (the distance between the mobile terminal and the object), using the distance calculated at block S210, the base length Y that separates the first distal lens and the second distal lens, by e.g. computing the value of YTan9 (Naval Ordinance and Gunnery Volume 2, Chapter 16F4, prepared by the Department of Ordnance and Gunnery United States Naval Academy, edited and produced by the Bureau of Naval Personnel NavPers 10798-A for sale by the Superintendent of Documents, U. S. Government Printing Office Washington 25, D.C., 1958 edition revised from the 1950 edition). The computed range is displayed and/or stored, at block S213 with preselected units (e.g. meter). Details of rangefinding methods have been described in literature and will not be repeated here.
- 10Fig. 3 shows a flow diagram of a method of determining a dimension of an object according to an embodiment.
At block S301, the user selects or defines a first point and a second point on an object on the screen of a mobile terminal. The selection may, for example, be performed using a pointer to select a point on the object and clicking on the point to confirm.
Alternatively or optionally, at block S302, the user selects or defines the object (the dimension of which is to be determined) on screen by, for example, using a focus box. Then, at block S303, the terminal (or the processing means of the terminal) identifies two or more points on the object for determining one or more dimensions of the object.
At block S304, the processing means detects the first and second points on an image captured through one of the first or second distal lenses, based on the points defined at block S301 or identified at block S303.
At block S305, the difference (or separation) between the first and second points on the image is determined. A range of the object (the distance between the object and the mobile terminal) is obtained using, for example, the range determination method described above, and the range is input at block S306. Other methods may alternatively be used to obtain the range of the object for use at block S306, for example by physical measurement, determination using a map, using a database, etc. At block S307, the dimension of the object is determined using the difference or separation determined at block S305 and the range input at block S306, for example using a trigonometric method. This step may be repeated, at block S308, e.g. for further images obtained via the same lens or the other lens, to obtain multiple results for averaging in order to reduce the effect of systematic errors and movement of the mobile terminal. Additionally or alternatively, at block S309, the calculation is repeated to obtain other dimension(s) of the object (e.g. width and height) using other points identified on the object.
The determined dimension(s) can then be displayed and/or stored, at block S310, in preselected units (e.g. meters). At block S311, the determined dimension(s) is used to generate a scale grid of the object that can be displayed, superimposed on the object, or embedded in the image file of the object, e.g. for storage or sharing. At block S312, the image file with embedded scale grid is output in a predetermined file format. According to the present embodiment, the predetermined file format allows the scale grid to be resized
- 11 in direct proportion with the object when the image is being resized (e.g. by zooming in or out).
Specific implementation examples using embodiments of the present invention have been contemplated, which will be described below. It should be understood that the examples described below are for illustration purposes only and are non-exhaustive.
E-Commerce applications
Embodiments of the present disclosure may be used in e-commerce applications, for example, custom fitting apps are anticipated for clothing retailers.
In the present example, a customer may create an image of his/her body or a part of the body with an embedded scale grid as described above. The created image may be stored on the customer’s mobile terminal and/or sent to a retailer for fitting. In preferred embodiments, the retailer may provide an app on its website that can be downloaded to the customer’s mobile terminal for receiving image files of different sizes of a garment with embedded scale grid. The app can then manipulate the customer’s stored image file to generate a composite image file (e.g. a garment superimposed on the customer’s body image) to enable custom fitting.
The standards for such a collaboration between the customer’s mobile terminal and the retailer’s app may be controlled by the operating system (OS) of the manufacturer of the mobile terminal and any suitable file format interchange standards, thus a commercial arrangement for the use of the customer’s image files by the retailer may be negotiated.
The present embodiment allows the customer to visualise how a garment and other accessories may look on them and how different sizes may fit. This increases the likelihood of the customer purchasing the item. Moreover, the customer is more likely to order the correct size, thus reducing time and resources that would otherwise be wasted (and possible damage) when multiple sizes of an item are ordered and delivered, many of which are subsequently returned.
Custom 3D Printing/Manufacturing
Embodiments of the present disclosure may be applied to custom 3D printing manufacturing, to provide custom-made products unique to customer requirements.
- 12An example of this may be the custom production of earphone/plugs. Custom production currently requires multiple visits to a location where measurements are made and then, fitting must be done when the product is shipped to the location.
In this application, accurate measurement of the dimensions of an object is required. It is anticipated that a customer uses a mobile terminal to create an image file or series of image files with embedded scale grid of an object to be made, or an object for which a product is to be made to fit, for example custom-made ear plugs or ear phones. The image file may be uploaded to the manufacturer’s website, where image recognition software may create a 3D image (or mirror image), based on the embedded dimensions of the object, to generate a file to be sent to a 3D printing device (in scale). The custom produced product may then be packaged and shipped to the customer without the need for the customer to be physically present for measurements.
According to the present embodiment, the customer may create one or more image files with embedded scale grid as described above, and upload the image files to the manufacturer’s website, where a 3D image of each ear is generated and then earplug mapped to the thus generated 3D image. Instructions for producing the mapped earplug may then be passed to a 3D printer for both left and then right ears. Colour/patterns may also be customised by the customer if desired. The earplugs may be a carrier for Bluetooth speakers or other communications devices, or a noise reduction device.
3D Images
Further embodiments have been contemplated in which input to the camera may be rapidly (electrically) switched between the two (or more if desired) distal lenses, provided on the same surface of the mobile terminal, to generate stereoscopic images. The present embodiments make use of the slight difference in the line of sight from an object to each of the two distal lenses, which are separated by a predetermined distance, to generate a stereoscopic image of the object on screen that the user may perceive as threedimensional. The generated 3D image(s) may be stored, manipulated and/or shared with third parties as desired.
- 13Remote Diagnostics
According to preferred embodiments, an image of an object may be stored with an embedded scale grid in a predetermined file format, where the embedded scale grid resizes in direct proportion with the object such that the dimension of the object may be read from the scale grid. Thus, further embodiments have been contemplated in which a stored image file of e.g. a body part, may be sent to a third party, e.g. a medical expert, for determination of one or more dimensions with respect to the body part, e.g. a wound. According to present embodiments, it is possible for the medical expert at a remote location to perform initial assessment of, for example, the seriousness of an injury, to determine appropriate treatment for the injury.
Determination of Speed and Acceleration
Further embodiments have been contemplated in which a sequence of images of an object moving relative to the mobile terminal may be captured. This may include cases in which the user is on the move (e.g. in a moving vehicle), the object is moving, and both the user and the object are moving relative to each other. Each image of the sequence may be time-stamped. The time difference between each pair of images may be determined, and the range of the object may be obtained according to embodiments of the rangefinding method described above for each image of the sequence, to compute the speed at which the object moves relative to the user. If desired, the acceleration of the object relative to the user may also be computed by determining a sequence of speed using the sequence of images.
Biometric
Further embodiments have been contemplated in which biometric dimensions may be determined from, e.g. images of fingerprint, facial features, etc., in addition to pattern recognition to enhance the accuracy and security of biometric identification systems.
Although the present invention has been described with reference to preferred embodiments, it will be understood by those skilled in the art that various changes in form and detail may be made without departing from the scope of the invention as set forth in the accompanying claims.
- 14Claims
1. A method of determining a distance using a mobile terminal that comprises first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal, the first and second distal lenses being connected to a camera by a respective connecting means, the method comprising:
capturing, by the camera, a first image of an object from the first distal lens; capturing, by the camera, a second image of the object from the second distal lens; comparing the first image with the second image; and determining a distance in relation to the object based on a result of comparing the first image with the second image using the predetermined distance.
2. The method of determining a distance according to claim 1, wherein comparing the first image with the second image comprises determining an angle of convergence between the first image and the second image, the angle of convergence being an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens.
3. The method of determining a distance according to claim 1 or 2, further comprising: selecting a reference point on the object in the first image; and identifying the reference point on the object in the second image, wherein comparing the first image with the second image comprises detecting a difference between a first position of the reference point within the first image and a second position of the reference point within the second image.
4. The method of determining a distance according to claim 3, wherein comparing the first image with the second image further comprises determining an angle of convergence using the detected difference between the first position and the second position, the angle of convergence being an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens.
5. The method of determining a distance according to claim 2 or 4, wherein determining a distance in relation to the object comprises determining a range from the mobile terminal to the object using the predetermined distance and the angle of convergence.
6. The method of determining a distance according to claim 5, wherein determining a distance in relation to the object further comprises:
determining a second range from the mobile terminal to the object using the predetermined distance and a second angle of convergence determined by comparing a third image of an object captured by the camera from the first distal lens with a fourth image of the object captured by the camera from the second distal lens; and averaging the range and the second range.
7. The method of determining a distance according to claim 5 or 6, wherein determining a distance in relation to the object further comprises determining a dimension of the object based on the determined range from the mobile terminal to the object.
8. The method of determining a distance according to claim 7, wherein determining the dimension of the object comprises:
selecting, in one of the first image or the second image, a first point on the object and a second point on the object spaced from the first point;
determining a difference between the first point and the second point within the first or second image; and determining the dimension of the object based on the determined difference between the first point and the second point.
9. The method of determining a distance according to claim 7 or 8, wherein determining a distance in relation to the object further comprises determining a second dimension of the object based on the determined range from the mobile terminal to the object.
10. The method of determining a distance according to claim 9, wherein determining the second dimension of the object comprises:
selecting, in one of the first image or the second image, a third point on the object and a fourth point on the object spaced from the third point;
determining a difference between the third point and the fourth point within the first or second image; and determining the second dimension of the object based on the determined difference between the third point and the fourth point.
11. The method of determining a distance according to any one of claims 7 to 10, further comprising generating a scale grid based on the determined dimension of the object.
12. The method of determining a distance according to claim 11, further comprising embedding the generated scale grid in the first image or the second image.
13. The method of determining a distance according to claim 12, further comprising storing the first image or the second image with the embedded scale grid in a predetermined format.
14. The method of determining a distance according to any one of claims 11 to 13, further comprising displaying, on a screen of the mobile terminal, the generated scale grid superimposed on the first image or the second image.
15. A mobile terminal comprising:
a camera configured to capture an image;
first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal;
connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light from each distal lens to the camera; and processing means configured to perform a method of determining a distance according to any preceding claim.
16. A mobile terminal comprising:
a camera configured to capture an image;
first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal;
connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light from each distal lens to the camera; and processing means configured to compare an image of an object captured via the first distal lens with an image of the object captured via the second distal lens, and determine a distance in relation to the object based on the comparison using the predetermined distance.
17. A mobile terminal comprising:
first and second cameras each configured to capture an image, the first and second cameras being disposed spaced apart by a predetermined distance on the same surface of the mobile terminal; and processing means configured to compare an image of an object captured by the first camera with an image of the object captured by the second camera, and determine a distance in relation to the object based on the comparison using the predetermined distance.
Claims (14)
- Claims5 1. A method of determining a distance using a mobile terminal that comprises first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal, the first and second distal lenses being connected to a camera by a respective connecting means, the method comprising: capturing, by the camera, a first image of an object from the first distal lens;10 capturing, by the camera, a second image of the object from the second distal lens;comparing the first image with the second image; and determining a distance in relation to the object based on a result of comparing the first image with the second image using the predetermined distance by determining an angle of convergence between the first image and the second image, the angle of15 convergence being an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens, determining a range from the mobile terminal to the object using the predetermined distance and the angle of convergence, and determining a dimension of the object based on the determined range from the mobile terminal to the object.
- 2. The method of determining a distance according to claim 1, further comprising: selecting a reference point on the object in the first image; and identifying the reference point on the object in the second image, wherein comparing the first image with the second image comprises detecting a25 difference between a first position of the reference point within the first image and a second position of the reference point within the second image.
- 3. The method of determining a distance according to claim 2, wherein the angle of convergence is determined using the detected difference between the first position and the30 second position.
- 4. The method of determining a distance according to any preceding claim, wherein determining a distance in relation to the object further comprises:determining a second range from the mobile terminal to the object using the35 predetermined distance and a second angle of convergence determined by comparing a third image of an object captured by the camera from the first distal lens with a fourth image of the object captured by the camera from the second distal lens; and averaging the range and the second range.40 5. The method of determining a distance according to any preceding claim, wherein determining the dimension of the object comprises:20 06 18 selecting, in one of the first image or the second image, a first point on the object and a second point on the object spaced from the first point;determining a difference between the first point and the second point within the first or second image; and
- 5 determining the dimension of the object based on the determined difference between the first point and the second point.
- 6. The method of determining a distance according to any preceding claim, wherein determining a distance in relation to the object further comprises determining a second10 dimension of the object based on the determined range from the mobile terminal to the object.
- 7. The method of determining a distance according to claim 6, wherein determining the second dimension of the object comprises:15 selecting, in one of the first image or the second image, a third point on the object and a fourth point on the object spaced from the third point;determining a difference between the third point and the fourth point within the first or second image; and determining the second dimension of the object based on the determined difference 20 between the third point and the fourth point.
- 8. The method of determining a distance according to any preceding claim, further comprising generating a scale grid based on the determined dimension of the object.25
- 9. The method of determining a distance according to claim 8, further comprising embedding the generated scale grid in the first image or the second image.
- 10. The method of determining a distance according to claim 9, further comprising storing the first image or the second image with the embedded scale grid in a30 predetermined format.
- 11. The method of determining a distance according to any one of claims 8 to 10, further comprising displaying, on a screen of the mobile terminal, the generated scale grid superimposed on the first image or the second image.
- 12. A mobile terminal comprising:a camera configured to capture an image;first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal;40 connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light from each distal lens to the camera; and20 06 18 processing means configured to perform a method of determining a distance according to any preceding claim.
- 13. A mobile terminal comprising:5 a camera configured to capture an image;first and second distal lenses disposed spaced apart by a predetermined distance on the same surface of the mobile terminal;connecting means provided to each of the first and second distal lenses, the connecting means connecting each distal lens to the camera and configured to direct light 10 from each distal lens to the camera; and processing means configured to compare an image of an object captured via the first distal lens with an image of the object captured via the second distal lens, and determine a distance in relation to the object based on the comparison using the predetermined distance by determining an angle of convergence between the first image 15 and the second image, the angle of convergence being an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens, determine a range from the mobile terminal to the object using the predetermined distance and the angle of convergence, and determine a dimension of the object based on the determined range from the mobile terminal to the 20 object.
- 14. A mobile terminal comprising:first and second cameras each configured to capture an image, the first and second cameras being disposed spaced apart by a predetermined distance on the same surface of 25 the mobile terminal; and processing means configured to compare an image of an object captured by the first camera with an image of the object captured by the second camera, and determine a distance in relation to the object based on the comparison using the predetermined distance by determining an angle of convergence between the first image and the second 30 image, the angle of convergence being an angle between a first line of sight to the object with respect to the first distal lens and a second line of sight to the object with respect to the second distal lens, determine a range from the mobile terminal to the object using the predetermined distance and the angle of convergence, and determine a dimension of the object based on the determined range from the mobile terminal to the object.
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| GB1714716.6A GB2566462A (en) | 2017-09-13 | 2017-09-13 | Mobile terminal |
| US16/491,644 US11252314B2 (en) | 2017-03-08 | 2018-03-08 | Mobile terminal with distal lenses and method of using mobile terminal to determine distance, range and dimension of an object |
| EP18711643.9A EP3593523A1 (en) | 2017-03-08 | 2018-03-08 | Mobile terminal |
| CN201880017266.3A CN110431833A (en) | 2017-03-08 | 2018-03-08 | mobile terminal |
| PCT/GB2018/050590 WO2018162909A1 (en) | 2017-03-08 | 2018-03-08 | Mobile terminal |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| GB1714716.6A GB2566462A (en) | 2017-09-13 | 2017-09-13 | Mobile terminal |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| GB201714716D0 GB201714716D0 (en) | 2017-10-25 |
| GB2566462A true GB2566462A (en) | 2019-03-20 |
Family
ID=60117332
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| GB1714716.6A Withdrawn GB2566462A (en) | 2017-03-08 | 2017-09-13 | Mobile terminal |
Country Status (1)
| Country | Link |
|---|---|
| GB (1) | GB2566462A (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2000078055A1 (en) * | 1999-06-11 | 2000-12-21 | Emile Hendriks | Acquisition of 3-d scenes with a single hand held camera |
| US20020135468A1 (en) * | 1997-09-22 | 2002-09-26 | Donnelly Corporation, A Corporation Of The State Of Michigan | Vehicle imaging system with accessory control |
| WO2014002492A1 (en) * | 2012-06-28 | 2014-01-03 | Sony Corporation | Image processing apparatus, imaging apparatus, and image processing method |
| CN106355621A (en) * | 2016-09-23 | 2017-01-25 | 邹建成 | Method for acquiring depth information on basis of array images |
-
2017
- 2017-09-13 GB GB1714716.6A patent/GB2566462A/en not_active Withdrawn
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020135468A1 (en) * | 1997-09-22 | 2002-09-26 | Donnelly Corporation, A Corporation Of The State Of Michigan | Vehicle imaging system with accessory control |
| WO2000078055A1 (en) * | 1999-06-11 | 2000-12-21 | Emile Hendriks | Acquisition of 3-d scenes with a single hand held camera |
| WO2014002492A1 (en) * | 2012-06-28 | 2014-01-03 | Sony Corporation | Image processing apparatus, imaging apparatus, and image processing method |
| CN106355621A (en) * | 2016-09-23 | 2017-01-25 | 邹建成 | Method for acquiring depth information on basis of array images |
Also Published As
| Publication number | Publication date |
|---|---|
| GB201714716D0 (en) | 2017-10-25 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20140307100A1 (en) | Orthographic image capture system | |
| US10116867B2 (en) | Method and apparatus for displaying a light field based image on a user's device, and corresponding computer program product | |
| CN113412614A (en) | Three-dimensional localization using depth images | |
| JP6657214B2 (en) | Accuracy measurement of image-based depth detection system | |
| KR20200049502A (en) | System and method for disparity estimation using cameras with different fields of view | |
| US10726580B2 (en) | Method and device for calibration | |
| KR102452575B1 (en) | Apparatus and method for compensating variation of images caused by optical image stabilization motion | |
| CN105324791B (en) | Coordinate computing device and method and image processing apparatus and method | |
| JP6312227B2 (en) | RGB-D imaging system, RGB-D image generation method, and apparatus for generating RGB-D image | |
| US10347029B2 (en) | Apparatus for measuring three dimensional shape, method for measuring three dimensional shape and three dimensional shape measurement program | |
| KR20130041297A (en) | Image processing device, image processing method and information recording medium | |
| KR20190122018A (en) | Electronic Device and the Method for Generating Depth Information thereof | |
| WO2014084181A1 (en) | Image measurement device | |
| US10055817B2 (en) | Systems and methods for determining pupillary distance and scale | |
| WO2020235067A1 (en) | Three-dimensional measurement system and three-dimensional measurement method | |
| KR20190019606A (en) | An apparatus for composing objects using depth map and a method thereof | |
| US11252314B2 (en) | Mobile terminal with distal lenses and method of using mobile terminal to determine distance, range and dimension of an object | |
| CN113939852A (en) | Object recognition device and object recognition method | |
| JP2002056379A (en) | 3D data generator | |
| JP5487946B2 (en) | Camera image correction method, camera apparatus, and coordinate transformation parameter determination apparatus | |
| JP7040660B1 (en) | Information processing equipment and information processing method | |
| GB2566462A (en) | Mobile terminal | |
| EP2866446B1 (en) | Method and multi-camera portable device for producing stereo images | |
| JP2020148712A (en) | Ranging system, ranging device, and arrangement method | |
| JP6868167B1 (en) | Imaging device and imaging processing method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| WAP | Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1) |