NEAR-EYE DISPLAY SYSTEM AND METHOD
1. A near-eye display system, comprising:
- a display component comprising a light source array disposed on one or more actuators;
an actuator control component communicatively coupled to the one or more actuators; and
a processor unit communicatively coupled to the light source array and the actuator control component;
wherein the light source array is connected with the one or more actuators such that the one or more actuators are capable of moving the light source array in accordance with a scan pattern, and the processor unit is configured to synchronize the illumination of a plurality of pixels of the light source array with the scan pattern.
A near-eye light field display for use with a head mounted display unit with enhanced resolution and color depth. A display for each eye is connected to one or more actuators to scan each display, increasing the resolution of each display by a factor proportional to the number of scan points utilized. In this way, the resolution of near-eye light field displays is enhanced without increasing the size of the displays.
- 1. A near-eye display system, comprising:
a display component comprising a light source array disposed on one or more actuators; an actuator control component communicatively coupled to the one or more actuators; and a processor unit communicatively coupled to the light source array and the actuator control component; wherein the light source array is connected with the one or more actuators such that the one or more actuators are capable of moving the light source array in accordance with a scan pattern, and the processor unit is configured to synchronize the illumination of a plurality of pixels of the light source array with the scan pattern.
- View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13)
- 14. A method, comprising:
measuring by one or more focus sensors an eye focus of a user'"'"'s eyes; determining by a focus correction modules a desired focus based on the measured eye focus; setting a focus of a first camera and a second camera based on the desired focus; capturing images within a field of view of the first camera and the second camera; processing the captured images; setting a focus of a first display and a second display; and providing the captured images to the first display and a second display; wherein the first display is disposed in front of a first eye of the user and the second display is disposed in front of a second eye of the user.
- View Dependent Claims (15, 16, 17, 18, 19, 20, 21, 22)
The disclosed technology relates generally to near-eye displays, and more particularly, some embodiments relate to near-eye displays with reduced size and improved resolution and color.
Head-mounted displays (“HMDs”) are generally configured such that one or more displays are placed directly in front of a person'"'"'s eyes. HMDs have been utilized in various applications, including gaming, simulation, and military uses. Traditionally, HMDs have comprised heads-up displays, wherein the user focuses on the display in front of the eyes, as images are traditionally displayed on a two-dimensional (“2D”) surface. Optics are used to make the display(s) appear farther away than it actually is, in order to allow for a suitable display size to be utilized so close to the human eye. Despite the use of optics, however, HMDs generally have low resolution because of trade-offs related to the overall weight and form factor of the HMD, as well as pixel pitch.
According to various embodiments of the disclosed technology, an improved near-eye display with enhanced resolution is provided. The near-eye display system includes a display component comprising a light source array disposed on one or more actuators; an actuator control component communicatively coupled to the one or more actuators; and a processor unit communicatively coupled to the light source array and that actuator control component, wherein the light source array is connected with the one or more actuators such that the one or more actuators are capable of moving the light source array in accordance with a scan pattern, and the processor unit is configured to synchronize the illumination of a plurality of pixels of the light source array with the scan pattern. In some embodiments, more than one display components may be implemented in a HMD.
According to an embodiment of the disclosed technology, a method of providing a near-eye light field display with enhanced resolution is provided. The method includes measuring by one or more focus sensors an eye focus of a user'"'"'s eyes; determining by a focus correction modules a desired focus based on the measured eye focus; setting a focus of a first camera and a second camera based on the desired focus; capturing images within a field of view of the first camera and the second camera; processing the captured images; setting a focus of a first display and a second display; and providing the captured images to the first display and a second display; wherein the first display is disposed in front of a first eye of the user and the second display is disposed in front of a second eye of the user.
Other features and aspects of the disclosed technology will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, which illustrate, by way of example, the features in accordance with embodiments of the disclosed technology. The summary is not intended to limit the scope of any inventions described herein, which are defined solely by the claims attached hereto.
The technology disclosed herein, in accordance with one or more various embodiments, is described in detail with reference to the following figures. The drawings are provided for purposes of illustration only and merely depict typical or example embodiments of the disclosed technology. These drawings are provided to facilitate the reader'"'"'s understanding of the disclosed technology and shall not be considered limiting of the breadth, scope, or applicability thereof. It should be noted that for clarity and ease of illustration these drawings are not necessarily made to scale.
The figures are not intended to be exhaustive or to limit the invention to the precise form disclosed. It should be understood that the invention can be practiced with modification and alteration, and that the disclosed technology be limited only by the claims and the equivalents thereof.
As discussed above, HMDs generally employ one or more displays placed in front of the human eye. 2D images are shown on the displays, and the eye focuses on the display itself. In order to provide a clear, focused image, optics placed between the eye and the display make the display appear farther away than the display may actually be in reality. In this way, the eye is capable of focusing beyond the space occupied by the display.
HMDs are generally either too large, have limited resolution, or a combination of both. This is due to the distance between pixels in the display, or pixel pitch. When there is sufficient distance between the display and the eye, pixel pitch does not impact resolution to a great extent, as the space between pixels is not as noticeable. However, HMDs place displays near the eye, making pixel pitch an important limiting factor related to resolution. In order to increase resolution, larger displays are necessary to increase the number of pixels in the display. Larger displays require larger optics to create the illusion of space between the eye and the display.
Traditionally, the display in a HMD is a 2D surface projecting a 2D image to each eye. Some HMDs utilize waveguides in an attempt to simulate 3D images. Waveguides, however, are complex, requiring precise design and manufacture to avoid errors in beam angle and beam divergence.
One solution that provides true three-dimensional images is the use of near-eye light field displays. Similar to light field cameras, a near-eye light field display creates a representation of light as it crosses a plane that provides information not only relating to the intensity of the light, but also to the direction of the light rays. Traditional 2D displays only provide information regarding the intensity of light on a plane. Some light field HMDs under development utilize waveguides with diffractive optical elements in an attempt to synthesize light rays on a plane. However, this approach is complex, requiring precise design and manufacture to avoid errors in light ray angle and divergence.
Embodiments of the technology disclosed herein are directed toward systems and methods for near-eye light field displays. More particularly, the various embodiments of the technology disclosed herein relate to near-eye light field displays providing enhanced resolution and color depth compared to conventional near-eye light field displays. As will be described in greater detail below, embodiments of the technology disclosed herein enable greater resolution and color by laterally moving a light source array, such as an LED or OLED display, while controlling the intensity of the light source array in synchronization with a scan pattern. Scanning the light source array while modulating the intensity of the pixels reduces the impact of pixel pitch, resulting in increased resolution without the need for larger displays or optics. In one embodiment, the intensity modulation of the pixels is achieved by turning the pixels on and off for a time duration that is dependent on the desired intensity. For example, if higher intensity is desired on a red pixel than on a blue pixel, the red pixel would be lit up longer than the blue pixel. Moreover, various embodiments employ a light-field display with enhanced resolution and 3D images without the need for complicated waveguides.
By employing embodiments of the systems and methods described below, it is possible to reduce the size and/or enhance the resolution and color of traditional HMDs, or convert a traditional HMD into a light field display.
As discussed above, various embodiments of the technology disclosed herein utilize synchronized illumination of the display during scanning while moving the light source array to eliminate the space between pixels.
The light source array 910 may be disposed on one or more actuators 920. In various embodiments, the one or more actuators 910 may include one or more of: voice coil motors (“VCMs”); shape memory alloy (“SMA”) actuators; piezoelectric actuators; MEMS actuators; a combination thereof; among others. In some examples, the one or more actuators 910 may comprise a MEMS actuator similar to the MEMS actuator disclosed in U.S. patent application Ser. No. 14/630,437, filed Feb. 24, 2015, the disclosure of which is hereby incorporated herein by reference in its entirety. Non-limiting examples of material for connecting the light source array 910 to the one or more actuators include: epoxy; solder; metal pastes; wire bonding; among others. To control the actuators 920, the improved near-eye display system 900 may include actuator control components 930, including electronics for controlling the actuators and sensors identifying the position of the actuators 920. In various embodiments, the actuator control components 930 may be communicatively coupled to the processor unit 940 so that illumination of the light source array 910 may be synchronized with the movement of the one or more actuators 920.
The improved near-eye display system 900 is capable of enhancing spatial resolution of images, i.e., how closely lines can be resolved in an image. In terms of pixels, the greater the number of pixels per inch (“ppi”), the clearer the image that may be resolved. When operated in accordance with the scan pattern discussed with respect to
The improved near-eye display system discussed with respect to
As discussed above,
The images from the one or more cameras 210 may be fed into processor unit 220 to compute the actual image at different depths based on the captured images. For example, where a traditional camera is used, the images from the one or more cameras 210 may comprise a series of pictures at different focal depths. To create the three dimensional actual image, the different captured images are processed to provide depth to the actual image.
To achieve enhanced resolution, the light field system 200 may include one or more display components 270, similar to the display components 950 discussed with respect to
The captured light field discussed above with respect to the one or more cameras 210 is utilized to compute a sequence for a driver associated with the light source array 260, indicating when the plurality of light sources comprising the light source array 260 should be turned on. In some embodiments, the light field system 200 may include one or more gyroscopes or accelerometers 230, providing information representative of the particular position of the user'"'"'s head. Furthermore, the images captured by the cameras may be processed to determine the motion of the user'"'"'s head as well. This information may be fed into the processor unit 220 to utilize in computing the light field to account for changes in the position of the user'"'"'s head.
To provide depth to the image, one or more actuators for scanning the light source array in the Z-direction may be included. By scanning in the Z-direction, the generated light field may be provided at the proper focus. Scanning in the Z-direction may be incorporated into the scan pattern of
Another benefit of scanning the display is the ability to include different sensors within the display without sacrificing resolution.
In order to ensure that images are processed within the persistence of vision of the human eye, scanning in the three dimensions may need to occur at high frequency. For example, in some embodiments the focus scan (Z-direction) may be provided by a 25 Hz frequency triangle wave to ensure a 50 Hz update rate for the entire image. Accordingly, the lateral scan (X-Y plane) may require a 500 Hz frequency scan in the X-axis and 5 kHz scan in the Y-axis to ensure that the entire image is displayed within one depth of focus. In some embodiments, the frequency of X-Y scanning may be reduced by simplifying the design.
To reduce the complexity associated with three-axis scanning as described above with respect to the light field system of
The focus of each camera 602, 604 may be controlled based on the focus of the user'"'"'s eyes. The basic light field system 600 may include eye focus sensors 608, 610 disposed within a display in front of the left eye and right eye, respectively. Each eye focus sensor 608, 610 may include one or more focus sensors in various embodiments. In some embodiments, the eye focus sensors 608, 610 may be disposed in the spaces between the pixels of a left display 612 and a right display 614. The eye focus sensors 608, 610 may be used to determine where a user'"'"'s eyes are focused. The information from the eye focus sensors 608, 610 may be fed into a focus correction module 616. The focus correction module 616 may determine the correct focus based on the point where the user'"'"'s eyes are focused, and provide this information to a display focus control 618. The display focus control 618 may provide this information to the camera focus control 606. The camera focus control 606 may utilize the focus information from the display focus control 618 to set the focus of each camera 602, 604. The vision of a user with eye focus problems (myopia or hyperopia, nearsighted or farsighted) can be corrected by setting the focus of the cameras to a different depth than the focus of the display. In some embodiments, the cameras 602, 604 may be one or more of a light field camera, a standard camera, an infrared camera, or some other image sensor, or a combination thereof. For example, in some embodiments the cameras 602, 604 may comprise a standard camera and an infrared camera, enabling the basic light field system 600 to provide both a normal view and an infrared view to the user.
The display focus control 618 may also utilize the desired focus from the focus correction module 616 to set the focus of the displays 612, 614, to the focus of each eye.
Once the cameras 602, 604 are set to the desired focus, the cameras 602, 604 may capture the scene within the field of view of each camera 602, 604. The images from each camera 602, 604 may be processed by a processor unit 620, 622. As illustrated, each camera 602, 604 has its own processor unit 620, 622, respectively. In some embodiments, a single processor unit may be employed for both cameras 602, 604. The processor unit 620, 622 may process the images from each camera 602, 604 in a similar fashion as described above with respect to
Although illustrated as separate components, aspects of the basic light field system 600 may be implemented as in a single component. For example, the focus correction 616, the display focus control 618, and the camera focus control 606 may be implemented in software and executed by a processor, such as processor unit 620, 622.
At 720, a desired focus is determined. The desired focus is determined based on the measured eye focus from 710. The desired focus may be different from the eye focus if the user has focus problems. For example, if the user has myopia (nearshightedness), the desired focus is further away than the measured eye focus. The desired focus may also be determined from the position of the eye, such as close if looking down, or the position of the eye with respect to the image, such as the same focus as a certain object in the scene, or some of other measurement of the eye. Based on the desired focus, the camera focus may be set to the desired focus at 730. In various embodiments, the camera focus may be set equal to the desired focus. In other embodiments, the camera focus may be set to a focus close to, but not equal to, the desired focus. In such embodiments, the camera focus may be set as close as possible based on the type of camera employed in the embodiment.
At 740, the cameras capture images of objects within the field of view of the cameras. In some embodiments, the field of view of the cameras may be larger than the displayed field of view to enable some ability to quickly update the display when there is rapid head movement, without the need for capturing a new image.
At 750, each display is set to the eye focus. In some embodiments, the eye focus is the same as the desired focus. In other embodiments, the desired focus is derived from the eye focus identified at 710. In some embodiments, the displays may be set to the eye focus before setting the camera focus at 730, after 730 but before the camera captures images at 740, or simultaneous to the actions at 730 and/or 740.
At 760, the images are displayed to each eye. The images are displayed to each eye via the respective display. In some embodiments, the images may be processed by a processor unit prior to being displayed, similar to the processing discussed above with respect to
As used herein, the term component might describe a given unit of functionality that can be performed in accordance with one or more embodiments of the technology disclosed herein. As used herein, a component might be implemented utilizing any form of hardware, software, or a combination thereof. For example, one or more processors, controllers, ASICs, PLAs, PALs, CPLDs, FPGAs, logical components, software routines or other mechanisms might be implemented to make up a component. In implementation, the various components described herein might be implemented as discrete components or the functions and features described can be shared in part or in total among one or more components. In other words, as would be apparent to one of ordinary skill in the art after reading this description, the various features and functionality described herein may be implemented in any given application and can be implemented in one or more separate or shared components in various combinations and permutations. Even though various features or elements of functionality may be individually described or claimed as separate components, one of ordinary skill in the art will understand that these features and functionality can be shared among one or more common software and hardware elements, and such description shall not require or imply that separate hardware or software components are used to implement such features or functionality.
Where components of the technology are implemented in whole or in part using software, in one embodiment, these software elements can be implemented to operate with a computing or processing component capable of carrying out the functionality described with respect thereto. One such example computing component is shown in
Referring now to
Computing component 800 might include, for example, one or more processors, controllers, control modules, or other processing devices, such as a processor 804. Processor 804 might be implemented using a general-purpose or special-purpose processing engine such as, for example, a microprocessor, controller, or other control logic. In the illustrated example, processor 804 is connected to a bus 802, although any communication medium can be used to facilitate interaction with other components of computing component 800 or to communicate externally.
Computing component 800 might also include one or more memory components, simply referred to herein as main memory 808. For example, preferably random access memory (RAM) or other dynamic memory, might be used for storing information and instructions to be executed by processor 804. Main memory 808 might also be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 804. Computing component 800 might likewise include a read only memory (“ROM”) or other static storage device coupled to bus 802 for storing static information and instructions for processor 804.
The computing component 800 might also include one or more various forms of information storage mechanism 810, which might include, for example, a media drive 812 and a storage unit interface 820. The media drive 812 might include a drive or other mechanism to support fixed or removable storage media 814. For example, a hard disk drive, a floppy disk drive, a magnetic tape drive, an optical disk drive, a CD or DVD drive (R or RW), or other removable or fixed media drive might be provided. Accordingly, storage media 814 might include, for example, a hard disk, a floppy disk, magnetic tape, cartridge, optical disk, a CD or DVD, or other fixed or removable medium that is read by, written to or accessed by media drive 812. As these examples illustrate, the storage media 814 can include a computer usable storage medium having stored therein computer software or data.
In alternative embodiments, information storage mechanism 810 might include other similar instrumentalities for allowing computer programs or other instructions or data to be loaded into computing component 800. Such instrumentalities might include, for example, a fixed or removable storage unit 822 and an interface 820. Examples of such storage units 822 and interfaces 820 can include a program cartridge and cartridge interface, a removable memory (for example, a flash memory or other removable memory module) and memory slot, a PCMCIA slot and card, and other fixed or removable storage units 822 and interfaces 820 that allow software and data to be transferred from the storage unit 822 to computing component 800.
Computing component 800 might also include a communications interface 824. Communications interface 824 might be used to allow software and data to be transferred between computing component 800 and external devices. Examples of communications interface 824 might include a modem or softmodem, a network interface (such as an Ethernet, network interface card, WiMedia, IEEE 802.XX or other interface), a communications port (such as for example, a USB port, IR port, RS232 port Bluetooth® interface, or other port), or other communications interface. Software and data transferred via communications interface 824 might typically be carried on signals, which can be electronic, electromagnetic (which includes optical) or other signals capable of being exchanged by a given communications interface 824. These signals might be provided to communications interface 824 via a channel 828. This channel 828 might carry signals and might be implemented using a wired or wireless communication medium. Some examples of a channel might include a phone line, a cellular link, an RF link, an optical link, a network interface, a local or wide area network, and other wired or wireless communications channels.
In this document, the terms “computer program medium” and “computer usable medium” are used to generally refer to media such as, for example, memory 808, storage unit 820, media 814, and channel 828. These and other various forms of computer program media or computer usable media may be involved in carrying one or more sequences of one or more instructions to a processing device for execution. Such instructions embodied on the medium, are generally referred to as “computer program code” or a “computer program product” (which may be grouped in the form of computer programs or other groupings). When executed, such instructions might enable the computing component 800 to perform features or functions of the disclosed technology as discussed herein.
While various embodiments of the disclosed technology have been described above, it should be understood that they have been presented by way of example only, and not of limitation. Likewise, the various diagrams may depict an example architectural or other configuration for the disclosed technology, which is done to aid in understanding the features and functionality that can be included in the disclosed technology. The disclosed technology is not restricted to the illustrated example architectures or configurations, but the desired features can be implemented using a variety of alternative architectures and configurations. Indeed, it will be apparent to one of skill in the art how alternative functional, logical or physical partitioning and configurations can be implemented to implement the desired features of the technology disclosed herein. Also, a multitude of different constituent module names other than those depicted herein can be applied to the various partitions. Additionally, with regard to flow diagrams, operational descriptions and method claims, the order in which the steps are presented herein shall not mandate that various embodiments be implemented to perform the recited functionality in the same order unless the context dictates otherwise.
Although the disclosed technology is described above in terms of various exemplary embodiments and implementations, it should be understood that the various features, aspects and functionality described in one or more of the individual embodiments are not limited in their applicability to the particular embodiment with which they are described, but instead can be applied, alone or in various combinations, to one or more of the other embodiments of the disclosed technology, whether or not such embodiments are described and whether or not such features are presented as being a part of a described embodiment. Thus, the breadth and scope of the technology disclosed herein should not be limited by any of the above-described exemplary embodiments.
Terms and phrases used in this document, and variations thereof, unless otherwise expressly stated, should be construed as open ended as opposed to limiting. As examples of the foregoing: the term “including” should be read as meaning “including, without limitation” or the like; the term “example” is used to provide exemplary instances of the item in discussion, not an exhaustive or limiting list thereof; the terms “a” or “an” should be read as meaning “at least one,” “one or more” or the like; and adjectives such as “conventional,” “traditional,” “normal,” “standard,” “known” and terms of similar meaning should not be construed as limiting the item described to a given time period or to an item available as of a given time, but instead should be read to encompass conventional, traditional, normal, or standard technologies that may be available or known now or at any time in the future. Likewise, where this document refers to technologies that would be apparent or known to one of ordinary skill in the art, such technologies encompass those apparent or known to the skilled artisan now or at any time in the future.
The presence of broadening words and phrases such as “one or more,” “at least,” “but not limited to” or other like phrases in some instances shall not be read to mean that the narrower case is intended or required in instances where such broadening phrases may be absent. The use of the term “module” does not imply that the components or functionality described or claimed as part of the module are all configured in a common package. Indeed, any or all of the various components of a module, whether control logic or other components, can be combined in a single package or separately maintained and can further be distributed in multiple groupings or packages or across multiple locations.
Additionally, the various embodiments set forth herein are described in terms of exemplary block diagrams, flow charts and other illustrations. As will become apparent to one of ordinary skill in the art after reading this document, the illustrated embodiments and their various alternatives can be implemented without confinement to the illustrated examples. For example, block diagrams and their accompanying description should not be construed as mandating a particular architecture or configuration.