Can someone who's tried the Oculus Rift tell me how it works in the following sense:
If I have a PC monitor at hands distance, obviously my eyes are focusing at the distance of the monitor. Or if I watch the sky, they are focusing at "infinity".
One of the problems with these devices is that since the panel/lcd in the Rift is physically only ~inch away from your eyes, unless there is something that two ophthalmologists I asked claimed to be impossible, your eyes are focusing just a couple inch away (bit like looking at ones nose). And I suspect that if there was some "magical optics" to solve this, they might need invidual adjustment, possibly dedicated adjustments for both eyes, atleast in my case, unless one is supposed to wear corrective lenses also with them.
I read hints suggesting that in some of the military planes and windshield displays they have somehow solved this, allowing one to focus "at the sky" while somehow projecting the CGI to a surface like the windshield or helmet visor. Allowing to keep focus at infinity while being able to read the stuff that's being projected from/at close distance (but "at infinity", as far as focusing ones eyes goes).
I've long wanted such technology as if one could swap from "infinite focus" to "near focus" while keeping the CGI-text readable, this would solve some of the reading related 'lazy focus' issues that come from extensive focusing of eyes at certain distance as the eye focus mechanism isn't getting practised enough.
If Kinects big deal is end to end response time/latency, I think the above focusing issue is the classic big deal with wearable computer displays (not panel resolution which some might believe is an issue - visible spacing between pixels is another real issue but irrelevant unless the focus issue is also solved). Has Oculus Rift really solved that or have I been right to ignore this hype?