Week1: different platform for Virtual and Augmented Reality


In my opinion, one of the advantages of Vive is that it requires a real space rather than creates a virtual space.To explore the space, we need walk in the space or change the body posture. This makes the virtual circumstance more engagable and immersive. Another point is that the Vive handle gives us the real feedback of the object we touch or seize. This character gives much more interaction with the virtual world. Last but not the least, as a potential developer, Viveport is open to all developers under the submission guidlines and tech standards.
One of the disadvantages is obvious, it requires a space to perform if we do not want to destroy anything in our home like the video shows in the class.
Most applicable: art, creativity, design, education, health, etc.


The most amazing thing of HoloLens is that it is controled by gesture. Through the game in HoloLens, we can see that the gesture control is sensitive enough for entertainment and education researches. The phd student showing the demo told me that she has developed an application that shows information of a patient by looking at his or her face through HoloLens. It is widely applied in medical field.
The disadvantage of HoloLens is also obvious: we have to look at the virtual space from some fixed, specific point. I believe that if the observing point is flexible, HoloLen can be much more popular. Also I think the interaction between human and machine is too weak as a VR access.
Most applicable: entertainment


I have been in CAVE2 for a lot of times and this impressive work always has different unique projects to show. The advantage is that in CAVE2 the 360 degree circumstance with surround sound is shocking real. The device is simple, just a pair of 3D glasses, and a motor control handle. Tracked by the sensors on each screen, the handle gives navigation in the virtual reality, which makes it easier to exhibit the models, datas, research cases, etc.
Disadvantage? Maybe too expensive for normal applications. Also a powerful computer is required to control all the screens.
Most applicable: science researches, art.


The most accessible application of all the demos. Easily used and developed.No headset needed. A lot of fun.
Disadvantage is that it is basically designed for children, which means it is limited in academic and research fields.
Most applicable: children's application


According to the Times, most users will be able to watch the films in 4K resolution.the app supports virtual reality playback for Google Cardboard as well as playback of 360-degree video for mobile touchscreen navigation. Google Cardboard is a cheaper access for VR. It can be viewed with or without device.
Disadvantage may be that it lacks interaction with users.
Most applicable: social media, tourism.

Week3: some thoughts on AR translation

Here goes some screens that google translate in different degrees of success.


From the screens we can see the fact that the completeness of translation might vary due to different distance between angles and distances to the characters. Font of the characters may be different. Google translation, the application, has the solution of the problem. It takes photos and let the users highlight the text. Yet things might change when it comes to a real time access.
Now suppose that we have an AR product, i.e. a pair of glasses or headset, with the function automatically translating a specific language to another language and replacing the original text from the real world, it sounds really covenient. So the benefit is that it is an easy access to AR translation. Think about the case that you are reading an article in another language, things get much easier when you put on your AR device. With pattern recognition, everything is translated into the desired language. The defect of the technique is also obvious. We can see the success of the translation may be in different degrees due to the distance and angle changes. Nobody wants to move the steps and nodding his or her head to adjust the text for translation. This makes the AR translation device narrowly applied Another point is that since it replace the original text, the font may be mixed and messed up and the meaning of the words might change.
In order to overcome the problems, the user should have the control of angle and distance to the text as well as the align of the text to get better effect on translation. The user need to adjust the distance and angles using some handheld devices. Also for more effecient access, the user needs to highlight the intended part of text using some devices.

Week4: Walking small

Here goes some screens that things in virtual circumstances in different scales.

Experience virtual reality in different scales is unique. Before I scale up, everything looks normal. Although it takes me some time to adapt the world which looks not exactly like the real world. After I scale up the world, it takes me a while to adapt the new world. I can see low resolution on some of the sideling edges. While the vertical edges does not change. Some textures, upon the light, changed after the sclare is changed. Colors may change a little bit. I also notice that there exits some a scale that makes the virtual circumstances looks most real, which is not 1:1.
The characteristic can be used in some situations where details need to explore. For example, a murder scene can be made as virtual reality. We can notice details when we scale up the circumstance, such as a hair or a snail. Another example of the technology, more close to science researches, is that we can explore something that we are not able to enter, such as a brain, a lung, an organ, or a group of gas, the solar system, etc. Overall, scaling in virtual reality can benefit the science a lot.


Here is the linkd to my youtube video:https://youtu.be/XvDInegeZPQ

I have started with the general model of EVL and build my desire office.The purpose of the project is to create a virtual reaistic world in practice so that I can have better understand how the process goes, how VR works and have deeper experience about VR. In my application, we can explore and collect things and throw them away, also we can turn the light on and off
I referred to a lot of materials to finish the project. Mostly, the source codes are from https://www.evl.uic.edu/aej/491/evl%20491.zip and unity3d.com. Some of my assets are found from asset store. https://www.assetstore.unity3d.com.
Here is my link: https://drive.google.com/drive/folders/0BwDejtfdn3XPUXIzVWpaQmhEbk0?usp=sharing
I CREATE 11 things by my own. Most of them are done by Blender.
a door:

a cup and a dish:

a lamp:

a brush pot and a pencil:

a mouse and a mouse pad:

a trash bin:

a paper ball:

a chair:

a ballon:

All of my 5 audio clips are from http://soundbible.com
Background music is Parakeet Talking. The sound that a lamp hits something is sword strike. When a book touches things, page_turn mark. There is also a click music when the light is turn on and off. Also the paper ball shoots around and sounds punch when collision.

Compared to navigation using teleporting, I prefer the wand-based navigation. We have to set a several teleporting points for navigation. That make things feel like locked. And at some angles, we cannot even see the things directly.While wanded-based navigation would be much easier. At least it makes me feel freedom. Compared to the actual room, except for the fake models I made, we can notice they are different. Not only the size but the textures make people feel unreal. Also Light, physics and sounds help a lot in the virtual circumstance.Things change a lot with these parameters. What can hardly be replicate in the virtual world is the texture. We can always tell the difference between kicking a fake trash bin in VR and kicking a real one.

Week6: classmates' projects

Peter Hanula's project

One of the projects I love most is this one. It contains a lot of fancy models including cats that interacts with users with cat sounds which make the office interesting. In my project, I have used not much interaction as his project. I like the light control switch with a feedback to users interaction. These can be achieved through an easy function called OnCollision(). I should have done more works about it. The fancy ceiling is also an attraction.

Timoth Choh's project

This one did great job combining a huge amount of models into an office without weird feelings. It is really simple to add those models. Also I like the sounds interaction attaching on the models. The sounds of the fan and television make it real in a VR circumstance. The light control is realized through a buttom. I like the trash bin and the feeling when stuff was put into the trash bin. All of these makes the circumstance real.

Week8: Pocket Universe experiences

Here goes some screens that Pocket Universe shows


The application as shown in the screenshots gives the names of the stars and ohter planet as well as the astrologies. There are several modes of viewing such as motion, time etc. It might be more interesting, at least it seems to me, to add some data source about the planets such as distances to us or the orbit directions. Another thing I would like to add in this view should be the time left to rise and set. Imagine if we are looking at the sky in the wild, if we have the approximate time as well as the location for something to show up, things might be more convenient and fun.

Week 11: Critical analysis for Holoportation

Microsoft's Holoportation has a huge progress in application in mobile unit. It is availabe to send video and audio from a moving vehcle. The major challenge of the project was bandwidth. So everything that is collaborated synchronously is minimized but still makes the quality of the transmission as well as possible. Temporal consistency is required for the project so that everything is collaborative. Also the remote rendering has a reprojection so that the quality of teleportation is good enough.
There are totally 2 cameras applied in the project. The more the camera is set, the higher the quality of teleportation would be. But as it comes to the long-distance transmission the quality seems less important. Since the bandwidth is reduced to 30-50 Mbps, there should be a lot of limitation in teleportation.
I think Holoportation accelerate the progress of mobile network to some extend. The 4G cellular data can support it successfully. I believe the 5G network would make Holoportation works and develops better with faster transmission speed and higher qulity.

Student choice topic

Dexta Touching sense gloves for VR

Since the fast development of Virtual reality, the digital world has changed a lot and been more and more virtual. Dexta Robotics has developed a glove for VR to simulate the touching sense so that the users can feel feedbacks from the virtual world. The VR gloves are designed to provide resistance to the wearer, corresponding to their interactions with the VR environment.
Although the immerssive technology that simulate vision and sounds are developing quickly,the key to immersion is still touch. The basic idea behind Dexta is to make the force feedback lightly, affordably and simulatedly.
Dexta has been able to manage the binary force feedback module on each single finger so that the feedback from the gloves are more simulated. They also developed a robotic hand act simutanuously simulating what exactly the glove do. And unlike other devices that use motors, the feed back of the glove comes from variable stifness.
The glove is a huge progress in improving the VR experience. Also it makes the product of VR expanded to some extend. I can imagine the application it used in long-distance surgeon, mining, training,etc.
The product looks so cool that it looks like it comes from future. If there is any issue on it, it is its business operation team They suspend the development according to the feedback from the VR market. And the project is closed temporarily. I believe with a better marketing strategy, the product would goes on successfully.
source sites:https://www.roadtovr.com/raising-55000-dexta-robotics-cancels-kickstarter-haptic-hand-exoskeleton/


Course projects goes below.

Project 1

Project 2

Project 3

About Me