NAME OF THE PARTICIPANT 1: VASA YESHWANTH BRANCH:ECE NAME OF THE INSTITUTION:TEEGALA KRISHNA REDDY ENGINEEREING COLLEGE CONTACT NUMBER:9985776771 EMAIL ID:yeshwanthvasa@gmail.com
NAME OF THE PARTICIPANT 2:KEDAR PATHAK BRANCH:ECE NAME OF THE INSTITUTION: TEEGALA KRISHNA REDDY ENGINEEREING COLLEGE CONTACT NUMBER:9494243056 EMAIL ID:kedarrocks.tkr@gmail.com
Introduction:
'Sixth Sense' is a wearable gestural interface that augments the physical world around us with digital information and lets us use natural hand gestures to interact with that information.
Although the miniaturization of computing devices allows us to carry computers in our pockets, keeping us continually connected to the digital world, there is no link between our digital devices and our interactions with the physical world. Information is confined traditionally on paper or digitally on a screen. SixthSense bridges this gap, bringing intangible, digital information out into the tangible world, and allowing us to interact with this information via natural hand gestures. Imagine being able to take a digital photograph by making the shape of a picture frame in the air with your hands, or dialing a phone number by simply tapping your palm with a finger or even navigating your way around a city by projecting a map onto the nearest wall, then using it as a touch screen to find a convenient subway station or bus stop. This is the promise of Sixth-Sense, a prototype wearable device invented by MIT engineer Pranav Mistry. Mysterys device consists of a video camera and LED projector combined into a pendant worn around the neck. This is linked to a smart phone worn on the hip, which houses the devices software and links it wirelessly to the internet. The camera not only takes pictures, but also records the users hand gestures. The software processes these gestures and interprets them into actions.
The increasing need for intelligent video surveillance in public, commercial and family applications makes automated video surveillance systems one of the main current application domains in computer vision. Intelligent video surveillance systems deal with the real-time monitoring of persistent and transient objects within a specific environment. Intelligent surveillance system has been evolution to third generation, known as automated wide-area video surveillance system. Combined computer vision technology, the distributed system is
3
autonomic, which can also be controlled by remote terminals. A low-cost intelligent wireless security and monitoring solution using moving object recognition technology is presented in this paper. The system has good mobility, which makes it a useful supplement of traditional monitoring system. It can also perform independent surveillance mission and can be extended to a distributed surveillance system. Limited by the memory consuming and computing capacity in a mobile phone, background subtraction algorithm is presented to be adopted in mobile phones. In order to be adapted to the slow and slight changes of the natural environment, a self-adaptive background model updated automatically and timely is detailed. When the subtraction of the current captured image and the background reaches a certain threshold, a moving object is thought to be in the current view, and the mobile phone will automatically notify the central control unit or the user through phone call, SMS, or other means. Thesproposedsalgorithmscansbes implemented in an embedded system with little memory consumption and storage space, so its feasible for mobile phones and other embedded platforms, and the proposed solution can be used in constructing mobile security monitoring system with low-cost hardware and equipments.
principle of moving object detecting is given. Limited by the memory consuming and computing capacity of a mobile phone, a background subtraction algorithm is presented for a adaptation. Then, a self-adaptive background Central control unit or the user through phone call, SMS (Short Message System) or other means. Other hardware includes mirror and colored caps to be used for fingers. The software of the technology uses the video stream, which is captured by the camera, and also tracks the location of the tips of the fingers to recognize the gestures. This process is done using some techniques of computer vision.
Module:
The sixth sense prototype is comprised of a pocket projector, a mirror and a camera. the hard ware components are coupled in a pendant like mobile wearable device. both the projector and the camera are connected to the mobile computing in the users pocket. the projector projects visual information enabling the surfaces, walls and physical objects around us to be used as interfaces; while the camera recognizes and tracks users hand gestures and physical objects using computer vision based techniques. the software program processes the video stream data captured by the camera end tracks the location of the colored marker (visual tracking fiducially) at the tip of the users fingers using simple computer vision techniques .the movements and arrangements of these fiducially are interpreted into gestures that acts as an interaction instructions for the projected applications interfaces. The maximum number tracked fingers is only constrained by the number of unique fiducials,thus sixth sense -user interaction
Interfacing:
You can carry the digital world with you wherever you carry it. you can start using any surface or any wall around you as an interface. This module as already told consists of three major parts .the projector, the camera, the mirror. If you lack a surface around you, then its easy as making your palm the Interface with your module. This module has the ability to set its interface making it convenient to the user handling it.
In our prototype, the first 10 frames are captured to calculate the background template (i=10). Moving objects can not be identified in these frames. If the moving objects move too slowly, I should be increased to reduce the tolerance.
In case of some random disturbances, each pixel will fluctuate in a small range even there is no expected moving objects in the scene. So there must be a strategy to judge it. A threshold is defined in the system. If the difference of one pixel between real time frame and template is
more than 10, then add 1 to the threshold. When differences of all pixels in the frame are all calculated, moving objects is thought to appear if the threshold is more then 3 percent of the total number of pixels in the frame.
J2ME Technology:
In this article, we have implemented a prototype on mobile telephones based on J2ME technology. Java Platform, Micro Edition (Java ME) is the most ubiquitous application platform for mobile devices across the globe. It provides a robust, flexible environment for applications running on a broad range of other embedded devices, such as mobile phones, PDAs, TV set-top boxes, and printers. Applications based on Java ME software are portable across a wide range of devices, yet leveraging each device's native capabilities.
Capturing an Image
Once the camera video is shown on the device, capturing an image is easy. All you need to do is call VideoControl's getSnapshot() method. The getSnapshot() method returns an array of bytes,
which is the image data in the format you requested. The default image format is PNG (Portable Network Graphic). byte[]raw=mVideoControl.getSnapshot(null); Image image = Image.createImage(raw, 0, raw.length);
Prototype
The system architecture is shown in Figure 2. In the prototype system, if the difference between real-time frame and template reaches a predefined threshold, moving objects are considered to
appear. Then the handset will send out an alert SMS. Since the device has good mobility, it can be put anywhere including those area not covered by other surveillance system. And it can be deployed rapidly in emergency.
Implementation:
The prototype has been implemented on Nokia N series of latest version. The first picture in the form is real time frame, which is got from the camera originally. The second image is the template image. If there are some moving objects being detected, the third picture will be displayed on the form. And some real time information is displayed below the pictures.
10
Since the first several frames must be stored to calculate the template, a big memory heap size is needed. The JVM heap cant provide so much memory. There are two methods to solve this problem. First, the frames can be stored to the EFS (Embedded File System). Second, the image size can be reduced. The first method can provide high resolution image data, which contains more detailed image features. But storing to EFS will take a much longer time than storing to memory, and much longer time is need in the following calculating processes. The second method will lost some detailed image features, but it can fully operated in memory and reduce much processing time. Considered the requirements of the real time ability, the second method is adopted. The image size is reduced to 160*120, which is still enough to be used to identify the moving objects. The performance for 100 frames is detailed in Table 2. The term Snapshot time is the time length to get the image though J2ME MMA API. It mainly depends on the capability of Hardware, Operating System and Java VM. The DIP (digital image process) time is the time length to perform the background subtraction algorithm, including images compare and template update. The Frame time is the total time to process a frame which equals to the sum of Snapshot time and DIP time. The Frame time Average is an averageof several frame time. The Template time is the total time to construct the background template. In this instance, the Snapshot time, DIP time and Frame time of the 100th frame are presented. And the Frame time Average of the 100 frames and Template time are also given. The background template can be built in less than half a minute. And the total time to calculate a frame is around one and a half seconds. It meets the requirements to be a family security monitoring system and an anti-theft system. The experiment demonstrates the feasibility of the proposed system. Some frames are magnified to be seen more clearly (See Figures 4 - 6). Figure 4 is the self-adaptive background template. As shown in Figure 5 and 6, a person run into the scene was identified immediately.
11
Applications:
The device has a huge number of Applications firstly, it is portable and easy to carry as you wear it in your neck. It is the combined technology of computer along with cell phone. It works when a person hang it on his neck and start projecting through the micro-projector attached to it. Your fingers works like the keyboard as well as the mouse.
Easy to Draw: A person can draw anything on paint with the help of his fingers. So, why to
use mouse for painting. The drawing application lets the user draw on any surface by observing the movement of index finger.
12
Capture photos with fingers: why to take camera on your holiday and no tension for the
photo space as this Sixth Sense computer will work like your camera. It caputes the photo, when you make a square with your fingers, highlighting which one you want to frame.
Phone Call: You can call to your friend by typing the numbers on your hand. It display the
keypad of the phone over your palm and the keys appears on the four finger... Use your another hand's finger to press the keys.
13
Check the brand of the Product: It helps you to choose the best brand product from the
super market.
Read Books easily: Check out the ratings of the Book you are going to buy, it checks the
ratings from the internet. And another amazing thing is that it reads the book for you
Newspapers: Did you saw the moving pictures of the Newspaper in the movie Harry Patter, it
is quite similar to it. It Searches the most appropriate video from the web by seeing the headlines or the caption of the News report.
Check your Flight Status: You can check the status of the flight while you are on Taxi.
Just place the ticket in front of the projector and it checks its status from the internet.
14
Google Maps
The sixth sense also implements map which lets the user display the map on any physical surface and find his destination by just using his gestures.
This technology can be used as a fifth sense rather than a sixth sense for the dumb. Using the gestures of the people this instrument can be used to generate voice and make the dumb speak. This would be a greatest achievement for the mankind.
16