With the 0.3 release of the structure sensor SDK from Occipital it brings 3D colour scanning to the iPad addon sensor. I scanned a womble toy for which the above image is the model that was sent by email and loaded into meshlab.
In the pursuit of new technologies I backed the Structure.io sensor from Occipital Inc on Kickstarter. Being one of the first 200 hundred devices I got to grips with the device and think that it is a great device and really can allow your applications to reach out into the real world.
This is a list of iOS image filtering, processing, OCR, utilities and helper libraries and frameworks that I have come across for iPhone and iPad development and thought that they may be useful in one places:
1. GPU Image
By Brad Larson, GPUImage is a fantastic fast processing library to apply gpu accelerated filters to both live camera feed, movies or stills. It works on iOS 4 and above. It includes a range of filters such as Sepia, Grayscale, Blur but you can write your own easily.
2. Core Image
This is an apple framework available from iOS 5. It leverages the GPU and you don’t need to know details of OpenGL, OpenGL ES.
Documentation: Apple Core Image Docs
I came across an article demonstrating the performance of the feature extraction and descriptor, which is useful when building opencv on iOS as it shows how the performance of your application will change across all of the devices. It compares ORB, BRISK, FREAK (using GoodFeaturesToTrack), FAST, BRIEF (using GoodFeaturesToTrack), SIFT and SURF. Continue reading “Feature extractor and descriptor OpenCV performance on iOS iPad and iPhones” »
I am giving a presentation at SAP Dcode (originally teched) 2014 on the following topic:
How can augmented reality and virtual reality make a difference? What technologies are available today and coming? Find out from three-time DemoJam winners who pioneered smart glasses AR connected to SAP HANA. This extends going beyond 2D interfaces using 3D, augmented reality, virtual reality, and more connected to SAP providing next-generation applications.
I look forward to seeing you there. Check out the session here.
Augmented and virtual reality are technologies of our time and seem to be expanding at an ever increasing pace. I wanted to look back at where the origins of these technologies have come from and the initial inventions and developments that spawned these new platforms.
For the origins of AR and VR Morton Helig is often credited. through his construction of the Sensorama machine starting in 1957. This was designed as a cinematic experience which tried to immerse all your senses. It delivered different images to each eye, played sound at you and even vibrated your chair and blew air at you. So this was a really immersive solution and functionality it has more sensory inputs than we have with todays augmented and virtual reality solutions. The first film created was a 10 minute motorbike ride through New York. Unfortunately this machine was commercial unviable partly because it is the size of an arcade machine and partly because it cost to much to make the films for. But this one machine shows some of the early ideas that have become Virtual and Augmented reality.
I am a fan of really great UI and functionality and wanted to add features such as autocomplete and showing me my purchase history right on the product page of when I bought it and for how much. To do this I have created a chrome extension called ‘Food Basket Accelerator‘ that provides these features. Here is some more details:
Autocomplete for Tesco.com – The extension pulls the products that are important to you and auto suggests them based on what you have typed and their priorities. The importance of the products is calculated based on a number of metrics so it will get more accurate over time.
Product purchase history for Tesco Groceries – When you have previously added a product to your basket you will then see it within the tesco.com website product details page. It shows you how long ago you purchased it, how much and what promotions were running. In short it gives you a better experience and more informed decision into making a purchase.
All of the data that is compiled by this app remains on your own machine. It is not recorded on a server anywhere and it’s primary and sole purpose is to help you with your Tesco Online shopping.
Get it now from the chrome webstore here
HANADeck is a prototype application we developed that harnesses the Oculus Rift, Leap Motion, Microsoft Kinect and Unity3d for an immersive 3D virtual world which connected with SAP HANA created an enterprise application for data and environment visualisation.
So why use gaming technology for enterprise visualisation? We work in 3D, it is much easier to understand information when it is shown in a way that resembles the real world. Especially with the drive for gesture control and really immersive experiences it is more important than ever to look for truly immersive experiences.
HANADeck uses the display of the 7 inch Oculus Rift with Stereoscopic vision, which delivers the 3D environment. The user is tracked by a microsoft kinect, as this provides the lateral and vertical movement to the virtual environment. It was also found to reduce the motion sickness some people experienced with the Rift as the system is providing motion feedback to the user. Also a Leap Motion controller was mounted on the front of the Rift at 45 degress, which works well at tracking the user’s hands and providing an interaction point. Between the Leap Motion, Microsoft Kinect and Oculus Rift there are all the parts to track and deliver the immersive experience to the user.
The experience focused on in this enterprise demo is taking the visualisations of power lines within the UK network. This is sitting on real world data and purpose is to allow the engineer to travel to electricity towers a lot more quickly and safely.
Our second demojam entry involve making every surface around you an interactive human interface to you business activities.
Smart Fridge / Smart Refrigerator – a transparent multi touch surface on the front of the fridge with controlled internal RGB lighting and lock. You can fill out everything from your expense report to your leave requests on it, whilst getting you one of the sunshine drinks.
Smart Window – Controlled with leap motion, Microsoft kinect you can do similar on any window as you look out.
At Keytree labs we have been working hard at for SAP Demojam 2013. We are putting forward two ideas and our first one.
HANADeck is virtual reality experience powered by the Oculus Rift, Leap Motion and Microsoft Kinect. This means that a user can walk around and manipulate with their hands a virtual reality space. For this use case we are using the National Grid pylon database. A service engineer can inspect the data associated with any tower on the network remotely.