4 October 2017 – I updated the code for Swift 4 and iOS 11. You can find it here.
When an iPhone is processing an image from one of its cameras, the main purpose is to make it look good for us. The iPhone has no interest in interpreting the image itself. It just wants to adjust the brightness and the colours, so that we can optimally enjoy the image.
The three guys in the painting below are all depictions of King Charles I, painted from different viewpoints by Anthonis van Dyck in 1636. In the middle we see a frontal view (“en face”), on the left a side view (“en profile”), and the most intriguing is the one on the right: the three-quarter view (“en trois quarts”). So here is the question that has bugged me for some time. Why is it called three-quarter, or 3/4? Three quarters of what?Continue reading “3/4 view”
Well. Maybe we first have to talk about what it means to “see”. Vision is an extremely rich natural phenomenon. Most of us humans have the uncanny ability to turn light into meaning – as do many other species in the animal kingdom. Vision is mainly used for navigation and recognition. We use our eyes to detect objects in our environment and use the shapes and layout of these objects to navigate our way through life.Continue reading “What can your iPhone see?”
This week I seriously started learning Swift. Swift is a novel programming language developed by Apple to replace Objective-C. I already like it and I definitely enjoy the learning process. Some iOS developers I know are talking about how they love Swift. I am not there yet, although I have found three things that may ignite my love.
In the industrial production process from idea to use, the consumer came last – almost as an after-thought. That is where the term “end-user” comes from. In the present era of smartphones and tablets and apps, an approach that focuses on the user, first and foremost, is uncontroversial.Continue reading “Distractions from user focus”
Although sometimes credited to the Renaissance artists and engineers, the camera obscura, or pinhole camera was already used by the Chinese in the 4th century BC and the Arabs in the 10th century AD. If you have never seen one in action, you are missing out. The images have a vibrant dreamlike quality, especially when objects in the scene are moving.
I don’t drive. I tried to get my driver’s license several times, but I failed. The main reason is quite ironic. My control of the car was up to standards, but all the examiners stated that I lacked the perceptual abilities to safely navigate traffic. At that moment, I had spend almost half of my life studying visual perception. That knowledge apparently does not transfer to my visual skills at all.
Being skilled in the art of drawing a convincing scene in linear perspective is no guarantee anymore for a successful career. For roughly four centuries this was a pretty good tool to have in your kit as a visual artist – from the moment that Filippo Brunelleschi gave his demonstration of a perspective rendering of the Baptistery in Florence in 1425, right up until Joseph Nicéphore Niépce took the first photograph of a view from a window in Saint-Loup-de-Varennes in 1826.
Winston Smith is the main character in the famous novel Nineteen Eighty-Four by George Orwell. In his apartment there is an “oblong metal plaque like a dulled mirror which formed part of the surface of the right-hand wall.” The description of this flatscreen television goes on to state that the “instrument (the telescreen, it was called) could be dimmed, but there was no way of shutting it off completely. The telescreen received and transmitted simultaneously.” This system provided a method for controlling citizens in the totalitarian society that Smith had to endure.
Yesterday, Microsoft launched a new product called Kinect. It is an add-on for the very popular game console Xbox 360 and allows for the user itself to be the controller. No more fiddling with weirdly shaped controllers. Just step in front of your television and you can control games with your own gestures (and your own words).