Home

The Human Media Lab is seeking a highly motivated PhD candidate to explore Neuromorphic Human-Computer Interaction, a cutting-edge field that is set to revolutionise interactive experiences.

Ontwerpkeuzes in zowel software als hardware hebben grote gevolgen voor de uiteindelijke gebruikservaring en, in het verlengde, ook de impact op de samenleving. Dat leren Herbert Blankesteijn en Ben van der Burg in deze nieuwe aflevering van De Technoloog. Te gast is Roel Vertegaal, Hoogleraar Mens-computer Interactie aan de Radboud Universiteit in Nijmegen.

Roel Vertegaal has been appointed Professor of Human-Computer Interaction at Radboud University’s Faculty of Science effective May 1, 2024.

Vertegaal will be teaching and researching the topic of human-computer interaction. ‘How can we design computers such that people can use them more effectively? My specialization is the invention of novel interaction techniques. For example, I invented the attention-aware interface in the iPhone as well as the foldable smartphone, which makes using phones a bit more like using paper. My chair will focus on the development of technologies that combat societal problems caused by the use of computers. A good example is smartphone addiction: attention-aware devices are a solution.’ To this end, the professor is currently working on a new theory for designing and testing user interfaces based on predictability of functions by the human brain, with applications in AI. 

(Note that the flicker in the video is due to the interaction of the camera shutter with the 45 projector shutters, and is not visible to users)

Projected self-levitating light field display allows users to hover in remote places

GLASGOW – This week, Human Media Lab researchers will unveil LightBee - the world’s first holographic drone display. LightBee allows people to project themselves into another location as a flying hologram using a drone that mimics their head movements in 3D - as if they were in the room.

“Virtual Reality allows avatars to appear elsewhere in 3D, but they are not physical and cannot move through the physical space” says Roel Vertegaal, Director of the Human Media Lab. “Teleconferencing robots alleviate this issue, but cannot always traverse obstacles. With LightBee, we're bringing actual holograms to physical robots that are not bound by gravity.”

Queen’s University’s Human Media Lab unveils world’s first interactive physical 3D graphics

Berlin – October 15th. The Human Media Lab at Queen’s University in Canada will be unveiling GridDrones, a new kind of computer graphics system that allows users to physically sculpt in 3D with physical pixels. Unlike in Minecraft, every 3D pixel is real, and can be touched by the user. A successor to BitDrones and the Flying LEGO exhibit – which allowed children to control a 3D swarm of drone pixels using a remote control – GridDrones allows users to physically touch each drone, dragging them out of a two-dimensional grid to create physical 3D models. 3D pixels can be given a spatial relationship using a smartphone app. This tells them how far they need to move when a neighbouring drone is pulled upwards or downwards by the user. As the user pulls and pushes pixels up and down, they can sculpt terrains, architectural models, and even physical animations. The result is one of the first functional forms of programmable matter. 

Queen’s Human Media Lab unveils the world's first rollable touch-screen tablet, inspired by ancient scrolls

BARCELONA - A Queen’s University research team has taken a page from history, rolled it up and created the MagicScroll – a rollable touch-screen tablet designed to capture the seamless flexible screen real estate of ancient scrolls in a modern-day device. Led by bendable-screen pioneer Dr. Roel Vertegaal, this new technology is set to push the boundaries of flexible device technology into brand new territory.

The device is comprised of a high-resolution, 7.5” 2K resolution flexible display that can be rolled or unrolled around a central, 3D-printed cylindrical body containing the device’s computerized inner-workings. Two rotary wheels at either end of the cylinder allow the user to scroll through information on the touch screen. When a user narrows in on an interesting piece of content that they would like to examine more deeply, the display can be unrolled and function as a tablet display. Its light weight and cylindrical body makes it much easier to hold with one hand than an iPad. When rolled up, it fits your pocket and can be used as a phone, dictation device or pointing device.

New light field displays effectively simulate teleportation

MONTREAL – This week, a Queen's University researcher will unveil TeleHuman 2 - the world’s first truly holographic videoconferencing system. TeleHuman 2 allows people in different locations to appear before one another life size and in 3D - as if they were in the same room.

“People often think of holograms as the posthumous Tupac Shakur performance at Coachella 2012,” says Roel Vertegaal, Professor of Human-Computer Interaction at the Queen’s University School of Computing. “Tupac's image, however, was not a hologram but a Pepper Ghost: A two-dimensional video projected on a flat piece of glass. With TeleHuman 2, we're bringing actual holograms to life.”

Using a ring of intelligent projectors mounted above and around a retro-reflective, human-size cylindrical pod, Dr. Vertegaal’s team has been able to project humans and objects as light fields. Objects appear in 3D as if inside the pod, and can be walked around and viewed from all sides simultaneously by multiple users - much like Star Trek's famed, fictional ‘Holodeck’. Capturing the remote 3D image with an array of depth cameras, TeleHuman 2 “teleports” live, 3D images of a human from one place to another - a feat that is set to revolutionize human telepresence. Because the display projects a light field with many images – one for every degree of angle – users need not wear a headset or 3D glasses to experience each other in augmented reality.

Experimenting with the Future of Play at LEGO® World Expo

Visitors will be able to experience and play with new technology combining LEGO® bricks and drones.

COPENHAGEN - February 15-18, children and families visiting the LEGO® World expo in Copenhagen, Denmark will have the chance to make their brick-building dreams take flight with a flock of interactive miniature drones developed by the Human Media Lab at Queen’s University in Canada in collaboration with the LEGO Group’s Creative Play Lab.

The system allows children to arrange LEGO elements into a shape of their choice and watch as a group of miniature drones takes flight to mimic the shape and colour of their creation in mid-air. With the aid of tiny sensors and gyroscopes, the system also tracks when the children move, twist and bend their designs. The drones faithfully replicate any shape alterations as an in-air animation.

Queen’s University’s Human Media Lab to unveil musical instrument for a flexible smartphone

KINGSTON - Researchers at the Human Media Lab at Queen’s University have developed the world’s first musical instrument for a flexible smartphone. The device, dubbed WhammyPhone, allows users to bend the display in order to create sound effects on a virtual instrument, such as a guitar or violin.

“WhammyPhone is a completely new way of interacting with sound using a smartphone. It allows for the kind of expressive input normally only seen in traditional musical instruments.” says Dr. Vertegaal.

Queen’s University’s Human Media Lab to unveil world’s first flexible lightfield-enabled smartphone.

KINGSTON - Researchers at the Human Media Lab at Queen’s University have developed the world’s first holographic flexible smartphone. The device, dubbed HoloFlex, is capable of rendering 3D images with motion parallax and stereoscopy to multiple simultaneous users without head tracking or glasses.

“HoloFlex offers a completely new way of interacting with your smartphone. It allows for glasses-free interactions with 3D video and images in a way that does not encumber the user.” says Dr. Vertegaal.

HoloFlex features a 1920x1080 full high-definition Flexible Organic Light Emitting Diode (FOLED) touchscreen display. Images are rendered into 12-pixel wide circular blocks rendering the full view of the 3D object from a particular viewpoint. These pixel blocks project through a 3D printed flexible microlens array consisting of over 16,000 fisheye lenses. The resulting 160 x 104 resolution image allows users to inspect a 3D object from any angle simply by rotating the phone.

Queen’s University’s Human Media Lab to unveil world’s first handheld with a fully cylindrical display at CHI 2016 conference in San Jose, CA.

Researchers at Queen’s University’s Human Media Lab have developed the world’s first handheld device with a fully cylindrical user interface. The device, dubbed MagicWand, has a wide range of possible applications, including use as a game controller.

Similar to the Nintendo Wii remote, but with a 340 degree cylindrical display, users are able to use physical gestures to interact with virtual 3D objects displayed on the wand. The device uses visual perspective correction to create the illusion of motion parallax; by rotating the wand users can look around the 3D object.

Queen’s University’s Human Media Lab to unveil world’s first wireless flexible smartphone; simulates feeling of navigating pages via haptic bend input

KINGSTON - Researchers at Queen’s University’s Human Media Lab have developed the world’s first full-colour, high-resolution and wireless flexible smartphone to combine multitouch with bend input. The phone, which they have named ReFlex, allows users to experience physical tactile feedback when interacting with their apps through bend gestures.

“This represents a completely new way of physical interaction with flexible smartphones” says Roel Vertegaal (School of Computing), director of the Human Media Lab at Queen’s University.

“When this smartphone is bent down on the right, pages flip through the fingers from right to left, just like they would in a book. More extreme bends speed up the page flips. Users can feel the sensation of the page moving through their fingertips via a detailed vibration of the phone. This allows eyes-free navigation, making it easier for users to keep track of where they are in a document.”

Queen’s University’s Roel Vertegaal says self-levitating displays are a breakthrough in programmable matter, allowing physical interactions with mid-air virtual objects

KINGSTON, ON – An interactive swarm of flying 3D pixels (voxels) developed at Queen’s University’s Human Media Lab is set to revolutionize the way people interact with virtual reality. The system, called BitDrones, allows users to explore virtual 3D information by interacting with physical self-levitating building blocks.

Queen’s professor Roel Vertegaal and his students are unveiling the BitDrones system on Monday, Nov. 9 at the ACM Symposium on User Interface Software and Technology in Charlotte, North Carolina. BitDrones is the first step towards creating interactive self-levitating programmable matter – materials capable of changing their 3D shape in a programmable fashion – using swarms of nano quadcopters. The work highlights many possible applications for the new technology, including real-reality 3D modeling, gaming, molecular modeling, medical imaging, robotics and online information visualization.

Queen’s University’s Human Media Lab present 3D Printed Touch and Pressure Sensors at Interact'15 Conference

Queen's professor Roel Vertegaal and students Jesse Burstyn, Nicholas Fellion, and Paul Strohmeier, introduced PrintPut, a new method for integrating simple touch and pressure sensors directly into 3D printed objects. The project was unveiled at the INTERACT 2015 conference in Bamberg, Germany: one of the largest conferences in the field of of human-computer interaction. PrintPut is a method for 3D printing that embeds interactivity directly into printed objects. When developing new artifacts, designers often create prototypes to guide their design process about how an object should look, feel, and behave. PrintPut uses conductive filament to offer an assortment of sensors that an industrial designer can easily incorporate into these 3D designs, including buttons, pressure sensors, sliders, touchpads, and flex sensors.

Queen’s University’s Human Media Lab and the Microsoft Applied Sciences Group unveil DisplayCover at MobileHCI'15

Queen’s professor Roel Vertegaal and student Antonio Gomes, in collaboration with the Applied Sciences Group at Microsoft, unveiled DisplayCover, a novel tablet cover that integrates a physical keyboard as well as a touch and stylus sensitive thin-film e-ink display. The technology was released at the ACM MobileHCI 2015 conference in Copenhagen - widely regarded as a leading conference on human-computer Interaction with mobile devices and services.

DisplayCover explores the ability to dynamically alter the peripheral display content based on usage context, while extending the user experience and interaction model to the horizontal plane, where hands naturally rest.