Oculus Rift Experiment – Is Virtual Reality Ready for Business Applications?

Introduction to Oculus Rift

The Oculus Rift is a new Virtual Reality (VR) headset designed to provide a truly immersive experience, allowing you to step inside your favorite video game, movie, and more. The Oculus Rift has a wide field of view, high-resolution display, and ultra-low latency head tracking unlike any VR headset before it.

Nerdery Lab Program

Lab partners Chris Figueroa and Scott Bromander collaborated on this Oculus Rift experiment; their respective Lab Reports are below. The Nerdery Lab program is an opportunity for employees to submit ideas for passion projects demonstrating cutting-edge technologies.  Nerds whose ideas show the most potential are given a week to experiment and produce something to show to other Nerds and the world at large.

Continue reading Oculus Rift Experiment – Is Virtual Reality Ready for Business Applications?

Apps That Know Where You Are: Our Experimentation With Apple’s iBeacon Technology

Introduction to the Lab Program:

Earlier this year, The Nerdery unveiled its Nerdery Labs program. It’s an opportunity for employees to submit ideas for projects demonstrating cutting-edge technologies. Those Nerd’s ideas which show the most potential are given a week of time to pursue it and produce something to show to other Nerds and the world at large.

I have a strong personal interest in extending user experiences beyond the bounds of traditional mobile apps by interfacing with external technologies. I saw the Nerdery Labs program as the perfect opportunity to pursue that interest…so I submitted a proposal to show the possibilities of Apple’s new iBeacon technology. I was tremendously excited when I heard my idea had been selected and as soon as I wrapped up the client project I was engaged with, I got to work!

Introduction to iBeacon

Buried in the ballyhoo surrounding the radical visual changes to iOS 7 was an all-new technology introduced by Apple: iBeacon. What it lacks in razzle-dazzle, it more than makes up for in enabling entirely new interactions and types of applications!

It is important to understand that iBeacon is not a device or a new piece of hardware like the TouchID thumbprint scanner. Instead, it is a public protocol or “profile” built on top of the Bluetooth LE (Low Energy) technology which has been present for several years in iOS devices: iPhone 4S and later, iPad 3rd Gen and later, and the 5th Gen iPod Touch. Bluetooth LE was released in 2010 as a lower-power, lower-speed alternative to traditional Bluetooth; devices broadcasting infrequently using Bluetooth LE can run for up to two years on a single coin-cell battery. Any device that announces itself using the iBeacon profile is an iBeacon, whether it is a small, dedicated radio device or an iDevice configured to broadcast as an iBeacon. Apple will not be producing any dedicated iBeacon hardware – that will be left to third parties. Android support for Bluetooth LE was added in 4.3 (Jelly Bean) so there will likely be Android iBeacons in the near future, too.

How iBeacon works

Figure 1-1. How iBeacon works

At its core, iBeacon is simply a “HERE I AM!” message broadcast roughly once per second to other devices within range of the Bluetooth radio (Figure 1-1). It has a few identifying characteristics so that apps can distinguish the iBeacons they’re interested in from a crowd. iBeacon broadcasts have no data payload; they simply identify themselves via a UUID (unique identifier) and 2 numbers, dubbed “major” and “minor”. You can think of the UUID as the application identifier: each app will use a different one (or more). An app can only listen for specific UUIDs provided by the developer, there is no way to see a list of all iBeacons visible to the device. The major and minor numbers have no intrinsic meaning, they are available for the app to use as the developer sees fit. A common scheme is to designate the major number as the general region and the minor as a specific location within that region. As an example, in an app for Macy’s, the UUID for all iBeacons in all Macy’s stores would be identical. The major number would refer to a particular Macy’s store (22 = San Francisco, 1 = NYC, etc.). The minor number would represent the different departments within the Macy’s store (14 = Women’s Apparel, 7 = Bedding, 29 = Men’s Shoes, etc). The numbers represent whatever what you decide as you plan out the app. The point is, major and minor could be used to identify more than just physical locations; people, pets, containers, kiosks, luggage, and many other objects that you want to keep track of as they are on-the-go could benefit from the technology.

Continue reading Apps That Know Where You Are: Our Experimentation With Apple’s iBeacon Technology

I Can See The Music – The TasteMapper Lab Experiment


My name is Kevin Moot and I am a Senior Software Developer at The Nerdery. I was fortunate to be involved in one of the Nerdery’s first projects of the Nerdery Labs program.

The Nerdery Labs program is in the same vein as Google’s famous “20% time” (in which Google developers are offered an opportunity to invest 20% of their time at work fostering their own personal side-projects). Our Nerdery Labs program presents a great opportunity for developers to bring their own personal projects to life and play with new technologies, creating experiments ranging from software “toys” to projects with potential business applications.

The Vision

Our concept was ambitious: create a diagram of the entire musical universe.

I teamed up with one of our Nerds in our User Experience department, Andrew Golaszewski, to conceptualize how we could visualize a musical universe in which users could explore an interconnected set of musical artists and genres, jumping from node to node in much the same way that your curiosity might take you from a Wikipedia article on “Socioeconomics of the Reformation Era” and somehow end up on a video clip of “The Funniest Baby Sloth Video Ever!

Assuming there was some huge data set out there which would give us insights into the composition of people’s personal musical libraries and playlists, could we find out which genres and artists are most commonly associated with one another? Are fans of Ozzie Osbourne likely to see Simon and Garfunkel popping up in their listening history? What other artists should a loyal follower of Arcade Fire be listening to? Do certain genres display more listener “stickiness” that others – that is, do fans of pop music statistically branch out more often to other genres than devotees of death metal?

To narrow this ambitious plan down to a more reasonable, bite-sized problem set, we decided to concentrate on depicting a single central artist at a time as a “root” node. Connected to this root note would be a set of the most closely related artists.

Thus was born the TasteMapper experiment.

Continue reading I Can See The Music – The TasteMapper Lab Experiment