This is a big one. It’s a packet analyzing network sniffer that outputs crazy rhythms, mapped to crazy sounds. We (me and @0xtosh) set it up at HAR2009 were people could log in to a linux VM, run some scripts, ours or their own, to scan the network and netglitch broadcasted a shoutcast stream to which you could tune into to listen to your own packets.
I hardly have time lately to code anything at home. I just dusted of the APC and I made slight progress going almost no sound to some sound…
That is in “dongle” mode. In stand alone mode it works fine. I even added a set of capacitors so you can switch between several sorts of sounds and a light sensitive resistor for that look-without-touching-I-can-play-this-APC-touch
I was successful in attaching it to the computer with the regular standard arduino’s + firmata and playing it from puredata.But this is an arduino pro mini (5v 16Mhz, from sparkfun). Apparently firmata isn’t entirely the same on these mini’s. Although Since arduino 0017 I can connect and send data, it’s not without errors. First of all
Is what I see when I try to run the pd patch. lot’s of those, thousands…
Another weird thing is that even though pin 13 is unconnected on the arduino, It’s the only pin that seems to work. I really don’t get it. Is there some sort of different mapping? normally only pin 9-10-11 are supposed to work.
physorg has some (old) news about a 3D LCD screen. They claim it’s 3D so it has to use lecticular system to deliver seperate image to both eyes. On the other hand they say it uses two LCD screens which are then projected. This would be good as this way you would not notice any interlacing. I have a hard time imagining it from this rather vague description and KDDI doens’t seem to have link on their site. KDDI is serious about R&D though…
Various techniques and software exists to bring more interaction and integration with reality. As this is a fairly new domain most of the projects are still in research phase. Some are already easy to implement, like motion tracking based on colorful objects or camera motion based on movements (accelerometers and other sensors). Blender is a good experimentation environment because it’s open and you can interface it with virtually everything using python.
some things already look stable and usable, like levelhead. This application is more then a proof of concept. It proves interesting new concepts are possible using these techniques. The tools are getting easier to use and the CPUs and GPUs are finally up to realtime processing and rendering.
current usable tools include:
osgart: a library that simplifies the development of Augmented Reality or Mixed Reality applications by combining the well-known ARToolKit tracking library with OpenSceneGraph.
ARToolKit: a software library for building Augmented Reality (AR) applications. Open source, with professional support from artoolworks inc. if needed.
Studierstube Tracker: a computer vision library for detection and pose estimation of 2D fiducial markers. It looks great, but it’s not open source.
If you know of good tools in the AR-area, please leave a note. Once the workflow is on track the developing of applications will be easier. Combining gestural data with 3d modeling tools so real time animation of 3D-models using the body of an actor as input is one that comes to mind. Another one could be to overlay your newspaper with interactive data, like graphs or models of the new product being reviewed. Already mobile phone apps exist that let you view additionalinformation about Landmark X when you point them at Landmark X. But also advertising comes to mind. Or using it as a try before you buy tool. Imagine you could put that new sofa on the spot where you always wanted it, without having to move it, assemble it and then notice the colors don’t match… just don’t try to sit on it.