27 Oct Update

just a heads up im stuck

Right now, the goal is to have a camera to detect if the orange ball is still on the plane- if yes, timer runs/ if not, timer stops. 

I have tried a few methods to achieve this but every one of it doesn’t work.

  • Processing + Macbook Facetime HD Camera
    Used the ‘Video’ library of Processing
    Problem as per screenshot:
    Unfortunately, I am on a macOS Catalina ver 10.15.5https://github.com/processing/processing-video/issues/134
    Tried the solution as suggested by ihaveaccount on this above forum but the string doesn’t return as they said it wouldwithout the video library, this eliminates the use of external webcams also (i tried with my DSLR as a webcam)
  • Processing + IP Camera
    Since I was hoping for the webcam to be portable anyway, I tried to use an IP Camera. 
    Problem: On my processing sketch, nothing shows up, no error, no video feed show- nothing. IP camera doesn’t show it is being connected to anything either.

    Not sure if its because im accessing processing via a Macbook and im using an android IP Camera app?
  • TouchDesigner + IP Camera
    Used the Video Stream in, and with RSTP network protocol, while yes there is a video feed, it is SOOO choppy and pretty much only refreshes when I “pulse” the node
  • TouchDesigner + Web Camera
    https://forum.derivative.ca/t/video-device-in-webcam-doesnt-work/12208/13
    Once again seems like everyone else with Mac ver after 10.14, webcam doesnt seem to work
    I did try with the latest beta ver of Touchdesigner however, it only works when I first start a timeline, once I save the file and close it and then reopen, its just a black screen again.

So now Im stuck. Im thinking maybe a Raspberry PI(since it runs Windows) with a webcam, or a camera module and maybe something along the line of python w OpenCV library, or it will be great if i could have processing on Raspberry PI. But this is a whole new set of problems because first I don’t have a Raspberry PI and just learning to work it is like 😵🤯😩.

Low-fidelity prototype

 

  • Front Back tracking
  • Had to lift my whole hand to change the angle of the servo
  • Couldn’t have my wrist in place and just adjust my hand
  • After inserting both motors to track both front-and-back and also 360 (in which this case still 180 because using servo), the motors are going CRAZZYYY until I use just 1 finger. Likely because its taking coordinates of all 5 fingers. But, still an issue was even though just 1 finger was stuck out, the leap motion is still trying to its best to get readings of all fingers.

  • Kinda works
  • Looks very rigid and tried using a board and a bead to simulate the game for now,
    realised UX might be abit confusing. because the bead starts off being still on the board so no point in moving your hand if the goal of the game is to balance the ball on the plane for as long as possible

After consultation:

  • DC motor might not be the most practical because cant control angle of rotation, only controls the count of rotation- in which needs sensors to send feedback also
  • Build a more stable structure, have the motors screwed on!
  • Adjust the code such that if the movement is too small, the values wont be sent over to arduino- to clean up the jerky-ness
  • Webcam to detect where the ball is , to start and stop timer

Final Pitch: BALLance

Previous pitch: https://oss.adm.ntu.edu.sg/ho0011an/final-project-pitch-build-me-up/
Short recap Basically, I was too fixated on the idea of using weighing scale as the input just because I wanted that specific aspect to tie the project to the theme ‘Distant Body‘. But, overall, the initial pitch didn’t really make sense and I was left off with the question- what more can you do with ‘reflection’? & advised to do experiments first.

So- that’s what I did.

Using TouchDesigner, I tried to move a red line to where there is motion.

Trying to replicate Daniel Rozin’s trash mirror but on a digital screen. This led to a couple of random ideas.

Brief explanations on some ideas
Idea 2 Thinking along the lines of trying to create a tangible user interface perhaps where the physical ball is able to move according to the projected graphics or vice versa where the projected graphics changes according to the placement of the ball.
Idea 3 Maybe a screen based device and user can manipulate the graphics in ‘3D’ space?
Could be in a form of 2 screens or maybe 1 screen but with acrylic to give a hologram look

What Harvey Moon said: “you don't need disguise or unreal or stype to do the virtual production stuff. you can do it…

Posted by TouchDesigner on Saturday, 26 September 2020

Idea 4 Individual modules will move up to replicate the shapes of what is detect on the webcam. EG pin needle toy

With all these ideas, I was further challenged to do something with the ball. To perhaps make it into a game.

So after deliberating more, my final sketch idea is an expansion of Idea 4.

It going to be a game and the only goal is: control the plane, remotely with your palm, such that the ball stays on the plane for as long as possible. Time starts when the hand is detected and stops once the ball drops off the plane. Longest time = Highest on the score board.

Flow:

Software:
TouchDesigner
Arduino IDE

Hardware:
Leapmotion
L298n driver + Dual/Double Shaft DC Motor + Single Shaft DC Motor
Arduino UNO

Week breakdown of tasks:

Sketch MultiModal: Online Dating Accessories

 

Initial Sketches

Went with Idea 2
Concept as explained in video

Initially, I figured in my head that I would need to 2 arduinos (1 for the user- stagnant and 1 for the potential love- portable). But it was only half-way through building the project that I realised even if I was to have 2 arduinos, both of them still need to be bounded together (ref to bottom) so there’s no difference in UX even if just used 1 arduino. Which was when my project reduced quite significantly and became hardware heavy instead- led to my problems stemming from hardware management.

First lesson of the day, don’t be rough. In attempt to put the “necklace” over my head, I tore the vibration motor. Trying to put the devices around me also meant I couldn’t conceal a huge bulk of wire and I needed to make a lot of “long” wire to be able to reach from the controller to the user’s neck. Mess.

Decided to make an accessory box instead.

Takeaways:

  • Bringing it forward, an aspect i can look into is, after realising that something doesn’t go according to plan, is to be flexible enough to trial it as something else. In this case, the necklace into maybe perhaps a belt. Working with what I have and improvising along the way.
  • There needs to be a way for the user to know what the different output means. So say if someone likes your photo, the ring would light up but when the user sees it- what is going to hint them to make that association? (something very simple i can think of is maybe-literally a heart pendant/gem that likes up for a ‘like’ or since the ‘super like’ icon is a star- maybe a motor could move in the direction of how you would draw a star)
  • Hardware UI to screen based UI

DOW Multimodal: Ball

Images taken from https://www.behance.net/gallery/60769739/Ball

Ball is a smart fire extinguisher. Ball was created with the blind in mind. The whole product comes with a sensor, the spray and a refill capsule.

Here’s how it works:

In case of fire, the sensors placed around the house would detect heat and smoke.

It will alert the user via the speaker on the fire extinguisher. It will also guide the user on how to use it.

When the fire extinguisher is activated, the nozzle is able to rotate and automatically find the source of fire using the heat detection camera.

So essentially:

The very obvious advantage of this product is how is it able to make up for the user’s visual capabilities. Through audio and heat detection, Ball is still able to give the user control but at the same time aids the user with this control they have. It could very well have been a vibration/smell output instead of audio, they would all be able to alert the user anyway. But in the context of a blind person who still has to get to the fire extinguisher itself, I felt audio was the smart and better mode because blind users are usually more sensitive to auditory information and use it to locate objects faster.

The mechanics is also easy. Ball sprays with a simple pull of a loop.

Considering that this is only aimed at small initial fires, I think there should be other forms of outputs to alert the user when the fire is too intense/ too close and when they should ditch the fight against the fire. This way Ball is able to maximise the potential of the information it receives to not just become a defence mechanism but also a guide to get out of the situation safely.

I am also not too sure what is going to happen if the fire is behind the user.

In conclusion, I chose to look at this device from a multimodal point of view because I believe it has potential to grow, to collect more data and transmitting them into other various forms of output. Maybe it’s able to somehow tap into the user’s mental perception of the space using audio beacons? Maybe it’s able to give hepatic feedback based on intensity of fire? Nonetheless, I can deny this is an innovative piece of assistive tech, tapping into both “low” and “high-tech”.

 

Sketch Multimodal

IDEA 1 – Office Space
imagine 2 co-workers trying to communicate with each other discretely 

  • Pen lifting up and down – “Call me”
  • Pen spinning in the pen holder – “10 more min to lunch!!!”
  • Pen holder vibrating – “Boss is walking over!!”

IDEA 2 – PhyDigital Dating Space
when all you can think about is whether anyone liked, started a chat, superliked etc etc you on Tinder

  • Earrings pulled down – “comment”
  • Necklace starts rotating – “like”
  • Mask gets pulled down – “chat”
  • Feeling wind – “super like”

IDEA 3 – Fighting Couple needs talking
you’re in bed but can’t go to sleep because you need to talk to the person who you fought with, who also happens to be in the same bed as you

  • 3 Red LEDs – “let’s talk”
  • 2 Red LEDs – “yes”
  • 1 Red LEDs – “no”
  • 3 Green LEDs – “sleeping?”
  • ETC

DOW Health: FreeStyle Libre

Here’s how you can be one step closer to being a Cyborg. First, have diabetes and then, get yourself a set of FreeStyle Libre.

FreeStyle Libre is designed for diabetic patients as an alternative to the traditional way for monitoring blood glucose levels. In past way, to check blood glucose level, patients had to prick their fingers with a lancet (sharp point needle) and then add a couple drops of blood onto a checking strip. This strip is then inserted into a meter, which would then give off their blood glucose level. Patients would have to do this procedure a minimum of 4 times a day (depending also on the type of diabetes) to manage levels and to reduce their risk of developing a range of diabetes-related complications.

So what FreeStyle Libre does for patients is that it eliminates the need for a fingerstick routine and with each scan, provides:

  • A real-time glucose result
  • An eight-hour historical trend
  • A directional trend arrow showing where glucose levels are headed

Freestyle Libre system consists of a small fibre which pierces the skin into the interstitial fluid, a thin layer of fluid that surrounds the cells of the tissues below your skin. It takes the glucose readings and stored in the sensor. A ‘reader’ device is passed over the sensor and the last 8 h of readings can be transferred to the reader. It is simple and discreet.

(And to sound less like a commercial:)

Think barcodes at the supermarket, but now you’re the product. A quick scan of the barcode and you get the price of the product. Similarly, a quick scan of the sensor and you get your glucose readings.

This is an image I took from their website, which you can see that besides the reader and the sensor, there’s also LibreLink, which is their app that can be used in place of the reader through NPC connection, and there’s LibreLinkUp App, which allows for your data to be shared through that online cloud system.

With that, it brings me to one of the biggest pros of this device. We all know that the world is currently in the midst of the coronavirus (COVID-19) outbreak, and having this system would mean that patients to continue to connect with healthcare professionals remotely. For diabetic patients, having their healthcare team be up-to-date with their progress means being able to strategise effectively and eventually shortening the time to achieving key clinical targets. So, by not having the virus outbreak break be a hindrance to patients recovery journey makes this device a real advantage.

Another pro point to this set-up would be the convenience that it brings. As compared to a simple scan on the phone/reader, I would believe the traditional finger pricking method to be a total nightmare- can you imagine the process of having to find a clean spot, wash your hands, sterilise the lancet, force out blood etc that many times a day? And not to mention, this process feels so intrusive! If it was me, I am not sure how much discipline it is going to take for me to keep up with it everyday. There bound to be patients out there who finds it hard to sustain this routine as well and in turn deteriorating their condition; so, having the option of eliminating the whole hassle seems like a huge plus for me!

One point of improvement could be an implementation of live feedback when glucose levels fluctuate drastically. It could simply be a visual indicator or a hepatic vibration, I think this would help patients be notified of their symptoms especially when they are distracted and can then take immediate action before situation gets worst. Such implementation would improve the device to a continuous mentoring system rather than one that only works in the flash.

I’ll be leaving out the analysis of scientific concepts and accuracy in readings because for obvious reasons that I too have no idea how to make it any better. But in conclusion, I do think the design of FreeStyle Libre is a real good example to show how much technology can improve the mundane chores of everyday life. By creating something up-to-date, something revolutionary, not only makes things so much more efficient but I assume it would also encourage patients become more engaged in their care and recovery process. FreeStyle Libre’s idea of a quick scan is undeniably life changing.

Sketch: LED Room

Final Video Presentation: https://youtu.be/SjWOerja13o

I created the video story with idea 2,3,5 and 6.

Idea 2 – used in 2 instances, the first scene when I picked up my phone and the second scene when I was hiding my phone from my brother. (Privacy)

Idea 3 – used in the last scene when my mum keeps nagging at me to eat and I just cant watch my show in peace. (Warning sign to others)

Idea 5 – used when I was going to start watching my Netflix show. (Ambient Lighting)

Idea 6 – used when I was replying to a text message. (Signal busyness?)

The main goal I set for myself when I was using prototyping using ZIGSIM and OSC Communications was to get the various feedback seamlessly, all in one run.

First problem I encountered was if I wanted to use 2 different sensors, on Wekinator, I would not have the full set of input readings. More explained in this video: https://youtu.be/8VaB7EYs04k
Solution after consultation was to have more than 1 port I am transmitting the data to, so each Wekinator file would just feel from 1 sensor and not have their readings messed up.

Second problem was by combining the various ideas together, I would have to note that some of my sensors don’t work well together. This was a big problem.

If you take a look at my FIRST DRAFT for my storyline, there was so many sensors that couldn’t be applied to purpose I wanted it for. First example would be using compass to cue ambient lighting. When I wanted to use compass at first, it was very specific, from 90 degrees to 180 degrees, in my head it should work every time. BUT unaware (/dumb) me forgot that the Earth’s North is constantly changing so there were many times I find myself coming back to my code that was once working, no longer working anymore. So I decided to use initially wanted to use acceleration but this might overlap with picking up my phone that could be read as a gesture if I was too aggressive with rotating it or if I didn’t have a reliable set of examples for accurate machine learning. I ended up using Quaternion for rotation sensing.

And in order for it not to infer with throwing phone (which also uses quaternion), my solution was to combine sensor readings in Processing itself instead of Wekinator. So I had to also incorporate 3D touch to trigger flashes to happen but the hue of reds are taken by the the x coordinates from Quaternion.

In the end, I chose to present my main prototype as a projection because with the LED Strip, I couldn’t achieve certain lighting as such black lighting (for privacy) or various tint of red was not obvious when theres many pixel going off at one time.

All in all, there was A LOT of trial and error but I genuinely enjoyed learning ZIGSIM and Wekinator. No doubt there are some ideas that I still have a hard time executing such as Idea 1 but I still see so much potential in this application. Very intriguing to know that my phone could essentially be my modern day mood ring!

 

DOW IoT: Descriptive Camera

http://mattrichardson.com/Descriptive-Camera/

In this post, I will be talking about Descriptive Camera created by Matt Richardson in 2012.

Have you ever gone into an expensive restaurant, open up the menu and instead of seeing beautiful pictures of exquisite food, all you are getting is a bunch of fancy culinary terms stringed up together, and somehow they expect you to know what to order?

In my opinion, the Descriptive Camera reflects that. After the shutter is being pressed, you would expect to see the photo taken, yet instead- all you are getting is black and white text. And then it’s up to your interpretation skills to decipher what was taken- just like how you would have to imagine what your food was going to look like.

In my own (and simplified) words, this was how the Descriptive Camera was built. We have a USB Webcam, a thermal printer, 3 LEDs, 1 button(acting as a shutter button), a BeagleBone (microcontroller) connected together via a series of Python scripts. When the shutter button is pressed, it would trigger the webcam to take a photo. The photo will then be sent via the Internet to a platform where there are people waiting complete tasks. This platform is called the Amazon Mechanical Turk, and it’s almost like a 24/7 workforce. The photo is sent together with a task, which is to describe what is in the photo. Someone out that, would do exactly that and send a description back, and this will be received once again by the Internet. This output would then be translated as a print via the thermal printer.

To all my visually-inclined readers who have no idea what they just read, here’s visual aid:

These days, when a photo is taken, there’s a lot of data that comes with it, in terms of date and time of photo taken, where was it taken, camera settings of how it was taken but not so often the contents of the photo such as what they’re doing, their environment or certain adjectives to the photo. The Descriptive Camera was created with this in mind and wanted something of the latter. Richardson also believe this could be incredibly useful in being able to search, filter, and cross-reference photo collections.

These are my thoughts-
Considering that this was created in 2012, a good 8 years ago, I think he really foreshadowed what we actually see today. On our Iphones, Apple automatically uses machine learning to identify repeated faces in our Photos app and collects them into the People album. Here’s an example of what I see in my gallery:

So, this camera, though doesn’t do it, speaks of a concept that is very relevant and practical.

I also appreciated the fact that Richardson decided to build a camera form instead of just using a smart phone built-in camera. The phone itself contains a truckload of data and with his intention of streamlining data, using a phone camera could compromise just that. And though this might sound shallow but having the tactility of a physical camera is just fun! Personally, the sensation of clicking cameras is very… satisfying.

The description receipt that comes at the end also feels sort of like a reward, also seem to reassemble a Polaroid camera. The novelty of having a physical print, away from the boring digital pixel compound, excites me!

On the flip side, Richardson did mention that each print is cost money, because you do need to pay the person who is transcribing the image, and especially at an instance. There is also an apparent lag time of 3-6minutes.

I think with technology these days, if there was a chance to reconfigure the system, we could definitely look at the method of face recognition using OpenCV, Python and Deep Learning, which seems to gain traction only a couple years back. And accompanying this should be object recognition AI, which is an even more recent technology. Overall, I believe this would eliminate the need for any humans at all and a shorter time before feedback.

In conclusion, the Descriptive Camera is a paradox in many ways, you expect photos but you get text, it has a touch of rustic-ness but speaks of an idea so pertinent to the present. All in all, this work got me thinking and I genuinely enjoyed learning more about it.