mac online apple blackjack http://www.euro-online.org

Live Electronic Performance: Noise sounds like telegram

telegram

Pure Data is pretty fun that is allowed to get a quick try with any possibility of linking and changing the value, which makes me feel less pressure and more motivation do the experiment of noise. This is the testing of how to make the noise with low/medium pitch – the outcome sounds like the combination of telegramming and whisper from data, which keeps me arising the further idea about improving the theme of duo performance, and it still needs to consult with mate next week.

Experiment: Quartz Composer with kinect

Screen Shot 2014-01-27 at 11.58.15


It is really exhausting to record that I found there are some boundaries of interactive visual expression after I tried Processing and Pure Data in terms of the form of particles, despite both of them have advantages in different aspects, such as audio or the flexibility of editing code.

I came across Quartz Composer when I was searching particle system with kinect, and then I have to say that I tried different software again….Actually, this experiment is a good start I reckon, which really saves me time to get a try and have fun playing with these stuff by using data stream.

Testing 1:
(1) add image
(2) change X|Y position and gravity.

Testing 2: Hand tracking
(1) connect QC with kinect (the tutorial of connecting can be found on Youtube)
(2) change the parameter

Testing 3: Hand tracking without body shape
get a try to add another library patch into it.

I will keep learning QC and PD after doing a series of experiment by different software – data stream is really suitable for those who have no idea with computer language to build confidence keep on going the next step.

Updated (24th January 2014)

Testing 4: Blur function with mouse
get a try based on template – black and white function has been added, and the image can be moved by mouse.

Testing 5: Blur function with video camera
make video camera be shown inside the surface of cube.

Updated (26th January 2014)

Testing 6: silhouette via kinect in QC
I am still not satisfied to see the final effect after doing a series of experiments, since the silhouette effect  remains a problem in this case. 

Updated (27th January 2014)

Testing 7: Hand tracking via kinect in QC
This might be the last testing of kinect project of Semester 1, which shows the concept of elapse of time. I changed data flow of particles and the parameters based on ‘kinect test’ of Quartz Composer, such as gravity and attraction, which circles can be shown radially when the user moves his/her both hands. At last, I feel a bit disappointed more or less due to the unsuccessful experiments, on the other hand, thanks sincerely for those who had supported quite a lot in terms of Processing and Kinect during we were doing a series of experiments about group project. By the way, it will be a good start in Semester 2 that sessions involve that learning Processing and doing the homework.

Bibliography
Challinor, R. (2011). Available at: Youtube,  Synapse for Kinect Quartz Composer Tutorial (Accessed: 21 January, 2014).
Robinson, G. and Buchwald, S. (2012). Learning quartz composer: A hands-on guide to creating motion graphics with quartz composer. Corporate and Government Sales, U.S.

References: Virtual element and Particles effect

20140106095113

Reference 1:

INTEGRATION is a real-time / interactive art project (S. A. Tawil 2013)

The similar work about our group project’s previous concept appears coincidently yesterday. Before that, we didn’t find similar work except the effects of particles. Regarding particles, dropping inside the silhouette based on Lin’s advice in November – the comparison between falling outside and filling inside.

Reference 2:

Dynamic Sand Effect by C. Guebert

We found some effect by Processing and Kinect at the beginning, such as sand effect (Jun’s advice), and focused on the one called Falling Sand (Guebert year unknown) after searching several ones that we could do the reference.

Available from: Openprocessing.org [Accessed 6 January, 2014 ].

Previous Reference Below:

Here are the previous references that I found before, the inspiration was constantly built by thinking about fluid characteristics of time which exists in the dimension that we live in but it does not stay anywhere. Time notion provokes me a question to the limitation of human cognition for a long time, as the motivation to ‘see’ time being and to let audience perceive its existence.

Relevant reference about ‘Time’
In spacetime, a coordinate grid that spans the 3+1 dimensions locates events(rather than just points in space), i.e., time is added as another dimension to the coordinate grid.

Spacetime, Available at: Wikipedia. (http://en.wikipedia.org/wiki/Spacetime) [08 January 2014].

Reference 3:

Installation Interactive Kinect + Processing (Brother System 2013)

 Available from: Youtube. Installation Interactive Kinect + Processing [Accessed 11 November 2013].

Reference 4:

CAN Kinect Physic Tutorial 

 Available from: Vimeo. CAN Kinect Physics Tutorial [Accessed 11 November, 2013].

Experiment: Particle effect testing (Pure Data)

particles_pd_testing

Share a fairly basic practice for testing particles that was made after understanding the PD tutorials in last week. The outcome of group project effect will be mainly handled by group member Wengchang who is familiar with AfterEffect. Although we focus on the combination of virtual element and photography in this project work, we three still learn interactive technique for the further development in either project work or individual one.

Here have a try and make it be possible if interactive effect between audience and particles can be achieved, though I still get stuck in the interactive step with lack of sufficient skills. (The camera effect was combined by a coincidence…not quite understand some code in it, but it seems a good combination!)

An idea about individual project comes from this testing – may I use PD Gem to fulfill the synchronised two-layer’s variation with sounds?

Testing:

Reference
morefun4art (2012) Available at: Youtube.  VA3460 Embodied Interaction [Accessed 26 December 2013].

Hotchkiss, J. (2010) Motion detection to midi with puredata. Available at:  http://hotchk155.blogspot.co.uk/2010/05/webcam-motion-detection-to-midi-with.html/ [Accessed 26 December 2013].

Group Project: Demo framing of pre-set environment (Photos by W.Lin)

demo1

This testing was focused on the effect of comparison between virtual particles and pre-set environment.

The demo framing of Pre-set environment and sequence motion:

Photos were photographed by W.Lin
Date: 24.12.2013
Location: Newcastle University

Inspiration: About pre-set environment

2

After watching several interactive artworks, such as Tamás Waliczky The Garden (1992), I am thinking about the virtual environment in digital art – how about if we could use pre-set real environment instead of the completely virtual space? The concept of ‘see’ time being will be achievable, if the audience is able to perceive the virtual particles (time flowing) in a ‘real’ environment.

Waliczky (1992) Available at:
http://www.waliczky.com/waliczky_start.htm/ [Accessed 26 2013]

Idea of developing: Group project

The conception of the project called ‘time elapsing’

In terms of the initial idea
The conception is to express the feeling of time elapsing.

At the beginning, the inspiration emerged about ‘interactive electronic clock by dim vision’. After made a discussion with fellows, we developed the initial idea instead of using interactive LED table and decided to focus on the connection between audience and virtual element. The concept of ‘clock’ also will be replaced by virtual element. The metaphor of falling particles is that time is passing continually, meanwhile, human’s behaviour will be affected and adjusted in certain time.

Idea of developing (edited 8 Jan 2014):

Time conception as an elusive notion can be relatively hard to explain, which provokes
the question to several fields, from psychology to religion, however, the computer vision provides
the possibility to explore the embodiment of time perception. I think we could use virtual element to achieve abstract notion become visible and then combined photographic theory and dynamic virtual effect in the real environment in order
to imply human cognition to time-consciousness that depends on individual’s events.