» Blog
This site relies heavily on Javascript. You should enable it if you want the full experience. Learn more.

Blog

new post

Blog-posts are sorted by the tags you see below. You can filter the listing by checking/unchecking individual tags. Doubleclick or Shift-click a tag to see only its entries. For more informations see: About the Blog.

  reset tags

addon-release core-release date devvvv gallery news screenshot stuff
delicious flickr vimeo
Order by post date popularity

credits wirmachenbunt / Atelier Markgraph

This is an attempt to show a little bit more of a project, than just a video. There are some interesting bits and pieces, relying on brilliant contributions and sometimes overlooked cookies.

But first, let's have the video anyway.

The EQC-Scanner is sort of a augmented, kinetic installation. One could argue with the term augmented here but it certainly adds information to the "real" layer. The whole thing is controlled by a touch screen, allowing you to pick topics or move the screens with your fingertip. It is all and all car technology communication, but in a playful package.

Ok,let's have a look inside.

Fractal Lightning

This is based on some articles like this LINK A pretty neat recursive routine to learn what recursive is. I used c# but i bet this is easy in VL, anyone ?

The plugin was used for the battery scene. Thank you captain obvious :)

Binary Operators

Usually, when you are in RS232 or some fancy protocol land, you have to decode and encode values efficiently. Like encode high values with little use of digits. I always come back to jens.a.e'sBitWiseOps, this is one of the overlooked contributions.
For this project, the plugins were used to encode the engine controller messages.

Analysis

The image above shows some sensor recordings of the screen movement. Creating this data viz in vvvv was quick and revealed a physical feedback loop. The violet curve shows the frame difference of the screen position. And while the real screen movement actually looked kind of smooth, the sensor data showed some heavy regular spikes. Obviously, the engines did some regular overshooting. Not a big problem and solvable on the hardware side. Interesting how data visualization can help to track down problems.

Vectorraygen + DX11.Particles

Controlling Particles in a meaningful way can be painful. Using vectorfields can bring some structure into the chaos. The tool Vectorraygen helps generating the vectors the easy way. It even has a node-based enviroment to drop some organic variation into your fields. And btw. the devs are very friendly.https://jangafx.com/

The tool was used to create the key visual, 500k floating particles along the car exterior.
DX11.Particles

Boygrouping

Sure, this is not a big secret, it's one of vvvv's selling points. But i have to say, it just works. Bringing together 3 machines was actually fun. Hopefully VL preserves this killer feature.
Boygrouping

Automata UI

This was pretty much the first time i really used my own plugin and it was surprisingly helpful. :) It consists of a initialization tree for the engines & sensors and an abstract transition model of the visual software part. This is an attempt to leave content out of the statemachine but rather use the state TOPIC for every topic in use. It might be harder to read and it doesn't allow you jump to a specific content, but it makes your software expandable (without touching Automata again).

Not sure if this is a best practise for everything, let's see how this works out in the future.
Automata UI

Content Management System

There is always the dream of a fully, totally dynamic vvvv project. Our ruby-on-rails based web tool helps to manage all texts and images. It even renders the texts as images, freeing vvvv from rendering crappy text but rather display smooth pre-rendered images. Most of the vvvv project is depended to the CMS but of course, there are some bits which are always hard-wired, like 3D stuff.
CMS

I hope you find this "article" informative, any questions, comments ?

Cheers

Chris
wirmachenbunt

u7angel, Monday, Nov 5th 2018 Digg | Tweet | Delicious 8 comments  

credits Students: Jan Barth (jancouver), Roman Stefan Grasy (roman_g), Mark Lukas, Markus Lorenz Schilling (dl-110) Academics: Prof. Hartmut Bohnacker, Prof. Ralf Dringenberg

Interaction designers develop our digital everyday life. They concept, sketch and create new possibilities of interaction, which are evaluated with functional and technical prototypes.

Prototyping Interfaces

The book “Prototyping Interfaces – Interactive Sketches with VVVV” covers within 280 pages the applied handling of interactive sketches with the visual programming language VVVV.


read more...
dl-110, Monday, Feb 13th 2012 Digg | Tweet | Delicious 17 comments  

credits Alexey Beljakov, Andrey Boyarintsev, Anna Kerbo, Nikolay Matveev, Julien Vulliet, Minoru Ito

Architectural Light Studio (The ALS) is unrivaled software suite designed for creating, editing and playing back architectural and art lighting scenarios.

The ALS editor is required for:

  • creating a three-dimensional geometric layout of lighting equipment in 3D
  • creating a developed layout view on a plane with address assignment and equipment grouping
  • creating visual effects scenarios

Main benefits of the new software:

  • real-time lighting equipment simulation with calculation of illuminance according to luminaire .ies files;
  • - more than 500 IntiLED luminaires with W, RGB, RGBW color modes and up to 40 pixels per device;
  • - handy tools for creating complex layouts for luminaires of varied types, with an option to assign to building model surfaces;
  • - full-fledged display of buildings and environment as 3D models as well as ability to display dynamic lighting effects in real time;
  • - tools for creating lighting scenarios at the multi sequenced timeline;
  • - support of the Art-Net protocol enabling use of a wide spectrum of additional equipment;
  • - support up to 256 Universes and several automatic DMX addressing modes.

The debut of the ALS took place on The Light+Building 2014 in Frankfurt from March 30 to April 4, 2014

The ALS is written in vvvv by Ivan Rastr electronic crafts.
Please take a few minutes for video preview of The ALS v.0.95 beta

For further information about The Architectural Light Studio please contact IntiLED office

IntiLED team:
Sergey Matveev, Evgeniy Moskov, Uljana Vinogradova, Anton Lysenko, Elena Belova, Oleg Yanchenkov, Denis Ichalovsky

Ivan Rastr electronic crafts team:
Alexey Beljakov vnm, Andrey Boyarintsev bo27, Anna Kerbo kerbo, Nikolay Matveev unc and specially invited Julien Vulliet vux Thanks to Minoru Ito mino for helping with AntTweakBar.

ivan.rastr, Tuesday, Apr 15th 2014 Digg | Tweet | Delicious 9 comments  
Inspiration Wall

Hi all,
we'd like to share our latest VVVV / VL project.
In collaboration with ART+COM. We breathed life into a kinetic installation, which explores a dynamic interaction between the digital and the physical. In this choreography the screens interact with generative images and vice versa in real time.

More information and project video:http://schnellebuntebilder.de/projects/inspiration-wall/

All the visuals are procedurally created in VVVV to interact with the movement of the screens.
We researched a couple of generative algorithms and approaches for this project and implemented a deferred shading pipeline for raymarched content. This was used for 3D reaction diffusion and growth systems and also to shade the 3D fluid system we've been developing over the last year.
lasal supported us with a couple of conents. A fair part of the project was implemented in VL. For example the pre-simulation of the screen movements, which ensured valid driving data for the motors. We also used the VL based timeline of lasal and implemented a content switcher system in VL, which selected contents based on a set of rules and distributed scene parameters like colors between the content patches.

Content Motion-Reel:https://vimeo.com/295602967

mburk, Thursday, Nov 29th 2018 Digg | Tweet | Delicious 11 comments  

credits Kyle McLean

Waifu Synthesis- real time generative anime

Bit of a playful project investigating real-time generation of singing anime characters, a neural mashup if you will.

All of the animation is made in real-time using a StyleGan neural network trained on the Danbooru2018 dataset, a large scale anime image database with 3.33m+ images annotated with 99.7m+ tags.

Lyrics were produced with GPT-2, a large scale language model trained on 40GB of internet text. I used the recently released 345 million parameter version- the full model has 1.5 billion parameters, and has currently not been released due to concerns about malicious use (think fake news).

Music was made in part using models from Magenta, a research project exploring the role of machine learning in the process of creating art and music.

Setup is using vvvv, Python and Ableton Live.
everyoneishappy.cominstagram.com/everyoneishappy/

StyleGan, Danbooru2018, GPT-2 and Magenta were developed by Nvidia,gwern.net/Danbooru2018, OpenAI and Google respectively.

Models
GPT-2 Text
everyoneishappy, Monday, Jul 1st 2019 Digg | Tweet | Delicious 1 comments  

i am from model-in-blender-render-in-vvvv

toyshung, Wednesday, Nov 2nd 2016 Digg | Tweet | Delicious 3 comments  

credits elliot woods, chris plant (catweasel)


https://vimeo.com/84964695

In April 2012 myself and Elliot Woods were approached to create an application to project onto a car on a turntable, the brief expanded as the project went on, until by the end, it was to be a different car (of the same model) for every show, they wanted to be able to do multiple shows per day and the alignment was to be done by a member of the staff at the Landrover experience.
A challenging brief.
The building was finally opened in December 2012, the car takes approximately 10-15 minutes to lineup using 3 sliders and a rotatary control on an iPad, with a couple of buttons to spin the turntable around so you can check the setup.
It using 7 projection design projectors all running at 1920x1200, 3 for the back wall, 2 for the car and 2 for the floor, we also had to minimise any visible shadows, content was rendered at 4k and is all run off 1 pc.

catweasel, Friday, Jan 24th 2014 Digg | Tweet | Delicious 9 comments  

credits For all vvvvers :)

Original
Dark
Bright
Urbankind, Sunday, Jan 1st 2012 Digg | Tweet | Delicious 2 comments  

credits Chris Plant (catweasel) , Tonfilm for VAudio.

https://vimeo.com/243221843

Harmonic Portal is a new work commissioned for Lumiere Festival Durham, 2017 http://www.lumiere-festival.com/durham-2017/
There are 5 pieces arranged around the walls of St Godrics Church, Castle Chare, Durham, each is a portal into the surface of the wall, the light amplifies surface details due to the angle of incidence and feels like a magnifying glass (or even an electron microscope at times.).
They question how we perceive colour, the inside and outside contrast and harmonise as time progresses, while this is mirrored by the sound track which reflects the colours, the frequencies of red, green and blue light have been dropped by 42 octaves and vary in proportion to those colours of light.
It is partially contemplating how our brains create colour from 3 frequencies, but also how limited a bandwidth that actually is, the fundamental notes are a flattened Ab, A and B.
http://www.colour-burst.com/2017/11/harmonic-portal-lumiere-festival-durham-uk/

title
title
title
title
title
catweasel, Wednesday, Nov 22nd 2017 Digg | Tweet | Delicious 1 comments  

credits DiMiX

Daniil Simkin’s “Falls the Shadow” is a commissioned performance for the Guggenheim’s Works and Process series.

Four 30-minute performances took place at the Guggenheim Museum NYC on September 4 and 5, with standing audience members observing the show from different levels of the venue, which is known for its continuous ramp that spirals up the rotunda structure and connects to all of the floors.

The work combines dance with live video projections in a collection of six pieces choreographed by Alejandro Cerrudo, with costumes by Maria Grazia Chiuri /Christian Dior.
Daniil Simkin, a principal dancer with the American Ballet Theater (ABT), is also coproducer and performer in the piece, along with ABT soloist Cassandra Trenary, Ana Lopez from Hubbard Street Dance Chicago, and dancer Brett Conway.

Projection design by Arístides García and Dmitrij Simkin.
Additional shader programing, Alexander Shaliapin.
Thanks to Robert Intolight for the particles library.

Tech info:
1 single camera infrared based tracking system.
Graphic server running several vvvv instances, multiple video playback and many interactive content modules.
Laptop running custom made VL timeliner, with fully integrated MIDI controllers.
360 panorama mapping.
Interactive/generative 4k texture.
Single Nvidia1080 graphic card with 2 DP TH Matrox
6 XGA outputs: 5 projectors+1 preview
5x Christie 14K with 0,73 lens

lasal, Saturday, Oct 7th 2017 Digg | Tweet | Delicious 0 comments  

anonymous user login

Shoutbox

~13d ago

~17d ago

joreg: The Winter Season of vvvv workshops is now over but all recordings are still available for purchase: https://thenodeinstitute.org/ws23-vvvv-intermediates/

~23d ago

schlonzo: Love the new drag and drop functionality for links in latest previews!

~1mth ago

joreg: Workshop on 29 02: Create Sequencers and Precise Clock Based Tools. Signup here: https://thenodeinstitute.org/courses/ws23-vvvv-08-create-sequencers-and-precise-clock-based-tools-in-vvvv-gamma/

~1mth ago

joreg: Workshop on 22 02: Unlocking Shader Artistry: A Journey through ‘The Book of Shaders’ with FUSE. Signup here: https://thenodeinstitute.org/courses/ws23-vvvv-12-book-of-shaders/

~2mth ago

joreg: Talk and Workshop on February 15 & 16 in Frankfurt: https://visualprogramming.net/blog/vvvv-at-node-code-frankfurt/

~2mth ago

woei: @Joanie_AntiVJ: think so, looks doable