RemNote Community
Community

Study Guide

📖 Core Concepts Interactive art – artwork that requires spectator input to fulfill its purpose (e.g., walking through, touching, contributing). Spectator roles – visitor / v‑user / immersant – the person who actively engages with the piece. Sensors & computers – motion, heat, proximity, temperature, etc., feed data to a computer that processes input and generates a responsive output. Immersive/VR – full‑field environments that engage all senses; the visitor interacts through sight, sound, motion, and sometimes haptics. Dialogue vs. monologue – interactive art = two‑way conversation (audience ↔ machine); generative art = one‑way output that may change without audience agency. 📌 Must Remember Early digital interactive art → mainstream late‑1990s; museums began dedicated exhibitions. Three core interaction types: Navigation – move through/around the work. Assembly – arrange or combine elements. Contribution – add new content or modify existing content. Conversation Theory (Gordon Pask) – two‑way information exchange that shaped 1970s interactive art. Key toolkits (pick the right one for hardware vs. software): Arduino, Max/MSP, Processing, OpenFrameworks, Pure Data, TouchDesigner. Hybrid discipline – artists + architects + engineers create custom sensors, actuators, projections, and networked communication. 🔄 Key Processes Sensor → Computer → Output Loop Detect input (motion, heat, proximity). Software (e.g., Max/MSP, Processing) interprets data. Generate visual/audio/kinetic response (LEDs, speakers, robotic movement). Designing an Interactive Installation Define desired spectator action (navigate, assemble, contribute). Choose appropriate sensor (IR motion, pressure, microphone). Prototype with Arduino (read sensor → send MIDI/OSC). Build interactive logic in a visual language (Max/MSP, TouchDesigner). Test for latency & robustness; iterate. VR Immersive Experience Creation Model 3‑D environment. Map head‑tracking and hand controllers to navigation & manipulation. Use game engine or custom OpenFrameworks/Processing sketch for real‑time rendering. 🔍 Key Comparisons Interactive art vs. Generative art – Interactive: audience acts, creates unique outcome. Generative: system acts on its own; audience is passive observer. Navigation vs. Assembly vs. Contribution – Navigation: movement through space (e.g., walk‑through installation). Assembly: physically rearrange pieces (e.g., modular sculpture). Contribution: add new content (e.g., collaborative digital collage). Arduino vs. Max/MSP – Arduino: low‑level hardware I/O, C/C++‑style programming. Max/MSP: high‑level visual language for audio‑visual signal flow; no soldering. ⚠️ Common Misunderstandings “All digital art is interactive.” – Only works that require spectator input qualify. “Generative = interactive.” – Generative systems can run autonomously; they lack the agency‑inviting dialogue. “Sensors automatically make a piece interactive.” – Sensors must be wired to logic that maps input to meaningful output. “VR is always immersive.” – Immersion depends on stimulus breadth (visual, audio, haptic); a simple 3‑D view may still be non‑immersive. 🧠 Mental Models / Intuition Conversation Model – Think of the artwork as a person that asks a question (sensor) and replies (output). Each visitor’s answer yields a different reply. Input‑Process‑Output (IPO) Pipeline – Treat every interactive piece as a black box: what you feed in (input) determines what comes out (output) after the internal algorithm runs. 🚩 Exceptions & Edge Cases Passive “responsive” installations – may react to environmental data (temperature) but lack spectator agency → classify as responsive rather than interactive. Hybrid works – a piece may combine navigation and contribution (e.g., VR world you walk through and also paint on). Large‑scale public façades – often limited to simple proximity triggers; deeper interaction may be constrained by safety or budget. 📍 When to Use Which Choose Arduino when you need direct hardware control (motors, LEDs, custom sensors). Choose Max/MSP for real‑time audio/video processing with minimal coding. Choose Processing for quick visual prototypes and easy export to standalone apps. Choose OpenFrameworks when you need C++ performance and integration with custom hardware. Choose Pure Data for interactive music and low‑latency audio synthesis. Choose TouchDesigner for large‑scale, GPU‑heavy visual installations (projection mapping, LED walls). 👀 Patterns to Recognize Sensor → visual/audio change pattern repeats across most installations. Narrative branching in interactive film/storytelling: each decision point splits the story tree. Spatial mapping – visitor position ↔ projected imagery or sound field (common in immersive rooms). Feedback loops – output influences subsequent input (e.g., moving lights that affect visitor path). 🗂️ Exam Traps “Interactive art always uses computers.” – Wrong; early works used video, satellite, or mechanical systems. “All VR works are interactive.” – Incorrect; a VR scene can be a passive 360° video with no user agency. Confusing “generative” with “interactive.” – Test‑takers may pick the answer that mentions “change in presence” without agency; remember the dialogue requirement. Mix‑up of tool capabilities – e.g., selecting Max/MSP for low‑level sensor wiring (Arduino is the correct answer). Assuming “installation art” = “interactive art.” – Only installations that require spectator input qualify; many are static.
or

Or, immediately create your own study flashcards:

Upload a PDF.
Master Study Materials.
Start learning in seconds
Drop your PDFs here or
or