Here's what we found for the Vision tag:
Our lab has extensive experience in using different sensing technology including eye tracking and facial emotion recognition (DiPaola et al 2013), as well as gesture tracking and bio sensing heart rate and EDA (Song & DiPaola, 2015) which both affect the generative system and can be used to understand the reception to the generated graphics (still, video, VR).http://ivizlab.sfu.ca/research/biosense
What is abstraction? Can you use AI techniques to model the semantics of an idea, object, or entity, where that understanding allows for abstraction of the meaning? We use several AI techniques including genetic programming, Neural Nets and Deep Learning to explore abstraction in its many forms. Mainly here in the visual and narrative arts.http://ivizlab.sfu.ca/research/cogabstract
Using Cognitive Science as a basis for our work, we attempt to model aspects of human creativity in AI. Specially we are using Neural Networks (and evolutionary systems) in the form of Deep Learning, CNNs, RNNs and other modern techniques to model aspects of human expression and creativity. We are known for modelling expression semantics and generation of visual art (stills, videos, VR) but have extended our work into expressive forms of linguistic (word based) narrative.http://ivizlab.sfu.ca/research/deepai
Can you build a computer parameterized model of the cognitive process or open methodology that human portrait painters use. This interdisciplinary toolkit research project automatically generates painterly output from people photographs and cognitive parameters.http://ivizlab.sfu.ca/research/painterly
Using new computer modelling techniques, we show that artists including Rembrandt, use vision based techniques (lost and found edges, center of focus techniques) to guide the eye path of the viewer through their paintings in significant ways...http://ivizlab.sfu.ca/research/rembrandt
Our open source toolkit / cognitive research in AI 3D Virtual Human (embodied IVA : Intelligence Virtual Agents) - a real-time system that can respond emotionally (voice, facial animation, gesture, etc) to a user in front of it via a host of gestural, motion and bio- sensor systems. The system uses SmartBody (USC) and MatLab Simulink as its control and AI system.http://ivizlab.sfu.ca/research/virthuman
Choi S K, DiPaola S,—Proc ACM Conf on Human Factors in Computing Systems (CHI '15), , pp. 579-588, 2015.
DiPaola S,—In Proc: Electronic Visualisation and the Arts, 10 pages, London, 2007.
DiPaola S,—In Proc: Electronic Visualisation and the Arts, 8 pages, London, 2008.
DiPaola S,—In Proc: Electronic Visualisation and the Arts, 7 pages, London, 2010.
Choi S K, DiPaola S,— Proc; Electronic Visualisation and the Arts, pp. 127-134. British Computer Society, London, 2013.
Seifi H, DiPaola S, Arya A, —International Journal of Computer Games Technology, International Journal of Computer Games Technology, vol. 2011, Article ID 164949, 7 pages, 2011.
DiPaola S, Riebe C, Enns J,—Leonardo, Vol 43, No 3, pp 145-151, 2010.
DiPaola S,—In Proc: SPIE Human Vision and Imaging, Int. Society for Optical Engineering, Keynote paper. pp 33-43, 2007.
DiPaola S,—In Proc: SPIE Human Vision and Imaging, Int. Society for Optical Engineering, 6 pages, 2009.
Riebe C, DiPaola S, Enns J,—Journal of Vision, Vol 9, No 8, pp 368 (abstract), 2009.