GANSlider: How Users Control Generative Models for Images using Multiple Sliders with and without Feedforward Information
Hai Dang1, Lukas Mecke2,3, Daniel Buschek1
1University of Bayreuth, 2Bundeswehr University Munich, 3LMU MunichPaper Link:
https://arxiv.org/abs/2202.00965Abstract: We investigate how multiple sliders with and without feedforward visualizations influence users' control of generative models. In an online study (N=138), we collected a dataset of people interacting with a generative adversarial network (\textit{StyleGAN2}) in an image reconstruction task. We found that more control dimensions (sliders) significantly increase task difficulty and user actions. Visual feedforward partly mitigates this by enabling more goal-directed interaction. However, we found no evidence of faster or more accurate task performance. This indicates a tradeoff between feedforward detail and implied cognitive costs, such as attention. Moreover, we found that visualizations alone are not always sufficient for users to understand individual control dimensions. Our study quantifies fundamental UI design factors and resulting interaction behavior in this context, revealing opportunities for improvement in the UI design for interactive applications of generative models. We close by discussing design directions and further aspects.
Video:
- Cuda Toolkit version 10+
- Python version 3.7
- MongoDB running on default port (27017)
- Docker
@inproceedings{10.1145/3491102.3502141,
author = {Dang, Hai and Mecke, Lukas and Buschek, Daniel},
title = {GANSlider: How Users Control Generative Models for Images Using Multiple Sliders with and without Feedforward Information},
year = {2022},
isbn = {9781450391573},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3491102.3502141},
doi = {10.1145/3491102.3502141},
abstract = {We investigate how multiple sliders with and without feedforward visualizations influence users’ control of generative models. In an online study (N=138), we collected a dataset of people interacting with a generative adversarial network (StyleGAN2) in an image reconstruction task. We found that more control dimensions (sliders) significantly increase task difficulty and user actions. Visual feedforward partly mitigates this by enabling more goal-directed interaction. However, we found no evidence of faster or more accurate task performance. This indicates a tradeoff between feedforward detail and implied cognitive costs, such as attention. Moreover, we found that visualizations alone are not always sufficient for users to understand individual control dimensions. Our study quantifies fundamental UI design factors and resulting interaction behavior in this context, revealing opportunities for improvement in the UI design for interactive applications of generative models. We close by discussing design directions and further aspects. },
booktitle = {CHI Conference on Human Factors in Computing Systems},
articleno = {569},
numpages = {15},
keywords = {generative adversarial network, interactive AI, dataset, user study, image manipulation},
location = {New Orleans, LA, USA},
series = {CHI '22}
}
The code of this repository is released under the Apache 2.0 license.
GANSLIDER_DEBUGGING = [0|1] # Set this environment variable if you need debugging. GANSLIDER_PROLIFIC_URL = "" # Set this environment variable to define the callback link to prolific. REACT_APP_ROOT_DIR = "/" **
- Setup submodules:
git submodule update --init --recursive
- Rename .env.example file to .env
- Run docker-compose:
docker-compose up --scale app=3
- Open Browser and navigate to:
http://localhost:8888/user-study?PROLIFIC_PID=mock_id&STUDY_ID=mock_id&SESSION_ID=mock_session
-
Setup submodules:
git submodule update --init --recursive
- [Quickfix - Google Drive Virus Check] Download the ffhq file from https://drive.google.com/uc?export=download&id=1FJRwzAkV-XWbxgTwxEmEACvuqF5DsBiV
- Rename the downloaded file to
stylegan2_ffhq_1024.pt
- Move that file to
.backend\ganspace\models\checkpoints\stylegan2\
-
Setup Backend
- Change to the backend directory
- Install dependencies:
pip install -r requirements.txt
- Install pytorch for cuda 10.2: https://pytorch.org/get-started/previous-versions/#v181
- Set Windows System Environment Variable:
GANSLIDER_BACKEND_SERVE=1
- Run:
uvicorn server:app --workers 1 --host 0.0.0.0 --port 8000
-
Setup Fronted:
- Change to frontend directory
- Run:
npm i
- Set Windows System Environment Variable:
REACT_APP_ROOT_DIR=/
- Run: ```npm run build``
-
Open Browser and navigate to:
http://localhost:8000/user-study?PROLIFIC_PID=mock_id&STUDY_ID=mock_id&SESSION_ID=mock_session