A site to display results for DEN.
- Isolate all semester-specific / special case logic from the main parts of the code. I.e. we want to translate all the data to some common format that isn't semester-specific.
- Make sure that single-user responses aren't visible in the front end. I.e. we're publishing aggregated data, and don't want to make the whole data set public - at least, not intentionally. We may want to provide breakdowns by demographic categories - i.e. graphing freshmen/sophomore/junior/senior separately.
- The front end should be agnostic to questions and pecularities of each semester's surveys - it should display what it's given. The idea is for the frontend to be a fairly generic graphing platform, only really coupled to the format of the json it is given to graph.
- ability to filter freeform answers as needed.
Each semester's data should be downloaded to the proper folder in raw_data. When backend/preprocess.py is run, this data will then be translated to python objects (Surveys and Questions in backend/survey_obj.py) by semester and survey-platform specific code. From these intermediate python objects, the data is then summarized, forming the json files in the summarized_data folder. Then the json files are used to create the csv files, containing just the comments.
The intended workflow is to take the csv files in summarized_data, upload them to google docs (with conversion turned on). Then fill in the include (y/n) column for each spreadsheet (1 per semester). When this is done, download the files as csv and run (for each semester):
./backend/comment_filtering.py make_new_json summarized_data/<original summarized json> <annotated csv>
The result will be saved as a json file in the filtered_data folder
To completely deploy everything:
- Attach the javascript files in frontend/ to the wiki
- Attach the json files from summarized_data/ to the wiki
- Take the body of displayChart.html, and copy-paste it onto the wiki. Rewrite the external scripts src's to point to the js attachments.
- Make sure the wiki version is able to load the json files. This is simply a matter of the javascript getting the path right. This could be a problem if you create a new page for the surveys (instead of the current one).
Currently, the best way to test is to run
python -m SimpleHTTPServer
From within the frontend folder. Copy the generated .json files from summarized_data (or from filtered_data - they use the same format) into the frontend folder, and then open displayChart.html in a browser (typically, as http://localhost:8000/displayChart.html)