diff --git a/knowledge/survey/browser.py b/knowledge/survey/browser.py index 4a76e83..5de6302 100644 --- a/knowledge/survey/browser.py +++ b/knowledge/survey/browser.py @@ -76,7 +76,8 @@ class SurveyView(ConceptView): if data: resp = Response(self.adapted, None) for qu in self.adapted.questions: - resp.values[qu] = data[qu.uid] + if qu.uid in data: + resp.values[qu] = data[qu.uid] qgAvailable = True for qg in self.adapted.questionGroups: if qg.uid in data: @@ -85,8 +86,8 @@ class SurveyView(ConceptView): qgAvailable = False if not qgAvailable: values = resp.getGroupedResult() - for qugroup, info, score in values: - resp.values[qugroup] = score + for v in values: + resp.values[v['group']] = v['score'] result.append(resp) return result @@ -106,22 +107,22 @@ class SurveyView(ConceptView): data[uid] = value response.values[question] = value values = response.getGroupedResult() - for qugroup, info, score in values: - data[self.getUidForObject(qugroup)] = score + for v in values: + data[self.getUidForObject(v['group'])] = v['score'] respManager.save(data) self.data = data self.errors = self.check(response) if self.errors: return [] - result = [dict(category=r[0].title, text=r[1].text, - score=int(round(r[2] * 100))) - for r in values] + result = [dict(category=r['group'].title, text=r['feedback'].text, + score=int(round(r['score'] * 100)), rank=r['rank']) + for r in values] if self.adapted.showTeamResults: teamData = self.getTeamData(respManager) - ranks, averages = response.getTeamResult(values, teamData) - for idx, qgdata in enumerate(result): - qgdata['rank'] = ranks[idx] - qgdata['average'] = int(round(averages[idx] * 100)) + values = response.getTeamResult(values, teamData) + for idx, r in enumerate(values): + result[idx]['average'] = int(round(r['average'] * 100)) + result[idx]['teamRank'] = r['rank'] return result def check(self, response): diff --git a/knowledge/survey/view_macros.pt b/knowledge/survey/view_macros.pt index b7b6a4d..56af418 100644 --- a/knowledge/survey/view_macros.pt +++ b/knowledge/survey/view_macros.pt @@ -13,21 +13,21 @@

Feedback

+ tal:content="structure python: + item.renderText(header, 'text/restructured')" /> + i18n:translate="">Score + + i18n:translate="">Team Score - @@ -37,12 +37,13 @@ tal:content="fbitem/text" />
Category Response Score %Rank Rank Team Score %Team Rank
+ - - +
diff --git a/locales/de/LC_MESSAGES/loops.mo b/locales/de/LC_MESSAGES/loops.mo index bbb2a9a..6d61ea8 100644 Binary files a/locales/de/LC_MESSAGES/loops.mo and b/locales/de/LC_MESSAGES/loops.mo differ diff --git a/locales/de/LC_MESSAGES/loops.po b/locales/de/LC_MESSAGES/loops.po index 00fd886..a674499 100644 --- a/locales/de/LC_MESSAGES/loops.po +++ b/locales/de/LC_MESSAGES/loops.po @@ -3,7 +3,7 @@ msgstr "" "Project-Id-Version: 0.13.0\n" "POT-Creation-Date: 2007-05-22 12:00 CET\n" -"PO-Revision-Date: 2014-05-07 12:00 CET\n" +"PO-Revision-Date: 2014-05-24 12:00 CET\n" "Last-Translator: Helmut Merz \n" "Language-Team: loops developers \n" "MIME-Version: 1.0\n" @@ -265,17 +265,17 @@ msgstr "Bitte beantworten Sie die angegebene Mindestanzahl an Fragen je Fragengr msgid "Obligatory question, must be answered" msgstr "Pflichtfrage, muss beantwortet werden" -msgid "Score %" -msgstr "Bewertung %" +msgid "Score" +msgstr "Bewertung" -msgid "Team Score %" -msgstr "Durchschnitt Team %" +msgid "Team Score" +msgstr "Durchschnitt Team" msgid "Rank" -msgstr "Rang im Team" +msgstr "Rang" msgid "Team Rank" -msgstr "Rang des Teams" +msgstr "Rang Team" # competence (qualification)