diff --git a/common/lib/xmodule/xmodule/js/src/peergrading/peer_grading_problem.coffee b/common/lib/xmodule/xmodule/js/src/peergrading/peer_grading_problem.coffee index de5b5be46a..4574ec3c93 100644 --- a/common/lib/xmodule/xmodule/js/src/peergrading/peer_grading_problem.coffee +++ b/common/lib/xmodule/xmodule/js/src/peergrading/peer_grading_problem.coffee @@ -459,6 +459,7 @@ class @PeerGradingProblem if response.actual_rubric != undefined calibration_wrapper.append("
Instructor Scored Rubric: #{response.actual_rubric}
") + if response.actual_feedback!=undefined calibration_wrapper.append("
Instructor Feedback: #{response.actual_feedback}
") # disable score selection and submission from the grading interface diff --git a/common/lib/xmodule/xmodule/peer_grading_module.py b/common/lib/xmodule/xmodule/peer_grading_module.py index 3916c76d37..de4922e60a 100644 --- a/common/lib/xmodule/xmodule/peer_grading_module.py +++ b/common/lib/xmodule/xmodule/peer_grading_module.py @@ -435,7 +435,6 @@ class PeerGradingModule(PeerGradingFields, XModule): try: response = self.peer_gs.save_calibration_essay(location, grader_id, calibration_essay_id, submission_key, score, feedback, rubric_scores) - log.debug("RESPONSE RESPONSE : {0}".format(response)) if 'actual_rubric' in response: rubric_renderer = combined_open_ended_rubric.CombinedOpenEndedRubric(self.system, True) response['actual_rubric'] = rubric_renderer.render_rubric(response['actual_rubric'])['html']