UQ Students should read the Disclaimer & Warning
Note: This page dates from 2005, and is kept for historical purposes.
<?xml version="1.0" encoding="iso-8859-1"?>
<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.1//EN"
"http://www.w3.org/TR/xhtml11/DTD/xhtml11.dtd">
<html xmlns="http://www.w3.org/1999/xhtml">
<head>
<title>COMP1501 - What I want</title>
<meta http-equiv="Content-Type" content="text/html; charset=iso-8859-1" />
</head>
<body>
<h1>COMP1501 Internet Interface Design – What I Want</h1>
<p>I want personal feedback including the criteria used and justification of my
result. I want to know how many people marked the assignments, and if one
or more of those who marked them gave inconsistent results, and if so, a full
and fair remark – free of bias, and if not, what efforts were made to
cross-mark or normalize personal marking differences between markers. </p>
<p>The criteria with which we will be assessed must be made available before the
assessment. The criterion that we were given to do our assignment is not the
same as that used to assess the assignment. Therefore, my, along with everyone
else’s, assignment was completed to meet different criterion than that
with which it was marked. I believe this is unfair and against university
policy; and I wish to know what Kevin thinks on this matter.</p>
<p>I want to know why some assignments that appear to have met most or all of
the criteria have achieved marks well below others that appear to have met
fewer of the criteria - or to put that simply, I want to know why crap sites
often got better marks than good ones.</p>
<p>The results just seem unfair, biased and wrong to me. I want to know why, and
getting feedback to see exactly how my site got the marks it did is the first
step in answering this.</p>
<p>I dug up a few relevant sections of the fine print:<br />
<a href="COMP1501-Assignment-2-query">Assessment principles</a><br />
</p>
</body>
</html>