• last updated 15 hours ago
Constraints
Constraints: committers
 
Constraints: files
Constraints: dates
Improved configurability of inclass exam

- added modal window for showing all configuration options

- depending on the state of an exam (whether students have stated

working on the exam or not) different options can be

altered. E.g. the grading scheme can be adjusted at any time, since

it is not visible to students prior to exam review.

- improved configurability of form-fields:

. new feature: fc_repository for shared definitions

(can be reused in the same configuration on multiple occasions)

. when form-constraints are defined with an empty short-spec,

the system tries to look this up from the fc_repository.

- extended answer manager:

. new method: student_submissions_exist

. renamed methods

. "student_submissions" -> "submissions"

(since result my include as well test run results)

. "get_answers" -> "get_answer_attributes"

to better distinguish from "submissions"

- base "merge_constraints" on dicts

- prefer forwarder over instance variable for referring

the question manager(QM)

- extended Message keys with "Configuration"

- added missing German translations in message keys

- bump version to 5.10.1d17

  1. … 9 more files in changeset.
use same idom for loading named forms as on other places

Added support for user-supplied grading schemes

Grading schemes provide a means to map achieved percentages to a numeric grade.

In the current versions, a fixed number of 5 grades is supported.

A grading scheme consists of a grounding scheme and grading boundaries.

- A grading scheme can be selected at exam definition time

- Lecturers can define their own grading schemes and reuse these between exams

- Available grounding schemes:

* no rounding (recommended for small exams, e.g. 5 minutes or 2 points)

* by percent (the calculated percentages are rounded to the provided number of digits)

* by points (the calculated points are rounded to the provided number of digits)

- The grounding precision can be defined by the user (e.g. to 2 digits)

- The grading boundaries represent percentages boundaries necessary for a certain grade

- When selecting no grading scheme, no grading information is provided in the exam protocol

(just percentages)

More changes:

- allow grading also, when student has not submitted the exam

- added percentage information in the grading-box (esp. useful for composite questions)

- renamed predefined grading schemes to more neutral terms

- provide easy-to-type names for question-manager, answer-manager and form-loader

- new utility for more robust list-comparions

- defined validating form-field type for grading boundaries

  1. … 11 more files in changeset.
file grading-procs.tcl was initially added on branch oacs-5-10.

reduce verbosity

improve spelling

  1. … 1 more file in changeset.
Document public api

added input validation for points in modal dialog

  1. … 2 more files in changeset.
fix spurious colons

  1. … 3 more files in changeset.
Added option to download results on an exam

The download is in form of an CSV file and consists of one line per

question/sub-question and including comments. This feature can be

selected after the exam was evaluated once via running the

exam-protocol.

  1. … 4 more files in changeset.
Limit display of questions in exam-protocol to max-items, when this is specified

revert change, since it is not clear from the description what it was for

Rename domNode argument to dom_node

tDOM defines a ::domNode command in the global namespace, which might potentially conflict with the argument name. Also, in certain configurations the api-doc machinery will treat the argument as a command usage and try to enforce e.g. the naming convention on it in acs-tcl.naming__proc_naming automated test.

added required argument to calc_grade (many thanks to Markus Moser for noting)

improve robustness when processing legacy questions

Added support for manual grading and individual feedback

- Lecturers can provide points and feedback comments

directly via exam protocol

- Grading is allowed, when student has submitted the exam or

the exam is not open

- Composite questions are graded at the sub-question level

- Manual grading have a higher priority than automatic grades

- Manual grading/comments can be undone by clearingfield

(showing then missing points, automated computed points, ...)

- Grades and comments are included in the exam review for

students ("Einsicht").

- Grading interactions are implemented as AJAX calls

(no need for redrawing the exam protocol, immediate feedback)

  1. … 6 more files in changeset.
remove quotes to avoid nested list interpretation

Add basic doc to public api

check if minutes are set

fix default values for short-text items

add titles for flag states

  1. … 2 more files in changeset.
add possibility to filter for state

tighten edit rights for answer workflows

VS: ----------------------------------------------------------------------

tighten policy for viewing an exam

fix collateral damage on cleanup (never trust to much on watching files)

Support for pool questions in the test-item family

Features:

- select random questions from some folder

(Currently siblings, i.e., folders of the same package instance)

- one exam can have multiple pool questions, potentially from other pools

- pools can be links to other folders (which are no siblings)

- The current folder can be used as well as a pool folder. In this

case, other not-used items can be selected as replacement items

given these match the specified filter characteristics

- Question filtering

* filter by item type:

allow one to select from all/some/some item types

(mc, sc, short text, reordering, composite, ...):

use as replacement items only examples of these types

* filter by minutes and/or points:

use as replacement items only items with matching points/minutes

* filter by language:

use as replacement items only items in a certain language

* Filter per item name pattern:

use as replacement items only items matching a name.

When (short-) names are used systematically, one can e.g.

use the date in the name and specify only items from e.g.

one year ago, via "*2020*", or from some chapter "*ch01", ...

Certainly, it is also possible to use different item

folders for this

Potential further steps:

- Currently, exams containing pool questions are treated as

auto-correctable (which implies automated exam-review (Einsicht), when

from the question pools only question from strictly closed questions

are selected (MC, SC, Ordering). Depending on the detailed settings,

also other item types could be possible (via correct-when), but this

requires a deeper analysis of every question, which is so far not

performed.

- Categorized items: technically, the infrastructure is mostly here to

allow also filtering by categories. This would allow one to select

e.g. from technical questions, case examples, knowledge transfer

questions, research-oriented questions... which are orthogonal to

the filtering currently available. Every lecturer can define

own categories depending on their needs, we could provide

university-wide category-trees, etc. Of course, one could also

define separate pools for these purposes, but categorizing is

probably more convenient and more flexible.

- Performance: when question pools become large (500+ questions) and

the cohorts as well (500+), the current version might require more

tuning. The only critical time is the exam-start, where the random

question placeholders have to be resolved for every student. The

approach from this weekend uses basic caching, but maybe this has to

be extended.

- Protecting selected questions: Question pools are more detached from

an exam than single exercises, a lecturer might have in mind. One

should not allow one to delete questions/question pools when these are

in use. Probably deletion should be a move to a trash-can, which is

actually, an issue for all exams, but getting more important with

pool questions.

- Handling of potential duplicates: When items are pulled from a

question pool, a replacement item is selected by making sure that

this item does not occur already in the query selection. Therefore,

one can safely draw two questions from the same question pool

without fearing that a student gets the same question twice.

This duplicate checking might require some fine-tuning:

* the system checks for duplicates in an exam via POOL/NAME.

* if a lecturer uses two pool-questions in an exam pointing to the same pool,

the systems makes sure, the same question is not used twice.

* however, if a teacher adds a question q1 to pool1 and the same question to pool2,

these two instances have different item ids and are regarded as two different

questions. One could make a test for only checking NAME (without POOL),

but then it might be the case that certain questions are not accepted

although these are different, because they have the same name.

- Statistics: so far, i have not provided any special answer

statistics for treating pool questions.

  1. … 9 more files in changeset.
keep track of position of a question in the form for correct scrambling also for filtering by form object

background: randomization depends on the per-user random values and by the position

of a test-item in an exam. If one is filtering just for the test-item,

the position of this item in the exam was lost. Now, we keep track

in the filtering cases

make sure, the end_clock is always an integer (as required by "clock format")

prefer xowf message catalog

provide means for omit the audio alarm control in the countdown timer

  1. … 4 more files in changeset.