Please note that DQF tools is NOT part of the Quality Dashboard!
DQF provides a vendor independent environment for evaluating translated content. Users can post-edit machine-translated segments to track productivity, evaluate adequacy and fluency of target sentences, compare translations and count errors based on an error-typology. The tools help establish return-on-investment and benchmark performance enabling users to take informed decisions.
To ensure evaluation results are reliable it is vital that best practices are applied. The DQF Tools help users apply best practices whether they are selecting their preferred engine, measuring productivity or evaluating the quality of translations.
The DQF Content Profiling feature is used to help select the most appropriate quality evaluation model for specific requirements. This leads to the Knowledge base where you find best practices, metrics, step-by-step guides, reference templates, and use cases.
Your underlying process, technology and resources will guide the choice of the quality evaluation model.
DQF-MQM Harmonized Error Typology
TAUS has made available the error review functionality to the open DQF API that allows TAUS members and users to integrate quality review in their day-to-day tool environment and track and benchmark their translation performance on the TAUS Quality Dashboard. The error review API is based on a harmonization of the DQF and MQM error typologies, that was funded by the European Commission as part of the QT21 project. SDL Trados Studio, SDL WorldServer, MateCat and SDL TMS will be among the first translation technologies that will implement the error review functionality of the TAUS Dynamic Quality Framework (DQF) tools.
More info on QT21: http://www.qt21.eu/wp-content/uploads/2015/11/QT21-D3-1.pdf
Click here to view the complete DQF-MQM hierarchy.