High-risk AI systems should perform consistently throughout their lifecycle and meet an appropriate level of accuracy, robustness and cybersecurity, in the light of their intended purpose and in accordance with the generally acknowledged state of the art. The Commission and relevant organisations and stakeholders are encouraged to take due consideration of mitigation of risks and negative impacts of the AI system. The expected level of performance metrics should be declared in the accompanying instructions of use. Providers are urged to communicate this information to deployers in a clear and easily understandable way, free of misunderstandings or misleading statements. The EU legislation on legal metrology, including on Measuring Instruments Directive (MID) and Non-automatic weighing instruments (NAWI) Directive, aims to ensure the accuracy of measurements and to help the transparency and fairness of commercial transactions. In this context, in cooperation with relevant stakeholders and organisation, such as metrology and benchmarking authorities, the Commission should encourage, as appropriate, the development of benchmarks and measurement methodologies for AI systems. In doing so, the Commission should take note and collaborate with international partners working on metrology and relevant measurement indicators relating to Artificial Intelligence.