How to Cope with the Future AI Act
If assisted tools are acknowledged to improve process of organisations, there are also growing concerns about their embedded algorithms which remain biased. Like previously with General Data Protection Regulation (GDPR), companies will have to comply with an upcoming EU regulation (https://artificialintelligenceact.eu/) linked to their use of AI assisted systems in order to prevent the perpetuation of historical patterns of discrimination (e.g., against women, certain age groups, persons with disabilities, or persons of certain racial or ethnic origins or sexual orientation).
- Trustworthy AI: what does it mean?
- Sources and risks of biased AI assisted tools for your organisation (illustrations: Amazon recruitment engine, compass algorithms for recidivism assessment, facial recognition)
- The AI act
- How to increase fairness in AI algorithms (metrics, bias mitigation methods, tool kits and programme, data collection, explainable AI)?
- Case study : an AI-assisted programme to match CVs with job offers avoiding age biases (use of the LIST technological demonstrator AMANDA).
Out of stock
Sorry, the course is fully booked. If you are interested in getting notified for future dates, please leave your name and email address below
June 29, 2023
June 29, 2023
Language(s) of the training
Languages spoken by the coach(es)
Expert from LIST
At the end of the training, participants are:
- aware of trustworthy AI main paradigms
- aware of the future regulation challenges for their organisations
- able to identify “risky” AI-assisted tools used in their companies
Day 1/1: 29.06.2023 08:30 – 12:30
Format and Location
This course takes place ON-SITE
Terres Rouges building
14, porte de France
No prerequisites necessary
This training is offered in collaboration with HOUSE OF TRAINING.
This training does not have any assessment or exams; a certificate of participation will be issued to participants.