Skip to content
Notes from the field

Notes from the field

The Ethics of Reward Shaping

The Ethics of Reward Shaping

Choosing an effective loss function is a critical part of training ML models. This thought provoking article reminds us to be critical in the choice of this function, especially as in many models the reward function itself is unclear – does a recommendation system (e.g. promoting new articles, or songs) simply create an echo chamber, or does it broadly converge on the mean? Which of these should score higher? If we penalise the system when users don't click on articles which violate their confirmation bias – are we acting ethically?
https://rob.al/2KoD1es
Musings on systems, information, learning, and optimization.

Share this:

  • Click to email a link to a friend (Opens in new window)
  • Click to share on Facebook (Opens in new window)
  • Click to share on LinkedIn (Opens in new window)
  • Click to share on Twitter (Opens in new window)
  • Click to share on WhatsApp (Opens in new window)
2018-05-03

linkedin cross-post

Post navigation

PREVIOUS
Is Open Source The AI Nirvana for Intel?
NEXT
Remote Design: How Zapier Is Building a Distributed Design Culture
Comments are closed.

Archives

The standard disclaimer…

The views, thoughts, and opinions expressed in the text belong solely to the me, and not necessarily to the my employer, organization, committee or other group that I belong to or am associated with.

Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
© 2023 Rob Aleck, licensed under CC BY-NC 4.0
Go to mobile version