Dimitri Coelho Mollo

Assistant Professor in Philosophy of Artificial Intelligence at Umeå University

AI Alignment through Reinforcement Learning from Human Feedback? Contradictions and Limitations


Unpublished


Adam Dahlgren Lindström, Leila Methnani, Lea Krause, Petter Ericson, Íñigo Martínez de Rituerto de Troya, Dimitri Coelho Mollo, Roel Dobbe
2024

DOI: arXiv:2406.18346

ArXiv
Cite

Cite

APA   Click to copy
Lindström, A. D., Methnani, L., Krause, L., Ericson, P., de Rituerto de Troya, Í. M., Mollo, D. C., & Dobbe, R. (2024). AI Alignment through Reinforcement Learning from Human Feedback? Contradictions and Limitations. ArXiv. https://doi.org/arXiv:2406.18346


Chicago/Turabian   Click to copy
Lindström, Adam Dahlgren, Leila Methnani, Lea Krause, Petter Ericson, Íñigo Martínez de Rituerto de Troya, Dimitri Coelho Mollo, and Roel Dobbe. “AI Alignment through Reinforcement Learning from Human Feedback? Contradictions and Limitations.” ArXiv, 2024.


MLA   Click to copy
Lindström, Adam Dahlgren, et al. AI Alignment through Reinforcement Learning from Human Feedback? Contradictions and Limitations. ArXiv, 2024, doi:arXiv:2406.18346.


BibTeX   Click to copy

@unpublished{adam2024a,
  title = {AI Alignment through Reinforcement Learning from Human Feedback? Contradictions and Limitations},
  year = {2024},
  doi = {arXiv:2406.18346},
  author = {Lindström, Adam Dahlgren and Methnani, Leila and Krause, Lea and Ericson, Petter and de Rituerto de Troya, Íñigo Martínez and Mollo, Dimitri Coelho and Dobbe, Roel},
  howpublished = {ArXiv}
}



Tools
Translate to