Follow
Vishal Maini
Title
Cited by
Cited by
Year
Scalable agent alignment via reward modeling: a research direction
J Leike, D Krueger, T Everitt, M Martic, V Maini, S Legg
arXiv preprint arXiv:1811.07871, 2018
2752018
Reducing sentiment bias in language models via counterfactual evaluation
PS Huang, H Zhang, R Jiang, R Stanforth, J Welbl, J Rae, V Maini, ...
arXiv preprint arXiv:1911.03064, 2019
1772019
Machine learning for humans
V Maini, S Sabri
Retrieved on May 1, 2022, 2017
972017
Building safe artificial intelligence: specification, robustness, and assurance
PA Ortega, V Maini, DMS Team
DeepMind Safety Research Blog, 2018
392018
Scalable agent alignment via reward modeling: A research direction. arXiv 2018
J Leike, D Krueger, T Everitt, M Martic, V Maini, S Legg
arXiv preprint arXiv:1811.07871, 1811
151811
the DeepMind safety team
PA Ortega, V Maini
Building safe artificial intelligence: specification, robustness, and assurance, 2018
132018
Machine learning for humans. 2017
V Maini, S Sabri
URL: https://medium. com/machinelearning-for-humans/why-machine-learning …, 2019
122019
Scalable agent alignment via reward modeling: a research direction. arXiv
J Leike, D Krueger, T Everitt, M Martic, V Maini, S Legg
arXiv preprint arXiv:1811.07871, 2018
82018
Machine Learning For Humans (6 X 9): Introduction to Machine Learning with Python
V Maini, S Sabri
Alanna Maldonado, 2023
22023
The system can't perform the operation now. Try again later.
Articles 1–9