Lecturas
Sobre las lecturas
Este curso es autocontenido: todo lo necesario para seguir las clases y completar los laboratorios se cubre en las sesiones presenciales. Las lecturas que se listan a continuacion son opcionales, pensadas para quienes quieran profundizar en algun tema o explorar aplicaciones adicionales. No es necesario leerlas antes de las sesiones.
Lecturas por dia
Dia 1: Fundamentos de IA y Machine Learning
Russell, S. & Norvig, P. (2021). Artificial Intelligence: A Modern Approach. Capitulo 1.
James, G., Witten, D., Hastie, T. & Tibshirani, R. (2021). An Introduction to Statistical Learning with Applications in R. Capitulos 1-2.
Kuhn, M. & Silge, J. (2022). Tidy Modeling with R. O’Reilly. Capitulos 1-3.
Wickham, H., Cetinkaya-Rundel, M. & Grolemund, G. (2023). R for Data Science (2da edicion). O’Reilly. Capitulos 1-4.
Dia 2: Aprendizaje supervisado
Muchlinski, D. et al. (2016). Comparing random forest with logistic regression for predicting class-membership. Political Analysis, 24(2), 168-185.
Mullainathan, S. & Spiess, J. (2017). Machine learning: An applied econometric approach. Journal of Economic Perspectives, 31(2), 87-106.
Athey, S. & Imbens, G. W. (2019). Machine learning methods that economists should know about. Annual Review of Economics, 11, 685-725.
James, G., Witten, D., Hastie, T. & Tibshirani, R. (2021). An Introduction to Statistical Learning. Capitulos 4-6 y 8.
Dia 3: Texto y aprendizaje no supervisado
Grimmer, J. & Stewart, B. (2013). Text as data: The promise and pitfalls of automatic content analysis methods for political texts. Political Analysis, 21(3), 267-297.
Silge, J. & Robinson, D. (2017). Text Mining with R: A Tidy Approach. O’Reilly.
Roberts, M. E., Stewart, B. M. & Tingley, D. (2019). stm: An R package for structural topic models. Journal of Statistical Software, 91(2), 1-40.
James, G., Witten, D., Hastie, T. & Tibshirani, R. (2021). An Introduction to Statistical Learning. Capitulos 10 y 12.
Dia 4: LLMs y aplicaciones
Gilardi, F., Alizadeh, M. & Kubli, M. (2023). ChatGPT outperforms crowd workers for text-annotation tasks. PNAS, 120(30).
Bail, C. A. (2024). Can generative AI improve social science? PNAS, 121(21).
Tornberg, P. (2024). Best practices for text annotation with large language models. Sociological Methods & Research.
Argyle, L. P. et al. (2023). Out of one, many: Using language models to simulate human samples. Political Analysis, 31(3), 337-351.
Vaswani, A. et al. (2017). Attention is all you need. Advances in Neural Information Processing Systems, 30.
Dia 5: Etica y sesgo algoritmico
O’Neil, C. (2016). Weapons of Math Destruction. Crown.
Barocas, S., Hardt, M. & Narayanan, A. (2023). Fairness and Machine Learning. MIT Press.
Eubanks, V. (2018). Automating Inequality: How High-Tech Tools Profile, Police, and Punish the Poor. St. Martin’s Press.
Raji, I. D. et al. (2020). Closing the AI accountability gap: Defining an end-to-end framework for internal algorithmic auditing. ACM Conference on Fairness, Accountability, and Transparency.
Bibliografia complementaria
Estas lecturas no son obligatorias, pero ofrecen perspectivas valiosas para profundizar en temas especificos:
- Athey, S. (2017). Beyond prediction: Using big data for policy problems. Science, 355(6324), 483-485.
- Breiman, L. (2001). Statistical modeling: The two cultures. Statistical Science, 16(3), 199-231.
- Crawford, K. (2021). Atlas of AI. Yale University Press.
- Gentzkow, M., Kelly, B. & Taddy, M. (2019). Text as data. Journal of Economic Literature, 57(3), 535-574.
- Grimmer, J., Roberts, M. E. & Stewart, B. M. (2022). Text as Data: A New Framework. Princeton University Press.
- Lazer, D. et al. (2020). Computational social science: Obstacles and opportunities. Science, 369(6507), 1060-1062.
- Molina, M. & Garip, F. (2019). Machine learning for sociology. Annual Review of Sociology, 45, 27-45.
- Salganik, M. J. (2018). Bit by Bit: Social Research in the Digital Age. Princeton University Press.
- Ziems, C. et al. (2024). Can large language models transform computational social science? Computational Linguistics, 50(1), 237-291.
- Benoit, K. et al. (2016). Crowd-sourced text analysis: Reproducible and agile production of political data. American Political Science Review, 110(2), 278-295.
- Blei, D. M., Ng, A. Y. & Jordan, M. I. (2003). Latent Dirichlet Allocation. Journal of Machine Learning Research, 3, 993-1022.
- Hastie, T., Tibshirani, R. & Friedman, J. (2009). The Elements of Statistical Learning (2da edicion). Springer. Disponible en hastie.su.domains.
- Knox, D. & Lucas, C. (2021). A dynamic model of speech for the social sciences. American Political Science Review, 115(2), 649-666.