参考文献一覧

Akiba, Takuya, Shotaro Sano, Toshihiko Yanase, Takeru Ohta, and Masanori Koyama. 2019. “Optuna: A Next-generation Hyperparameter Optimization Framework.” In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2623–31. Anchorage AK USA: ACM. https://doi.org/10.1145/3292500.3330701.
Albert, A., and J. A. Anderson. 1984. “On the Existence of Maximum Likelihood Estimates in Logistic Regression Models.” Biometrika 71 (1): 1–10. https://doi.org/10.1093/biomet/71.1.1.
Bengio, Yoshua, Nicolas Boulanger-Lewandowski, and Razvan Pascanu. 2013. “Advances in Optimizing Recurrent Networks.” In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, 8624–28. Vancouver, BC, Canada: IEEE. https://doi.org/10.1109/ICASSP.2013.6639349.
Bergstra, James, Rémi Bardenet, Yoshua Bengio, and Balázs Kégl. 2011. “Algorithms for Hyper-Parameter Optimization.” In Proceedings of the 24th International Conference on Neural Information Processing Systems, 2546–54. NIPS’11. Red Hook, NY, USA: Curran Associates Inc.
Bergstra, James, and Yoshua Bengio. 2012. “Random Search for Hyper-Parameter Optimization.” Journal of Machine Learning Research 13 (1): 281–305.
Breiman, Leo. 1996. “Bagging Predictors.” Machine Learning 24 (2): 123–40. https://doi.org/10.1023/A:1018054314350.
Breiman, Leo, and Philip Spector. 1992. “Submodel Selection and Evaluation in Regression. The X-Random Case.” International Statistical Review / Revue Internationale de Statistique 60 (3): 291. https://doi.org/10.2307/1403680.
Brier, Glenn W. 1950. “Verification of Forecasts Expressed in Terms of Probability.” Monthly Weather Review 78 (1): 1–3. https://doi.org/10.1175/1520-0493(1950)078<0001:VOFEIT>2.0.CO;2.
Brochu, Eric, Vlad M. Cora, and Nando de Freitas. 2010. “A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning.” arXiv:1012.2599 [Cs], December.
Buitinck, Lars, Gilles Louppe, Mathieu Blondel, Fabian Pedregosa, Andreas Mueller, Olivier Grisel, Vlad Niculae, et al. 2013. API Design for Machine Learning Software: Experiences from the Scikit-Learn Project.” In ECML PKDD Workshop: Languages for Data Mining and Machine Learning, 108–22.
Burbidge, John B., Lonnie Magee, and A. Leslie Robb. 1988. “Alternative Transformations to Handle Extreme Values of the Dependent Variable.” Journal of the American Statistical Association 83 (401): 123–27. https://doi.org/10.1080/01621459.1988.10478575.
Chapman, Peter, Janet Clinton, Randy Kerber, Tom Khabaza, Thomas Reinartz, C. Russell H. Shearer, and Robert Wirth. 2000. CRISP-DM 1.0: Step-by-step Data Mining Guide.” In.
Chen, Tianqi, and Carlos Guestrin. 2016. XGBoost: A Scalable Tree Boosting System.” In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD ’16, 785–94. New York, New York, USA: ACM Press. https://doi.org/10.1145/2939672.2939785.
Collett, David. 2003. Modelling Binary Data. 2. ed. Chapman & Hall/CRC Texts in Statistical Science Series. Boca Raton, Fla.: Chapman & Hall/CRC.
Davison, A. C., and E. J. Snell. 1991. “Residuals and Diagnostics.” In, edited by D. V. Hinkley, N. Reid, and E. J. Snell. Statistical Theory and Modelling : In Honour of Sir David Cox, FRS. Chapman and Hall.
Davison, A. C., and C.-L. Tsai. 1992. “Regression Model Diagnostics.” International Statistical Review / Revue Internationale de Statistique 60 (3): 337. https://doi.org/10.2307/1403682.
Defazio, Aaron, Francis Bach, and Simon Lacoste-Julien. 2014. SAGA: A Fast Incremental Gradient Method With Support for Non-Strongly Convex Composite Objectives.” Arxiv.
Duchi, John, Elad Hazan, and Yoram Singer. 2011. “Adaptive Subgradient Methods for Online Learning and Stochastic Optimization.” J. Mach. Learn. Res. 12 (July): 2121–59.
Efron, B. 1979. “Bootstrap Methods: Another Look at the Jackknife.” The Annals of Statistics 7 (1): 1–26. https://doi.org/10.1214/aos/1176344552.
Fan, Jianqing, and Runze Li. 2001. “Variable Selection via Nonconcave Penalized Likelihood and Its Oracle Properties.” Journal of the American Statistical Association 96 (456): 1348–60. https://doi.org/10.1198/016214501753382273.
Géron, Aurélien. 2017. Hands-on Machine Learning with Scikit-Learn and TensorFlow: Concepts, Tools, and Techniques to Build Intelligent Systems. O’Reilly Media, Inc.
Hastie, Trevor, Robert Tibshriani, and Jerome Friedman. 2009. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. Second. Springer.
Hayashi, Fumio. 2000. Econometrics. Princeton: Princeton University Press.
He, Xinran, Stuart Bowers, Joaquin Quiñonero Candela, Junfeng Pan, Ou Jin, Tianbing Xu, Bo Liu, et al. 2014. “Practical Lessons from Predicting Clicks on Ads at Facebook.” In Proceedings of 20th ACM SIGKDD Conference on Knowledge Discovery and Data Mining - ADKDD’14, 1–9. New York, NY, USA: ACM Press. https://doi.org/10.1145/2648584.2648589.
Healy, Kieran. 2018. Data Visualization: A Practical Introduction. Princeton, NJ: Princeton University Press.
Ke, Guolin, Qi Meng, Thomas Finley, Taifeng Wang, Wei Chen, Weidong Ma, Qiwei Ye, and Tie-Yan Liu. 2017. LightGBM: A Highly Efficient Gradient Boosting Decision Tree.” In Advances in Neural Information Processing Systems 30, edited by I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, 3146–54. Curran Associates, Inc.
Kingma, Diederik P., and Jimmy Ba. 2015. “Adam: A Method for Stochastic Optimization.” In International Conference on Learning Representations.
Lakshmanan, Valliappa, Sara Robinson, and Michael Munn. 2020. Machine Learning Design Patterns: Solutions to Common Challenges in Data Preparation, Model Building, and MLOps. First edition. Beijing Boston Farnham: O’Reilly.
Lefortier, Damien, Anthony Truchet, and Maarten de Rijke. 2015. “Sources of Variability in Large-Scale Machine Learning Systems.” In Machine Learning Systems (NIPS 2015 Workshop).
Matejka, Justin, and George Fitzmaurice. 2017. “Same Stats, Different Graphs: Generating Datasets with Varied Appearance and Identical Statistics Through Simulated Annealing.” In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, 1290–94. Denver Colorado USA: ACM. https://doi.org/10.1145/3025453.3025912.
Pierce, Donald A., and Daniel W. Schafer. 1986. “Residuals in Generalized Linear Models.” Journal of the American Statistical Association 81 (396): 977–86. https://doi.org/10.1080/01621459.1986.10478361.
Prokhorenkova, Liudmila, Gleb Gusev, Aleksandr Vorobev, Anna Veronika Dorogush, and Andrey Gulin. 2018. CatBoost: Unbiased Boosting with Categorical Features.” In Proceedings of the Neural Information Processing Systems 2018, edited by S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, 6639–49. Palais des Congrès de Montréal, Montréal Canada: Curran Associates, Inc.
Shadish, William R., Thomas D. Cook, and Donald T. Campbell. 2001. Experimental and Quasi-Experimental Designs for Generalized Causal Inference. Boston: Houghton Mifflin.
Shi, Qinfeng, James Petterson, Gideon Dror, John Langford, Alex Smola, Alex Strehl, and S. V. N. Vishwanathan. 2009. “Hash Kernels.” In Proceedings of the Twelth International Conference on Artificial Intelligence and Statistics, edited by David van Dyk and Max Welling, 5:496–503. Proceedings of Machine Learning Research. Hilton Clearwater Beach Resort, Clearwater Beach, Florida USA: PMLR.
Shi, Qinfeng, James Petterson, Gideon Dror, John Langford, Alex Smola, and S. V. N. Vishwanathan. 2009. “Hash Kernels for Structured Data.” Jornal of Machine Learning Research 10 (December): 2615–37.
Vehtari, Aki, Andrew Gelman, and Jonah Gabry. 2017. “Practical Bayesian Model Evaluation Using Leave-One-Out Cross-Validation and WAIC.” Statistics and Computing 27 (5): 1413–32. https://doi.org/10.1007/s11222-016-9696-4.
Watanabe, Sumio. 2010. “Asymptotic Equivalence of Bayes Cross Validation and Widely Applicable Information Criterion in Singular Learning Theory.” Journal of Machine Learning Research 11 (April): 3571–94.
Weinberger, Kilian, Anirban Dasgupta, John Langford, Alex Smola, and Josh Attenberg. 2009. “Feature Hashing for Large Scale Multitask Learning.” In Proceedings of the 26th Annual International Conference on Machine Learning - ICML ’09, 1–8. Montreal, Quebec, Canada: ACM Press. https://doi.org/10.1145/1553374.1553516.
Wickham, Hadley, and Garrett Grolemund. 2016. R for Data Science: Import, Tidy, Transform, Visualize, and Model Data. 1st ed. Sebastopol, CA: O’Reilly. https://r4ds.had.co.nz/.
Yeo, I.-K. 2000. “A New Family of Power Transformations to Improve Normality or Symmetry.” Biometrika 87 (4): 954–59. https://doi.org/10.1093/biomet/87.4.954.
Yi, Jeonghee, Ye Chen, Jie Li, Swaraj Sett, and Tak W Yan. 2013. “Predictive Model Performance: Offline and Online Evaluations.” In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD ’13, 1294. New York, New York, USA: ACM Press. https://doi.org/10.1145/2487575.2488215.
Zeiler, Matthew D. 2012. ADADELTA: An Adaptive Learning Rate Method.” Arxiv.
Zheng, Alice, and Amanda Casari. 2018. Feature Engineering for Machine Learning: Principles and Techniques for Data Scientists. 1st ed. O’Reilly Media, Inc.
Zhou, Zhi-Hua. 2012. Ensemble Methods: Foundations and Algorithms. CRC Press.
Zou, Hui, and Trevor Hastie. 2005. “Regularization and Variable Selection via the Elastic Net.” Journal of the Royal Statistical Society: Series B (Statistical Methodology) 67 (2): 301–20. https://doi.org/10.1111/j.1467-9868.2005.00503.x.
Нестеров, Юрый Е. 1983. “Метод Решения Задачи Выпуклого Программирования Со Скоростью Сходимости \(O\bigl(\frac1{k^2}\bigr)\).” Докл. АН СССР 269 (3): 543–47.
中川裕志. 2015. 機械学習. Edited by 東京大学工学教程編纂委員会. 東京大学工学教程 情報工学. 丸善出版.
久保拓哉. 2012. データ解析のための統計モデリング入門. 岩波書店.
伊庭幸人, ed. 2018. ベイズモデリングの世界. 岩波書店.
大石進一, 荻田武史, 柏木雅英, 劉雪峰, 尾崎克久, 山中脩也, 高安亮紀, et al. 2018. 精度保証付き数値計算の基礎. 東京: コロナ社.
富岡亮太. 2015. スパース性に基づく機械学習. 機械学習プロフェッショナルシリーズ. 東京: 講談社.
川野秀一, 松井秀俊, and 廣瀬慧, eds. 2018. スパース推定による統計モデリング. 6. 共立出版.
杉山将. 2013. イラストで学ぶ機械学習: 最小二乗法による識別モデル学習を中心に. 東京: 講談社.
杉山将, 山田誠, ドゥ・プレシマーティヌス・クリストフェル, and リウソン. 2014. 非定常環境下での学習:共変量シフト適応,クラスバランス変化適応,変化検知.” 日本統計学会誌 44 (1).
沖本竜義. 2010. 経済・ファイナンスデータの計量時系列分析. 朝倉書店.
海野裕也, 岡野原大輔, and 得居誠也. 2015. オンライン機械学習. 機械学習プロフェッショナルシリーズ. 東京: 講談社.
赤穂昭太郎. 2018. “ガウス過程回帰の基礎.” システム/制御/情報 62 (10): 390–95. https://doi.org/10.11509/isciesci.62.10_390.