• Nenhum resultado encontrado

• Aplicar outros testes estatísticos às variações prequential.

• Estender a análise da metodologia abordada, utilizando outras métricas de avaliação como o tempo e consumo de memória.

• Investigar quais seriam os melhores parâmetros a utilizar em cada uma das variações prequential para obter resultados que sejam mais próximos da acurácia real. • Experimentar o desempenho dos enfoques prequential com outros conjuntos de da-

69

REFERÊNCIAS

AGRAWAL, R.; IMIELINSKI, T.; SWAMI, A. Database mining: A performance perspective. IEEE transactions on knowledge and data engineering, IEEE, v. 5, n. 6, p. 914–925, 1993.

ATTAR, V.; CHAUDHARY, P.; RAHAGUDE, S.; CHAUDHARI, G.; SINHA, P. An instance-window based classification algorithm for handling gradual concept drifts. In: CAO, L.; BAZZAN, A. L. C.; SYMEONIDIS, A. L.; GORODETSKY, V. I.; WEISS, G.; YU, P. S. (Ed.). Agents and Data Mining Interaction: 7th International Workshop on

Agents and Data Mining Interation, ADMI 2011, Taipei, Taiwan, May 2-6, 2011, Revised Selected Papers. Berlin, Heidelberg: Springer Berlin Heidelberg, 2012. p. 156–172. ISBN

978-3-642-27609-5. Disponível em: <http://dx.doi.org/10.1007/978-3-642-27609-5_11>. BAENA-GARCIA, M.; CAMPO-ÁVILA, J. D.; FIDALGO, R.; BIFET, A.; GAVALDà, R.; MORALES-BUENO, R. Early drift detection method. In: Int. Workshop on

Knowl. Discovery from Data Streams. [S.l.: s.n.], 2006. p. 77–86.

BARROS, R. S. M. Advances in Data Stream Mining with Concept Drift. 2017. Professorship (Full) Thesis. Centro de Informática, Universidade Federal de Pernambuco, Brazil.

BARROS, R. S. M.; SANTOS, S. G. T. C.; GONÇALVES JR., P. M. A boosting-like online learning ensemble. In: Proceedings of IEEE International Joint Conference on

Neural Networks (IJCNN). Vancouver, Canada: [s.n.], 2016. p. 1871–1878.

BERNSTEIN, S. The theory of Probabilities. 1946. Gastehizdat Publishing House, Moscow.

BIFET, A. Adaptive stream mining: Pattern learning and mining from evolving data

streams. [S.l.]: IOS Press, 2010.

BIFET, A.; FRANK, E. Sentiment knowledge discovery in twitter streaming data. In: SPRINGER. International Conference on Discovery Science. [S.l.], 2010. p. 1–15. BIFET, A.; FRANK, E. Sentiment knowledge discovery in twitter streaming data. In: SPRINGER. International Conference on Discovery Science. [S.l.], 2010. p. 1–15.

BIFET, A.; GAVALDA, R. Learning from time-changing data with adaptive windowing. In: Proceedings of the 7th SIAM International Conference on Data Mining (SDM’07). Minneapolis, MN, USA: [s.n.], 2007. p. 443–448.

BIFET, A.; HOLMES, G.; KIRKBY, R.; PFAHRINGER, B. Moa: Massive online analysis. Journal of Machine Learning Research, v. 11, n. May, p. 1601–1604, 2010. BIFET, A.; HOLMES, G.; PFAHRINGER, B.; KIRKBY, R.; GAVALDÀ, R. New ensemble methods for evolving data streams. In: ACM. Proceedings of the 15th ACM

SIGKDD international conference on Knowledge discovery and data mining. [S.l.], 2009.

REFERÊNCIAS 70

BIFET, A.; MORALES, G. de F.; READ, J.; HOLMES, G.; PFAHRINGER, B. Efficient online evaluation of big data stream classifiers. In: ACM. Proceedings of the 21th ACM

SIGKDD International Conference on Knowledge Discovery and Data Mining. [S.l.],

2015. p. 59–68.

BLUMAN, A. Elementary Statistics: A Step by Step Approach. Ninth. New York, USA: McGraw-Hill, 2014.

BRZEZIŃSKI, D. Mining data streams with concept drift. Dissertação (Mestrado) — Master’s thesis, Poznan University of Technology, 2010.

BRZEZINSKI, D.; STEFANOWSKI, J. Prequential auc for classifier evaluation and drift detection in evolving data streams. In: SPRINGER. International Workshop on New

Frontiers in Mining Complex Patterns. [S.l.], 2014. p. 87–101.

DAWID, A.; VOVK, V. Prequential probability: Principles and properties. Bernoulli, v. 5, n. 1, p. 125–162, 1999.

DAWID, A. P. Present position and potential developments: Some personal views: Statistical theory: The prequential approach. Journal of the Royal Statistical Society.

Series A (General), v. 147, n. 2, p. 278–292, 1984.

DEMŠAR, J. Statistical comparisons of classifiers over multiple data sets. Journal of

Machine learning research, v. 7, n. Jan, p. 1–30, 2006.

DOMINGOS, P.; HULTEN, G. Mining high-speed data streams. In: ACM. Proceedings

of the sixth ACM SIGKDD international conference on Knowledge discovery and data mining. [S.l.], 2000. p. 71–80.

DU, L.; SONG, Q.; JIA, X. Detecting concept drift: an information entropy based method using an adaptive sliding window. Intelligent Data Analysis, IOS Press, v. 18, n. 3, p. 337–364, 2014.

FACELI, K.; LORENA, A. C.; GAMA, J.; CARVALHO, A. Inteligência artificial: Uma abordagem de aprendizado de máquina. Rio de Janeiro: LTC, v. 2, p. 192, 2011.

FRÍAS-BLANCO, I. Nuevos métodos para el aprendizaje en flujos de datos no

estacionarios. Tese (Doutorado) — Universidad de Granma, Cuba, 2014.

FRÍAS-BLANCO, I.; CAMPO-ÁVILA, J. del; RAMOS-JIMÉNEZ, G.; MORALES- BUENO, R.; ORTIZ-DÍAZ, A.; CABALLERO-MOTA, Y. Online and non-parametric drift detection methods based on hoeffding’s bounds. IEEE Transactions on Knowledge

and Data Engineering, v. 27, n. 3, p. 810–823, 2015.

FRIEDMAN, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. Journal of the american statistical association, Taylor & Francis, v. 32, n. 200, p. 675–701, 1937.

GAMA, J.; MEDAS, P.; CASTILLO, G.; RODRIGUES, P. Learning with drift detection. In: SPRINGER. Brazilian Symposium on Artificial Intelligence. [S.l.], 2004. p. 286–295. GAMA, J.; SEBASTIÃO, R.; RODRIGUES, P. On evaluating stream learning algorithms. Machine Learning, v. 90, n. 3, p. 317–346, 2013.

REFERÊNCIAS 71

GAMA, J.; SEBASTIÃO, R.; RODRIGUES, P. P. Issues in evaluation of stream learning algorithms. In: ACM. Proceedings of the 15th ACM SIGKDD international conference

on Knowledge discovery and data mining. [S.l.], 2009. p. 329–338.

GAMA, J.; ŽLIOBAIT ˙E, I.; BIFET, A.; PECHENIZKIY, M.; BOUCHACHIA, A. A survey on concept drift adaptation. ACM Computing Surveys (CSUR), ACM, v. 46, n. 4, p. 44, 2014.

GOMES, H. M.; BARDDAL, J. P.; ENEMBRECK, F.; BIFET, A. A survey on ensemble learning for data stream classification. ACM Computing Surveys (CSUR), ACM, v. 50, n. 2, p. 23, 2017.

GONÇALVES JR., P. M.; BARROS, R. S. M. Rcd: A recurring concept drift framework.

Pattern Recognition Letters, Elsevier, v. 34, n. 9, p. 1018–1025, 2013.

HAN, J.; PEI, J.; KAMBER, M. Data mining: concepts and techniques. [S.l.]: Elsevier, 2011.

HOEFFDING, W. Probability inequalities for sums of bounded random variables.

Journal of the American Statistical Association, v. 58, p. 13–30, 1963.

HULTEN, G.; SPENCER, L.; DOMINGOS, P. Mining time-changing data streams. In: ACM. Proceedings of the seventh ACM SIGKDD international conference on Knowledge

discovery and data mining. [S.l.], 2001. p. 97–106.

JOHN, G. H.; LANGLEY, P. Estimating continuous distributions in bayesian classifiers. In: MORGAN KAUFMANN PUBLISHERS INC. Proceedings of the Eleventh conference

on Uncertainty in artificial intelligence. [S.l.], 1995. p. 338–345.

JOHNSON, O. Information Theory and the Central Limit Theorem. [S.l.]: Imperial College Press, London, 2004.

LARSON, R.; FARBER, B. Elementary Statistics: Picturing the World. Fourth. [S.l.]: Pearson, 2010.

LEMAIRE, V.; SALPERWYCK, C.; BONDU, A. A survey on supervised classification on data streams. In: Business Intelligence. [S.l.]: Springer, 2015. p. 88–125.

MACIEL, B.; SANTOS, S.; BARROS, R. A lightweight concept drift detection ensemble. In: Proc. 27th IEEE Intern. Conf. on Tools with Artificial Intelligence (ICTAI). Vietri sul Mare, Italy: [s.n.], 2015. p. 1061–1068.

MARON, O.; MOORE, A. W. Hoeffding races: Accelerating model selection search for classification and function approximation. Robotics Institute, p. 263, 1993.

MARRÓN, D.; READ, J.; BIFET, A.; ABDESSALEM, T.; AYGUADÉ, E.; HERRERO, J. R. Echo state hoeffding tree learning. In: Proceedings of The 8th Asian Conference on

Machine Learning. [S.l.: s.n.], 2016. p. 382–397.

MASUD, M.; GAO, J.; KHAN, L.; HAN, J.; THURAISINGHAM, B. M. Classification and novel class detection in concept-drifting data streams under time constraints. IEEE

REFERÊNCIAS 72

MCDIARMID, C. On the method of bounded differences. Surveys in combinatorics, v. 141, n. 1, p. 148–188, 1989.

MINKU, L. L.; WHITE, A. P.; YAO, X. The impact of diversity on online ensemble learning in the presence of concept drift. IEEE Transactions on Knowledge and Data

Engineering, IEEE, v. 22, n. 5, p. 730–742, 2010.

MINKU, L. L.; YAO, X. Ddd: A new ensemble approach for dealing with concept drift.

IEEE transactions on knowledge and data engineering, IEEE, v. 24, n. 4, p. 619–633,

2012.

MITCHELL, T. M. Machine Learning. 1. ed. New York, NY, USA: McGraw-Hill, Inc., 1997.

MONTGOMERY, D. C.; RUNGER, G. C. Applied statistics and probability for

engineers. [S.l.]: John Wiley & Sons, 2010.

NEMENYI, P. Distribution-free multiple comparisons. Princeton University, 1963. NISHIDA, K.; YAMAUCHI, K. Detecting concept drift using statistical testing. In:

Proceedings of 10th International Conf. on Discovery Science (DS’07). [S.l.: s.n.], 2007.

(LNCS, v. 4755), p. 264–269.

PEARS, R.; SAKTHITHASAN, S.; KOH, Y. Detecting concept change in dynamic data streams. Machine Learning, v. 97, n. 3, p. 259–293, 2014.

PESARANGHADER, A.; VIKTOR, H. Fast hoeffding drift detection method for evolving data streams. In: Machine Learning and Knowledge Discovery in Databases. [S.l.]: Springer, 2016. (LNCS, v. 9852), p. 96–111.

RAFTER, J. A.; ABELL, M. L.; BRASELTON, J. P. Multiple comparison methods for means. Siam Review, SIAM, v. 44, n. 2, p. 259–278, 2002.

ROBERTS, S. W. Control chart tests based on geometric moving averages.

Technometrics, v. 1, n. 3, p. 239–250, 1959.

ROSS, G.; ADAMS, N.; TASOULIS, D.; HAND, D. Exponentially weighted moving average charts for detecting concept drift. Pattern Recognition Letters, v. 33, n. 2, p. 191–198, 2012.

SAKTHITHASAN, S.; PEARS, R.; KOH, Y. One pass concept change detection for data streams. In: Advances in Knowledge Discov. and Data Mining. [S.l.]: Springer, 2013, (LNCS, v. 7819). p. 461–472.

SANTOS, S.; JR., P. G.; SILVA, G.; BARROS, R. Speeding up recovery from concept drifts. In: Machine Learning and Knowl. Discovery in Databases. [S.l.]: Springer, 2014, (LNCS, v. 8726). p. 179–194.

SANTOS, S. G. T. C.; BARROS, R. S. M.; GONÇALVES JR., P. M. Optimizing the parameters of drift detection methods using a genetic algorithm. In: Proceedings of The

27th IEEE International Conference on Tools with Artificial Intelligence (ICTAI’15).

REFERÊNCIAS 73

VITTER, J. S. Random sampling with a reservoir. ACM Transactions on Mathematical

Software (TOMS), ACM, v. 11, n. 1, p. 37–57, 1985.

WALPOLE, R.; MYERS, S. M. R.; YE, K. Probabilidad y estadística para ingeniería y

ciencias. 9. ed. [S.l.]: Pearson, 2012. ISBN 978 -607-32-1417-9.

WIDMER, G.; KUBAT, M. Effective learning in dynamic environments by explicit context tracking. In: SPRINGER. Machine learning: ECML-93. [S.l.], 1993. p. 227–243. ŽLIOBAIT ˙E, I.; BIFET, A.; READ, J.; PFAHRINGER, B.; HOLMES, G. Evaluation methods and decision theory for classification of streaming data with temporal dependence. Machine Learning, Springer, v. 98, n. 3, p. 455–482, 2015.

74

APÊNDICE A – PROVAS DE

CONVERGÊNCIA DO ERRO PREQUENTIAL

A seguir, são apresentados os teoremas que demonstram a convergência do erro Prequen- tial para algoritmos de aprendizagem consistentes utilizando as variações da metodologia Prequential.

A.1 Limite do erro Prequential utilizando uma janela básica (Basic

Documentos relacionados