Image Captioning menurut Scientific Revolution Kuhn dan Popper
Image captioning is one area in artificial intelligence that elaborates between computer vision and natural language processing. The focus on this process is an architecture neural network that includes many layers to solve the identification object on the image and give the caption. This architecture has a task to display the caption from object detection on one image. This paper explains about the connection between scientific revolution and image captioning. We have conducted the methodology by Kuhn's scientific revolution and relate to Popper's philosophy of science. The result of this paper is that an image captioning is truly science because many improvements from many researchers to find an effective method on the deep learning process. On the philosophy of science, if the phenomena can be falsified, then an image captioning is the science.
R. K. Srihari, â€œUse of Captions and Other Collateral Text in Understanding Photos,â€ 1994, pp. 1â€“32.
J. Gu, G. Wang, J. Cai, and T. Chen, â€œAn Empirical Study of Language CNN for Image Captioning,â€ in IEEE International Conference on Computer Vision An, 2017, pp. 1231â€“1240.
Y. U. and T. H. Andrew Shin, â€œImage Captioning with Sentiment Terms via Weakly-Supervised Sentiment Dataset,â€ in British Machine Vision Conference, 2016, p. 53.1-53.1.
J. Aneja and A. G. Schwing, â€œConvolutional Image Captioning.â€
B. Dai, â€œContrastive Learning for Image Captioning,â€ in Advances in Neural Information Processing Systems Conferece, 2017, no. 30, pp. 898â€“907.
K. Shuster, S. Humeau, H. Hu, A. Bordes, and J. Weston, â€œEngaging Image Captioning via Personality,â€ in IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 12516â€“12526.
B. Dai and S. Fidler, â€œA Neural Compositional Paradigm for Image Captioning,â€ no. NeurIPS, pp. 1â€“11, 2018.
K. Papineni, S. Roukos, T. Ward, and Z. Wei-Jing, â€œBLEU: a Method for Automatic Evaluation of Machine Translation,â€ in Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL), 2002, pp. 311â€“318.
J. Devlin et al., â€œLanguage Models for Image Captioning : The Quirks and What Works,â€ Comput. Lang., pp. 100â€“105, 2015.
V. Kougia, J. Pavlopoulos, and I. Androutsopoulos, â€œA Survey on Biomedical Image Captioning,â€ 2016.
X. Li, X. Song, L. Herranz, Y. Zhu, and S. Jiang, â€œImage Captioning with both Object and Scene Information,â€ in 24th ACM international conference on Multimedia, 2016, pp. 1107â€“1110.
D. Shin and I. Kim, â€œDeep Image Understanding Using Multilayered Contexts,â€ Math. Probl. Eng., vol. 2018, pp. 1â€“11, 2018.
Z. Gan et al., â€œSemantic compositional networks for visual captioning,â€ Proc. - 30th IEEE Conf. Comput. Vis. Pattern Recognition, CVPR 2017, vol. 2017â€“Janua, pp. 1141â€“1150, 2017.
A. Karpathy and L. Fei-Fei, â€œDeep Visual-Semantic Alignments for Generating Image Descriptions,â€ IEEE Trans. Pattern Anal. Mach. Intell., vol. 39, no. 4, pp. 664â€“676, 2017.
X. He, B. Shi, X. Bai, G. Xia, and Z. Zhang, â€œImage Caption Generation with Part of Speech Guidance,â€ Pattern Recognit. Lett., vol. 0, pp. 1â€“9, 2017.
J. Mun, L. Yang, Z. Ren, N. Xu, B. Han, and A. Go, â€œStreamlined Dense Video Captioning.â€
J. Aneja and A. G. Schwing, â€œConvolutional Image Captioning,â€ Comput. Vis. Pattern Recognit., pp. 5561â€“5570, 2017.
G. Ding, M. Chen, S. Zhao, H. Chen, and J. Han, â€œNeural Image Caption Generation with Weighted Training and Reference,â€ Cognit. Comput., p. 10.1007, 2018.
C. Gan, Z. Gan, X. He, and J. Gao, â€œStyleNet : Generating Attractive Visual Captions with Styles,â€ in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 955â€“964.
S. Bai and S. An, â€œA survey on automatic image caption generation Shuang,â€ Neurocomputing, vol. 311, pp. 291â€“304, 2018.
T. Kuhn, The Structure of Scientific Revolutions, 4 (2012). University of Chicago Press, 1962.
D. Mahayana, Filsafat Ilmu Pengetahuan. Bandung, Indonesia: ITB Press, 2018.
J. Deng, W. Dong, R. Socher, L. Li, K. Li, and L. Fei-fei, â€œImageNet : A Large-Scale Hierarchical Image Database,â€ in IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2009, pp. 248â€“255.
â€œF. Serafino, G. Pio and M. Ceci, â€˜Ensemble Learning for Multi-Type Classification in Heterogeneous Networks,â€™ in,â€ vol. 30, no. 12, p. 8525379, 2018.
H. Kaiming, Z. Xiangyu, R. Shaoqing, and S. Jian, â€œDeep Residual Learning for Image Recognition,â€ Comput. Vis., pp. 1â€“9, 2016.
A. Krizhevsky, I. Sutskever, and G. E. Hinton, â€œImageNet Classification with Deep Convolutional Neural Networks,â€ Adv. Neural Inf. Process. Syst. 25 (NIPS 2012), vol. 25, pp. 1â€“9, 2012.
Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, â€œGradient-Based Learning Applied to Document Recognition,â€ in IEEE, 1998.
K. Simonyan and A. Zisserman, â€œVery Deep Convolutional Networks for Large-Scale Image Recognition,â€ arxiv, pp. 1â€“14, 2015.
J. Donahue et al., â€œLong-term Recurrent Convolutional Networks for Visual Recognition and Description,â€ pp. 1â€“14, 2015.
V. Kougia, J. Pavlopoulos, and I. Androutsopoulos, â€œA Survey on Biomedical Image Captioning,â€ in Proceedings of the Second Workshop on Shortcomings in Vision and Langua, 2016, pp. 26â€“36.
H. Agrawal et al., â€œnocaps : novel object captioning at scale,â€ in IEEE International Conference on Computer Vision (ICCV) 2019, 2019.
V. Batra, Y. He, and G. Vogiatzis, â€œNeural Caption Generation for News Images,â€ pp. 1726â€“1733, 2016.