The PDF file you selected should load here if your Web browser has a PDF reader plug-in installed (for example, a recent version of Adobe Acrobat Reader).
If you would like more information about how to print, save, and work with PDFs, Highwire Press provides a helpful Frequently Asked Questions about PDFs.
Alternatively, you can download the PDF file directly to your computer, from where it can be opened using a PDF reader. To download the PDF, click the Download link above.
BibTex Citation Data :
@article{JOIV709, author = {Dennis Setiawan and Maria Astrid Saffachrissa and Shintia Tamara and Derwin Suhartono}, title = {Image Captioning with Style Using Generative Adversarial Networks}, journal = {JOIV : International Journal on Informatics Visualization}, volume = {6}, number = {1}, year = {2022}, keywords = {Stylized image captioning; SeqCapsGAN; sentiments or styles; Generative Adversarial Network (GAN); capsule; discriminator; generator.}, abstract = {Image captioning research, which initially focused on describing images factually, is currently being developed in the direction of incorporating sentiments or styles to produce natural captions that reflect human-generated captions. The problem this research tries to solve the problem that captions produced by existing models are rigid and unnatural due to the lack of sentiment. The purpose of this research is to design a reliable image captioning model that incorporates style based on state-of-the-art SeqCapsGAN architecture. The materials needed are MS COCO and SentiCaps datasets. Research methods are done through literature studies and experiments. While many previous studies compare their works without considering the differences in components and parameters being used, this research proposes a different approach to find more reliable configurations and provide more detailed insights into models’ behavior. This research also does further experiments on the generator part that have not been thoroughly investigated. Experiments are done on the combinations of feature extractor (VGG-19 and ResNet-50), discriminator model (CNN and Capsule), optimizer (Adam, Nadam, and SGD), batch size (8, 16, 32, and 64), and learning rate (0.001 and 0.0001) by doing a grid search. In conclusion, more insights into the models’ behavior can be drawn, and better configuration and result than the baseline can be achieved. Our research implies that research in comparative studies of image recognition models in image captioning context, automated metrics, and larger datasets suited for stylized image captioning might be needed for furthering the research in this field.}, issn = {2549-9904}, pages = {26--32}, doi = {10.30630/joiv.6.1.709}, url = {http://joiv.org/index.php/joiv/article/view/709} }
Refworks Citation Data :
@article{{JOIV}{709}, author = {Setiawan, D., Saffachrissa, M., Tamara, S., Suhartono, D.}, title = {Image Captioning with Style Using Generative Adversarial Networks}, journal = {JOIV : International Journal on Informatics Visualization}, volume = {6}, number = {1}, year = {2022}, doi = {10.30630/joiv.6.1.709}, url = {} }Refbacks
- There are currently no refbacks.

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
__________________________________________________________________________
JOIV : International Journal on Informatics Visualization
ISSN 2549-9610 (print) | 2549-9904 (online)
Organized by Department of Information Technology - Politeknik Negeri Padang, and Institute of Visual Informatics - UKM and Soft Computing and Data Mining Centre - UTHM
W : http://joiv.org
E : joiv@pnp.ac.id, hidra@pnp.ac.id, rahmat@pnp.ac.id
View JOIV Stats
is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.