Story Generator Bahasa Indonesia dengan Skip-Thoughts

M Mustofa(1*), Dhomas Hatta Fudholi(2),

(1) Universitas Islam Indonesia
(2) Universitas Islam Indonesia
(*) Corresponding Author

Abstract


Currently, there are many studies that want computers to be able to imitate human creativity in stringing words into writing like a writer. This study aims to use the RNN algorithm to produce automatic story writing in Indonesian. The main contribution in this research is the creation and evaluation of the RNN algorithm based on the skip-thoughts model using an Indonesian language dataset. The skip-thoughts model consists of an encoder in the form of single GRU layer with 500 hidden units, and two decoders with single GRU layer each with 500 hidden units. The function of the encoder is to do the word mapping process from the input sentence, while the decoder predicts the sentence before (previous decoder) and the sentence after (next decoder) from the input sentence. The dataset used in the model training is in the form of stories in Indonesian with the genres of folklore and short stories. The model training process is run in 100 epochs, using the ADAM optimizer to get the optimal model. Based on the results of the assessment of respondents who have a background as writers, the folklore model shows a fairly good rating (average score of 65) for the S-P-O-K criteria, and a low rating for criteria of linkage between sentences (average score of 38) and the context of the whole story (average score of 32). The short story of life model shows a good rating (average score of 73) for the S-P-O-K criteria, and a low rating for the linkage between sentences criteria (average score of 48), and the context of the whole story (average score of 42). Based on the results of the assessment, the skip-thoughts model used in the Indonesian story generator has worked well, but it can still be improved by increasing the number of training datasets for each story genre used, as well as being more specific in determining the genre in order to obtain story integrity better.

Full Text:

PDF

References


N. G. Polson and V. O. Sokolov, “Deep Learning,” International Series in Operations Research and Management Science, Jul. 2018, doi: 10.1007/978-3-319-68837-4_17.

J. Brownlee, “What is Deep Learning,” 2019. https://machinelearningmastery.com/what-is-deep-learning/ (accessed Jan. 10, 2020).

S. T. Ounpraseuth, “Gaussian Processes for Machine Learning,” J Am Stat Assoc, 2009, doi: 10.1198/jasa.2008.s219.

A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” 2012.

A. L. Caterini and D. E. Chang, “Recurrent neural networks,” in SpringerBriefs in Computer Science, 2018. doi: 10.1007/978-3-319-75304-1_5.

S. Hochreiter and J. Schmidhuber, “Long Short-Term Memory,” Neural Computation, 1997, doi: 10.1162/neco.1997.9.8.1735.

J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evaluation of gated recurrent neural networks on sequence modeling,” 2014.

R. Kiros et al., “Skip-Thought Vectors,” Advances in Neural Information Processing Systems, Jun. 2015, [Online]. Available: http://arxiv.org/abs/1506.06726

K. Cho et al., “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” 2014.

S. Agarwal, “My Thoughts on Skip-Thoughts,” 2018. https://medium.com/@sanyamagarwal/my-thoughts-on-skip-thoughts-a3e773605efa




DOI: http://dx.doi.org/10.30645/j-sakti.v6i2.479

Refbacks

  • There are currently no refbacks.



J-SAKTI (Jurnal Sains Komputer & Informatika)
Published Papers Indexed/Abstracted By:


Jumlah Kunjungan :

View My Stats