TY - JOUR
T1 - RNN-Stega
T2 - Linguistic Steganography Based on Recurrent Neural Networks
AU - Yang, Zhong-Liang
AU - Guo, Xiao-Qing
AU - Chen, Zi-Ming
AU - Huang, Yong-Feng
AU - Zhang, Yu-Jin
PY - 2019/5
Y1 - 2019/5
N2 - Linguistic steganography based on text carrier autogeneration technology is a current topic with great promise and unpredictable challenges. Limited by the text automatic generation technology or the corresponding text coding methods, the quality of the steganographic text generated by previous methods is inferior, which makes its imperceptibility unsatisfactory. In this paper, we propose a linguistic steganography based on Recurrent Neural Networks (RNN-Stega), which can automatically generate high-quality text covers on the basis of a secret bitstream that needs to be embedded. We trained our model with a large number of artificially generated samples and obtained a good estimate of the statistical language model. In the text generation process, we propose Fixed-Length Coding (FLC) and Variable-Length Coding (VLC) to encode words based on their conditional probability distribution. We designed several experiments to test the proposed model from the perspectives of information embedding efficiency, information imperceptibility and information hidden capacity. Experimental results show that the proposed model outperforms all the previous related methods and achieves the state of the art performance.
AB - Linguistic steganography based on text carrier autogeneration technology is a current topic with great promise and unpredictable challenges. Limited by the text automatic generation technology or the corresponding text coding methods, the quality of the steganographic text generated by previous methods is inferior, which makes its imperceptibility unsatisfactory. In this paper, we propose a linguistic steganography based on Recurrent Neural Networks (RNN-Stega), which can automatically generate high-quality text covers on the basis of a secret bitstream that needs to be embedded. We trained our model with a large number of artificially generated samples and obtained a good estimate of the statistical language model. In the text generation process, we propose Fixed-Length Coding (FLC) and Variable-Length Coding (VLC) to encode words based on their conditional probability distribution. We designed several experiments to test the proposed model from the perspectives of information embedding efficiency, information imperceptibility and information hidden capacity. Experimental results show that the proposed model outperforms all the previous related methods and achieves the state of the art performance.
KW - automatic text generation
KW - Linguistic steganography
KW - recurrent neural network
UR - http://www.scopus.com/inward/record.url?scp=85054234170&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-85054234170&origin=recordpage
U2 - 10.1109/TIFS.2018.2871746
DO - 10.1109/TIFS.2018.2871746
M3 - RGC 21 - Publication in refereed journal
AN - SCOPUS:85054234170
SN - 1556-6013
VL - 14
SP - 1280
EP - 1295
JO - IEEE Transactions on Information Forensics and Security
JF - IEEE Transactions on Information Forensics and Security
IS - 5
ER -