DSpace Repository

Vector Representation of Bengali Word Using Various Word Embedding Model

Show simple item record

dc.contributor.author Rafat, Ashik Ahamed Aman
dc.contributor.author Salehin, Mushfiqus
dc.contributor.author Khan, Fazle Rabby
dc.contributor.author Hossain, Syed Akhter
dc.contributor.author Abujar, Sheikh
dc.date.accessioned 2021-11-29T04:18:34Z
dc.date.available 2021-11-29T04:18:34Z
dc.date.issued 2020
dc.identifier.uri http://dspace.daffodilvarsity.edu.bd:8080/handle/123456789/6496
dc.description.abstract To transfer human understanding of language to a machine we need word embedding. Skipgram, CBOW, and fastText is a model which generate word embedding. But finding pretrained word embedding model for the Bengali language is difficult for researchers. Also, training word embedding is time-consuming. In this paper, we discussed different word embedding models. To train those models, we have collected around 500000 Bengali articles from various sources on the internet. Among them, we randomly chose 105000 articles. Those articles have 32 million words. We trained them on SkipGram and CBOW model of Word2Vec, fastText. We also trained those words in Glove model. Among the all result fastText (Word2Vec) gave us a satisfactory result. en_US
dc.language.iso en_US en_US
dc.publisher Scopus en_US
dc.subject Bengali Words en_US
dc.subject Skip Gram en_US
dc.subject CBOW en_US
dc.subject Word2Vec en_US
dc.subject FastText en_US
dc.subject Glove en_US
dc.subject Word Embedding en_US
dc.title Vector Representation of Bengali Word Using Various Word Embedding Model en_US
dc.type Article en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Browse

My Account

Statistics