Deep Embeddings for Brand Detection in Product Titles

2019 
In this paper, we compare various techniques to learn expressive product title embeddings starting from TF-IDF and ending with deep neural architectures. The problem is to recognize brands from noisy retail product names coming from different sources such as receipts and supply documents. In this work we consider product titles written in English and Russian. To determine the state-of-the-art on openly accessed “Universe-HTT barcode reference” dataset, traditional machine learning models, such as SVMs, were compared to Neural Networks with classical softmax activation and cross entropy loss. Furthermore, the scalable variant of the problem was studied, where new brands are recognized without retraining the model. The approach is based on k-Nearest Neighbors, where the search space could be represented by either TF-IDF vectors or deep embeddings. For the latter we have considered two solutions: (1) pretrained FastText embeddings followed by LSTM with Attention and (2) character-level Convolutional Neural Network. Our research shows that deep embeddings significantly outperform TF-IDF vectors. Classification error was reduced from 13.2% for TF-IDF approach to 8.9% and to 7.3% for LSTM embeddings and character-level CNN embeddings correspondingly.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    0
    References
    0
    Citations
    NaN
    KQI
    []