Sentiment classification of user reviews for mobile games that rely on direct advertising (direct ads) is crucial for understanding player perceptions and improving user experience. This study aims to compare the performance of two deep learning architectures, Long Short-Term Memory (LSTM) and multilingual Bidirectional Encoder Representations from Transformers (BERT) in classifying sentiment in reviews into three categories, positive, negative, and neutral. The dataset used consists of reviews from games employing direct ads, which underwent rule-based labeling and text preprocessing. The LSTM model was built from scratch using a custom embedding layer, while the multilingual BERT model was fine-tuned using a transfer learning approach. Evaluation was conducted based on accuracy, precision, recall, and F1-score metrics. Experimental results show that multilingual BERT achieves superior validation loss compared to LSTM (0.37 vs. 0.44). BERT also outperforms LSTM significantly in terms of F1-score and its ability to understand multilingual linguistic context. However, LSTM demonstrates advantages in computational efficiency and training speed. These findings offer practical recommendations for developers in selecting an appropriate sentiment analysis model based on accuracy requirements and resource availability.