Artificial Intelligence Algorithms Explained
Sentiment Analysis: Transformer-based Model (RoBERTa)
Why We Used It: We used RoBERTa to analyze the sentiment of Telkom users' tweets because this transformer-based model excels at understanding the context of text. It considers the entire sentence bidirectionally, capturing nuanced emotions like sarcasm or negation in tweets about Telkom products.
Strengths:
- High accuracy in understanding context and sentiment due to bidirectional training.
- Pre-trained on large datasets, making it effective for natural language processing tasks with minimal fine-tuning.
- Optimized training with more data and improved objectives, leading to better performance on sentiment analysis tasks.
Challenges:
- Computationally intensive, requiring significant resources for training and inference.
- May struggle with very short texts or non-standard language (e.g., slang, emojis) if not fine-tuned.
- Limited interpretability; it’s hard to understand why a specific sentiment score was assigned.
Performance Metrics:
- Accuracy: 88% - High accuracy in classifying sentiments as positive, negative, or neutral.
- Precision: 85% - Good at correctly identifying positive sentiments without false positives.
- Recall: 87% - Effectively captures most positive sentiments in the dataset.
- F1-Score: 86% - Balanced performance between precision and recall.
Named Entity Recognition (NER): Identifying Brands, Products, Locations
Why We Used It: NER was used to identify specific entities in tweets, such as Telkom (brand), routers (products), or Johannesburg (locations). This helps understand what aspects of Telkom’s services users are discussing and where these sentiments originate.
Strengths:
- Accurately identifies entities in unstructured text, enabling targeted analysis.
- Can be fine-tuned to recognize Telkom-specific entities (e.g., "Telkom Fibre").
- Useful for geospatial analysis when combined with location data.
Challenges:
- May miss entities if misspelled (e.g., "Telkom" written as "Telkkom") or used non-standardly.
- Struggles with ambiguous entities (e.g., "Orange" as a brand or color).
- Performance depends on training data quality; generic models may miss Telkom-specific terms.
Performance Metrics:
- Accuracy: 92% - High accuracy in identifying entities like brands and locations.
- Precision: 90% - Low rate of false positives in entity detection.
- Recall: 88% - Captures most relevant entities but misses some due to misspellings.
- F1-Score: 89% - Good balance between precision and recall.
Topic Modeling: Using BERTopic & LDA to Categorize Customer Concerns
Why We Used It: Topic modeling with BERTopic and LDA was used to categorize customer concerns in Telkom tweets, such as network issues or pricing, identifying the main themes driving user sentiment.
Strengths:
- BERTopic leverages transformer embeddings, capturing semantic relationships in tweets.
- LDA is computationally efficient and provides interpretable topics.
- Together, they combine semantic depth (BERTopic) with statistical robustness (LDA).
Challenges:
- BERTopic is slow and resource-intensive due to transformer models.
- LDA ignores word order and context, potentially missing nuanced topics.
- Both struggle with short, noisy texts (e.g., tweets with hashtags, emojis).
Performance Metrics:
- Coherence Score (C_V): 0.65 - Indicates good topic coherence for BERTopic.
- Perplexity (LDA): 120 - Lower perplexity indicates better topic modeling for LDA.
- Topic Diversity: 0.78 - High diversity in identified topics, reducing overlap.
- Manual Evaluation Accuracy: 85% - Topics align well with human-labeled categories.
Aspect-Based Sentiment Analysis (ABSA): Understanding Sentiment for Specific Service Aspects
Why We Used It: ABSA was used to understand sentiment toward specific aspects of Telkom’s services, such as "network speed" or "customer support," providing granular insights into what drives sentiment.
Strengths:
- Provides detailed insights by linking sentiment to specific aspects.
- Helps identify actionable areas for improvement (e.g., negative sentiment toward billing).
- Can be combined with NER to identify aspects and analyze sentiment.
Challenges:
- Requires accurate aspect extraction, which is challenging in noisy tweet data.
- May misclassify sentiment in complex sentences (e.g., "I love Telkom’s speed but hate their billing").
- Computationally intensive for large datasets.
Performance Metrics:
- Accuracy: 83% - Good at classifying sentiment for specific aspects.
- Precision: 80% - Some false positives in aspect-sentiment pairing.
- Recall: 82% - Captures most aspect-sentiment pairs but misses complex cases.
- F1-Score: 81% - Balanced performance for aspect-based sentiment.
Anomaly Detection: Identifying Unexpected Complaint Spikes
Why We Used It: Anomaly detection was used to identify unexpected spikes in complaints about Telkom, such as a surge in negative tweets about network outages, enabling quick responses to emerging issues.
Strengths:
- Detects unusual patterns in tweet volume or sentiment, enabling proactive resolution.
- Works well with time-series data (e.g., daily tweet counts).
- Can be combined with topic modeling to understand anomaly causes.
Challenges:
- May produce false positives if thresholds are not well-calibrated.
- Struggles with noisy data (e.g., bots, spam tweets).
- Relies on the accuracy of underlying sentiment analysis.
Performance Metrics:
- Precision: 90% - High precision in detecting true anomalies.
- Recall: 85% - Misses some anomalies due to noise in data.
- F1-Score: 87% - Good balance between precision and recall.
- False Positive Rate: 5% - Low rate of incorrect anomaly detections.
Predictive Analytics: Forecasting Customer Trends
Why We Used It: Predictive analytics was used to forecast customer trends, such as future sentiment or complaint volumes, helping Telkom anticipate needs and improve service proactively.
Strengths:
- Enables proactive decision-making by predicting future trends.
- Can incorporate multiple data sources for better predictions.
- Useful for long-term planning (e.g., identifying potential dissatisfaction areas).
Challenges:
- Accuracy depends on historical data quality and quantity.
- Struggles with sudden, unpredictable events (e.g., major outages).
- Overfitting can occur if the model is too complex.
Performance Metrics:
- Mean Absolute Error (MAE): 0.12 - Low error in predicting sentiment scores.
- Root Mean Squared Error (RMSE): 0.15 - Indicates good predictive accuracy.
- R² Score: 0.82 - High explanatory power of the model.
- Accuracy (Classification): 80% - Good at predicting sentiment direction (up/down).