Research Improvement Consistency Validation
Consistency in improvement is validated in the following ways:
Checking for optimal values of W1 and W2 :
This step ensures the optimality of our pipeline.
Checking for different seeds with optimal (W1, W2) Values Received:
This step ensures the consistency because ideally our results should not get affected by varying values of seeds
EXPERIMENT using APD:
We have taken the best result obtained for APD. Since our RMSE already showed vast improvement, we prioritized best-NDCG improvement test case to run this experiment. We ran the experiment with the condition : W1+W2=1.0 and iterating W1: [0.0,1.0](with step size 0.1)
Best_Case: NDCG
(
Social Relationship Descriptor= All;
Group Decision Strategies= Max Satisfaction, Avg Satisfaction, Minimum Misery;
Expertise Descriptor=yes;
Dissimilarity Descriptor= APD;
(W1, W2)=(0.8, 0.2);
(MF Type)=Default;
LR Schedule=Exponential Decay with Step Size
)
EXPERIMENT using VD:
We have taken the best result obtained for APD. Since our NDCG did not show any improvement, we prioritized best-RMSE improvement test case to run this experiment. We ran the experiment with the condition : W1+W2=1.0 and iterating W1: [0.0,1.0](with step size 0.1)
Best_Case: RMSE
(
Social Relationship Descriptor= All;
Group Decision Strategies= Max Satisfaction, Avg Satisfaction, Minimum Misery;
Expertise Descriptor=yes;
Dissimilarity Descriptor= VD;
(W1, W2)=(0.8, 0.2);
(MF Type)=Default;
LR Schedule=Exponential Decay
)
OBSERVATIONS:
Optimal Values Obtained for (w1,w2) for each of the following cases:
APD: (0.9, 0.1)
VD: (0.9, 0.1)
This is an improved parameter than baseline algorithms where the authors have used (0.8, 0.2)
Once we have found our optimal (W1, W2) pair, our next step is to ensure consistency in the improvement of results. The baseline algorithm was deterministic in nature because it had no randomness in it. Using Matrix Factorization, we have introduced randomness in our proposed method in the following 2 ways:
We initialize the Latent Factors(P, Q) using randomness.
We also shuffle our data to start Matrix Factorization Training using randomness.
Therefore, it is important to validate whether our result is consistent throughout or is it just a special case for a particular seed coincidentally.
For this, we have run the best case test cases with optimal (w1,w2) on 50 seeds from [0,50). Finally we calculated the mean of these 50 iterations to compare with baseline error values.
APD Results:
For the APD NDCG error, we can see that the predicted mean (over all seeds) is higher than the baseline algorithm's NDCG
For the APD RMSE error, we can see that the predicted mean (over all seeds) is lower than the baseline algorithm's RMSE.
VD Results:
For the VD NDCG error, we can see that the predicted mean (over all seeds) is lower than the baseline algorithm's NDCG. Here we found no improvement.
For the VD RMSE error, we can see that the predicted mean (over all seeds) is lower than the baseline algorithm's RMSE
RQ 1: What is the relative performance of various combinations of group recommendation techniques when compared to one another?
We have generated valid combinations and compared them against each other in the Datasets and Results Section
RQ 2: Is it possible to get an optimal pipeline for group recommendation techniques, based on efficiency or performance?
We have concluded the following optimal Pipelines:
Combined NDCG + RMSE for APD: Baseline Algorithm(APD) + Default Matrix Factorization + Learning Rate Scheduler using exponential decay with step size + (w1,w2)=(0.9,0.1)
Best Case RMSE for VD : Baseline Algorithm(VD) + Default Matrix Factorization + Learning Rate Scheduler using exponential decay + (w1,w2)=(0.9,0.1)
RQ 3: Based on the comparative analysis to find optimal solution, can we also find the factors/ techniques contributing more to optimality and the ones contributing less or deviating/hindering the optimality?
Factors contributing to optimality:
Combination of Social Relationship + Group Decision Strategies + Weight Descriptor + Dissimilarity + Matrix Factorization
When using APD, we use Matrix Factorization with Learning Rate Scheduler and step size
When using VD, we use Matrix Factorization with Learning Rate Scheduler
Factors hindering optimality:
Any combination except taking into account all factors (above instance)
Baseline-Included Matrix Factorization
When using VD, we use Matrix Factorization with Learning Rate Scheduler with step size