Projects

Online Product Review using Sentiment Analysis


Code:-

JPJA2389


Abstract:-

Unsupervised Cross-domain Sentiment Classification is the task of adapting a sentiment classifier trained on a particular domain (source domain), to a different domain (target domain), without requiring any labeled data for the target domain. By adapting an existing sentiment classifier to previously unseen target domains, we can avoid the cost for manual data annotation for the target domain. We model this problem as embedding learning, and construct three objective functions that capture: (a) distributional properties of pivots (i.e., common features that appear in both source and target domains), (b) label constraints in the source domain documents, and (c) geometric properties in the unlabeled documents in both source and target domains. Unlike prior proposals that first learn a lower-dimensional embedding independent of the source domain sentiment labels, and next a sentiment classifier in this embedding, our joint optimisation method learns embeddings that are sensitive to sentiment classification. Experimental results on a benchmark dataset show that by jointly optimising the three objectives we can obtain better performances in comparison to optimizing each objective function separately, thereby demonstrating the importance of task-specific embedding learning for cross-domain sentiment classification. Among the individual objective functions, the best performance is obtained by (c). Moreover, the proposed method reports cross-domain sentiment classification accuracies that are statistically comparable to the current state-of-the-artembedding learning methods for cross-domain sentiment classification.


Existing System:-

One popular solution to cross-domain sentiment classificationis to first project the source and the target features into the same lower-dimensional embedding, and subsequently learn a sentiment classifier on this embedded feature space.</li> <li>This approach is particularly attractive when there is little overlap between the original source and the target feature spaces. Similarly distributed words in the source and the target domains get mapped to closer points in the embedded space, thereby reducing the mismatch of features in the two domains.</li> <li>Prior work on cross-domain sentiment classification use unlabeled data from the source and the target domains to first learn a low-dimensional embedding for the two domains.</li> <li>Next, labeled reviews in the source domain are projected onto this embedding.Finally, a binary sentiment classifier is trained using the projected source domain labeled training instances.</li>


Disadvantages of Existing System:-

A limitation of existing two-step approach that decouples the embedding learning and sentiment classifier training is that the embeddings learnt in the first step is agnostic to the sentiment of the documents, which is the ultimate goal in cross-domain sentiment classification.</li> <li>This method does not consider source domain labeled data during the PLSR step, which makes the projection agnostic to sentiment classification.</li> <li>Supervised dimensionality reduction methods are prone to over fitting when the number of labeled instancesare small.</li>


Proposed System:-

We propose an unlabeled cross-domain sentiment classification method using spectral embeddings where we project both the words and the documents into the same lower dimensional embedding.</li> <li>The embedding learnt by our method enforces three important requirements.</li> <li>First, a set of domain independent features (also known as pivots) areselected from the source and target domains which must be mapped as close as possible in the embedded space.</li> <li>Second,friend closeness and enemy dispersion of the source domain labeled documents must be preserved. In other words, positively labeled documents must be embedded closer to each other and far from the negatively labeled documents. Likewise,negatively labeled documents must be embedded closer to each other and far from the positively labeled documents.</li> <li>Third, within each domain, the local geometry among the documents must be preserved. For example,unlabeled neighbour documents in the source domainmust be embedded closer to each other in the embedded space whereas, unlabeled neighbour documents in the targetdomain must be embedded closer to each other in the embedded space. Here, neighbour documents refer to similar documents in terms of their text content.</li> <li>We model each of the above-mentioned requirements as an objective function,and jointly optimise all three objective functions.</li>


Advantages of Proposed System:-

The proposed method can be easily extended to more than two sentiment classes.</li> <li>Our experimental results on a benchmark data set formulti-domain sentiment classification demonstrate that by jointly optimising the three objectives in many cases we obtain better classification accuracies than if we had optimized each objective separately.</li> <li>Even in cases where joint optimisation does not improve over the separately trained objectives, the performance obtained by the joint optimization method is never below that obtained by the best individually trained methods.</li> <li>This result shows the importance of learning embeddings that are sensitive to the final task at hand, which is sentiment classification.</li> <li>Moreover, the proposed method significantly outperforms several base lines and previously proposed embedding learning methods when applied to cross-domain sentiment classification.</li>


Hardware Requirements:-


Software Requirements:-

  • Operating system : Windows 7.</li> <li style="text-align: justify;">Coding Language : JAVA/J2EE</li> <li style="text-align: justify;">Tool : Netbeans 7.2.1</li> <li style="text-align: justify;">Database : MYSQL</li>

Cost:-

Rs 2000


Tools Used

Java

Cost

₹Rs 2000