Home OALib Journal OALib PrePrints Submit Ranking News My Lib FAQ About Us Follow Us+
 Title Keywords Abstract Author All
Search Results: 1 - 10 of 100 matches for " "
 Page 1 /100 Display every page 5 10 20 Item
 Computer Science , 2008, Abstract: We present a general approach for collaborative filtering (CF) using spectral regularization to learn linear operators from "users" to the "objects" they rate. Recent low-rank type matrix completion approaches to CF are shown to be special cases. However, unlike existing regularization based CF methods, our approach can be used to also incorporate information such as attributes of the users or the objects -- a limitation of existing regularization based CF methods. We then provide novel representer theorems that we use to develop new estimation methods. We provide learning algorithms based on low-rank decompositions, and test them on a standard CF dataset. The experiments indicate the advantages of generalizing the existing regularization based CF methods to incorporate related information about users and objects. Finally, we show that certain multi-task learning methods can be also seen as special cases of our proposed approach.
 Loc Nguyen Journal of Data Analysis and Information Processing (JDAIP) , 2016, DOI: 10.4236/jdaip.2016.43011 Abstract: Traditional collaborative filtering (CF) does not take into account contextual factors such as time, place, companion, environment, etc. which are useful information around users or relevant to recommender application. So, recent aware-context CF takes advantages of such information in order to improve the quality of recommendation. There are three main aware-context approaches: contextual pre-filtering, contextual post-filtering and contextual modeling. Each approach has individual strong points and drawbacks but there is a requirement of steady and fast inference model which supports the aware-context recommendation process. This paper proposes a new approach which discovers multivariate logistic regression model by mining both traditional rating data and contextual data. Logistic model is optimal inference model in response to the binary question “whether or not a user prefers a list of recommendations with regard to contextual condition”. Consequently, such regression model is used as a filter to remove irrelevant items from recommendations. The final list is the best recommendations to be given to users under contextual information. Moreover the searching items space of logistic model is reduced to smaller set of items so-called general user pattern (GUP). GUP supports logistic model to be faster in real-time response.
 Computer Science , 2015, Abstract: Neighbourhood-based Collaborative Filtering (CF) has been applied in the industry for several decades, because of the easy implementation and high recommendation accuracy. As the core of neighbourhood-based CF, the task of dynamically maintaining users' similarity list is challenged by cold-start problem and scalability problem. Recently, several methods are presented on solving the two problems. However, these methods applied an $O(n^2)$ algorithm to compute the similarity list in a special case, where the new users, with enough recommendation data, have the same rating list. To address the problem of large computational cost caused by the special case, we design a faster ($O(\frac{1}{125}n^2)$) algorithm, TwinSearch Algorithm, to avoid computing and sorting the similarity list for the new users repeatedly to save the computational resources. Both theoretical and experimental results show that the TwinSearch Algorithm achieves better running time than the traditional method.
 Computer Science , 2015, Abstract: There is much empirical evidence that item-item collaborative filtering works well in practice. Motivated to understand this, we provide a framework to design and analyze various recommendation algorithms. The setup amounts to online binary matrix completion, where at each time a random user requests a recommendation and the algorithm chooses an entry to reveal in the user's row. The goal is to minimize regret, or equivalently to maximize the number of +1 entries revealed at any time. We analyze an item-item collaborative filtering algorithm that can achieve fundamentally better performance compared to user-user collaborative filtering. The algorithm achieves good "cold-start" performance (appropriately defined) by quickly making good recommendations to new users about whom there is little information.
 Journal of Software , 2011, DOI: 10.4304/jsw.6.6.993-1000 Abstract: Memory-based collaborative filtering (CF) is applied to help users to find their favorite items in recommender systems. Up to now, this approach has been proven successful in recommender systems, such as e-commerce systems. The idea of this approach is that the interest of a particular user will be more consistent with those who share similar preference with him or her. Therefore, it is critical that an appropriate similarity measure should be selected for making recommendations. This paper proposes a new similarity measure named adjusted Euclidean distance (AED) method which unifies all Euclidean distances between vectors in different dimensional vector spaces. Our AED enjoy the advantages that it takes both the length of vectors and different dimension-numbers of vector spaces into consideration. Based on two datasets MovieLens and Book-Crossing, we conduct experiments comparing our AED with two notable existing methods. The experimental results demonstrate that our AED improves the accuracy of prediction and recommendation.
 Zijiang Zhu Communications and Network (CN) , 2018, DOI: 10.4236/cn.2018.103009 Abstract: Collaborative filtering algorithm is the most widely used and recommended algorithm in major e-commerce recommendation systems nowadays. Concerning the problems such as poor adaptability and cold start of traditional collaborative filtering algorithms, this paper is going to come up with improvements and construct a hybrid collaborative filtering algorithm model which will possess excellent scalability. Meanwhile, this paper will also optimize the process based on the parameter selection of genetic algorithm and demonstrate its pseudocode reference so as to provide new ideas and methods for the study of parameter combination optimization in hybrid collaborative filtering algorithm.
 Computer Science , 2013, Abstract: Collaborative filtering or recommender systems use a database about user preferences to predict additional topics or products a new user might like. In this paper we describe several algorithms designed for this task, including techniques based on correlation coefficients, vector-based similarity calculations, and statistical Bayesian methods. We compare the predictive accuracy of the various methods in a set of representative problem domains. We use two basic classes of evaluation metrics. The first characterizes accuracy over a set of individual predictions in terms of average absolute deviation. The second estimates the utility of a ranked list of suggested items. This metric uses an estimate of the probability that a user will see a recommendation in an ordered list. Experiments were run for datasets associated with 3 application areas, 4 experimental protocols, and the 2 evaluation metrics for the various algorithms. Results indicate that for a wide range of conditions, Bayesian networks with decision trees at each node and correlation methods outperform Bayesian-clustering and vector-similarity methods. Between correlation and Bayesian networks, the preferred method depends on the nature of the dataset, nature of the application (ranked versus one-by-one presentation), and the availability of votes with which to make predictions. Other considerations include the size of database, speed of predictions, and learning time.
 Computer Science , 2012, Abstract: Collaborative filtering (CF) allows the preferences of multiple users to be pooled to make recommendations regarding unseen products. We consider in this paper the problem of online and interactive CF: given the current ratings associated with a user, what queries (new ratings) would most improve the quality of the recommendations made? We cast this terms of expected value of information (EVOI); but the online computational cost of computing optimal queries is prohibitive. We show how offline prototyping and computation of bounds on EVOI can be used to dramatically reduce the required online computation. The framework we develop is general, but we focus on derivations and empirical study in the specific case of the multiple-cause vector quantization model.
 Computer Science , 2012, Abstract: Collaborative filtering is a rapidly advancing research area. Every year several new techniques are proposed and yet it is not clear which of the techniques work best and under what conditions. In this paper we conduct a study comparing several collaborative filtering techniques -- both classic and recent state-of-the-art -- in a variety of experimental contexts. Specifically, we report conclusions controlling for number of items, number of users, sparsity level, performance criteria, and computational complexity. Our conclusions identify what algorithms work well and in what conditions, and contribute to both industrial deployment collaborative filtering algorithms and to the research community.
 Journal of Emerging Technologies in Web Intelligence , 2010, DOI: 10.4304/jetwi.2.4.300-309 Abstract: The growing popularity of Social Networks raises the important issue of trust. Among many systems which have realized the impact of trust, Recommender Systems have been the most influential ones. Collaborative Filtering Recommenders take advantage of trust relations between users for generating more accurate predictions. In this paper, we propose a semantic recommendation framework for creating trust relationships among all types of users with respect to different types of items, which are accessed by unique URI across heterogeneous networks and environments. We gradually build up the trust relationships between users based on the rating information from user profiles and item profiles to generate trust networks of users. For analyzing the formation of trust networks, we employ Tindex as an estimate of a user’s trustworthiness to identify and select neighbors in an effective manner. In this work, we utilize T-index to form the list of an item’s raters, called TopTrustee list for keeping the most reliable users who have already shown interest in the respective item. Thus, when a user rates an item, he/she is able to find users who can be trustworthy neighbors even though they might not be accessible within an upper bound of traversal path length. An empirical evaluation demonstrates how T-index improves the Trust Network structure by generating connections to more trustworthy users. We also show that exploiting Tindex results in better prediction accuracy and coverage of recommendations collected along few edges that connect users on a Social Network.
 Page 1 /100 Display every page 5 10 20 Item