Institution
Company•Sunnyvale, California, United States•
About: LinkedIn is a company organization based out in Sunnyvale, California, United States. It is known for research contribution in the topics: Social network & User interface. The organization has 1600 authors who have published 1720 publications receiving 24932 citations. The organization is also known as: LinkedIn Corporation & linkedin.com.
Topics: Social network, User interface, Ranking (information retrieval), Service (business), Web search query
Papers published on a yearly basis
Papers
More filters
••
10 Aug 2015TL;DR: This work presents two case studies where high-performance generalized additive models with pairwise interactions (GA2Ms) are applied to real healthcare problems yielding intelligible models with state-of-the-art accuracy.
Abstract: In machine learning often a tradeoff must be made between accuracy and intelligibility. More accurate models such as boosted trees, random forests, and neural nets usually are not intelligible, but more intelligible models such as logistic regression, naive-Bayes, and single decision trees often have significantly worse accuracy. This tradeoff sometimes limits the accuracy of models that can be applied in mission-critical applications such as healthcare where being able to understand, validate, edit, and trust a learned model is important. We present two case studies where high-performance generalized additive models with pairwise interactions (GA2Ms) are applied to real healthcare problems yielding intelligible models with state-of-the-art accuracy. In the pneumonia risk prediction case study, the intelligible model uncovers surprising patterns in the data that previously had prevented complex learned models from being fielded in this domain, but because it is intelligible and modular allows these patterns to be recognized and removed. In the 30-day hospital readmission case study, we show that the same methods scale to large datasets containing hundreds of thousands of patients and thousands of attributes while remaining intelligible and providing accuracy comparable to the best (unintelligible) machine learning methods.
1,301 citations
••
TL;DR: Deep learning has emerged as a powerful machine learning technique that learns multiple layers of representations or features of the data and produces state-of-the-art prediction results as mentioned in this paper, which is also popularly used in sentiment analysis in recent years.
Abstract: Deep learning has emerged as a powerful machine learning technique that learns multiple layers of representations or features of the data and produces state-of-the-art prediction results. Along with the success of deep learning in many other application domains, deep learning is also popularly used in sentiment analysis in recent years. This paper first gives an overview of deep learning and then provides a comprehensive survey of its current applications in sentiment analysis.
917 citations
••
26 Apr 2010TL;DR: This work develops an unsupervised model to estimate relationship strength from interaction activity and user similarity and evaluates it on real-world data from Facebook and LinkedIn, showing that the estimated link weights result in higher autocorrelation and lead to improved classification accuracy.
Abstract: Previous work analyzing social networks has mainly focused on binary friendship relations. However, in online social networks the low cost of link formation can lead to networks with heterogeneous relationship strengths (e.g., acquaintances and best friends mixed together). In this case, the binary friendship indicator provides only a coarse representation of relationship information. In this work, we develop an unsupervised model to estimate relationship strength from interaction activity (e.g., communication, tagging) and user similarity. More specifically, we formulate a link-based latent variable model, along with a coordinate ascent optimization procedure for the inference. We evaluate our approach on real-world data from Facebook and LinkedIn, showing that the estimated link weights result in higher autocorrelation and lead to improved classification accuracy.
725 citations
••
02 Feb 2017TL;DR: Recurrent Recommender Networks (RRN) are proposed that are able to predict future behavioral trajectories by endowing both users and movies with a Long Short-Term Memory (LSTM) autoregressive model that captures dynamics, in addition to a more traditional low-rank factorization.
Abstract: Recommender systems traditionally assume that user profiles and movie attributes are static. Temporal dynamics are purely reactive, that is, they are inferred after they are observed, e.g. after a user's taste has changed or based on hand-engineered temporal bias corrections for movies. We propose Recurrent Recommender Networks (RRN) that are able to predict future behavioral trajectories. This is achieved by endowing both users and movies with a Long Short-Term Memory (LSTM) autoregressive model that captures dynamics, in addition to a more traditional low-rank factorization. On multiple real-world datasets, our model offers excellent prediction accuracy and it is very compact, since we need not learn latent state but rather just the state transition function.
650 citations
•
01 Sep 2005TL;DR: In this paper, a method and system for evaluating the reputation of a member of a social networking system is disclosed, which is consistent with the embodiment of the invention of the technology.
Abstract: A method and system for evaluating the reputation of a member of a social networking system is disclosed. Consistent with an embodiment of the invention, one or more attributes associated with a social networking profile of a member of a social network are analyzed. Based on the analysis, a ranking, rating or score is assigned to a particular category of reputation. When requested, the ranking, rating or score is displayed to a user of the social network.
644 citations
Authors
Showing all 1600 results
Name | H-index | Papers | Citations |
---|---|---|---|
David Zhang | 111 | 1027 | 55118 |
Xi Chen | 105 | 1547 | 52533 |
Eric J. Huang | 72 | 201 | 22172 |
Haixun Wang | 63 | 277 | 16473 |
D. Hardtke | 53 | 134 | 15991 |
Lu Chen | 46 | 113 | 9092 |
Erez Petrank | 44 | 153 | 5798 |
Dennis P. Wall | 43 | 158 | 6789 |
Christopher J. Dawson | 43 | 260 | 5434 |
Gerald Francis McBrearty | 42 | 307 | 6517 |
Erich J. Windhab | 42 | 284 | 5808 |
Shipeng Yu | 40 | 94 | 7049 |
James Lee Hafner | 39 | 103 | 12097 |
James R. Kraemer | 38 | 222 | 3941 |
Glenn Fung | 38 | 147 | 6159 |