Integrating Summarization and Retrieval For Enhanced Personalization Via Large Language Models
Integrating Summarization and Retrieval For Enhanced Personalization Via Large Language Models
Table 2: Prompts used for summarization. Additional tokens were used for Vicuna summaries to match the expected prompt
format for that model but the content was the same for Vicuna and ChatGPT.
Task Prompt
LaMP-1 Write a summary, in English, of the research interests and topics of a researcher who has published the following papers.
Only generate the summary, no other text.
LaMP-2 Look at the following past articles this journalist has written and determine the most popular category they write in.
Answer in the following form: most popular category: <category>
LaMP-3 Based on this user’s past reviews, what are the most common scores they give for positive and negative reviews? Answer
in the following form: most common positive score: <most common positive score>, most common negative score: <most
common negative score>
LaMP-4 Given this author’s previous articles, try to describe a template for their headlines. I want to be able to accurately predict
the headline gives one of their articles. Be specific about their style and wording, don’t tell me anything generic.
LaMP-5 Given this author’s previous publications, try to describe a template for their titles. I want to be able to accurately predict
the title of one of the papers from the abstract. Only generate the template description, nothing else.
LaMP-7 Given this person’s previous tweets, try to describe a template for their tweets. I want to take a generic sentence and
rephrase it to sound like one of their tweets, with the same style/punctuation/capitalization/wording/tone/etc. as them.
Only give me the template description, nothing else.
Table 3: Results for FlanT5-base model fine-tuned on LaMP benchmark tasks. Baseline: retrieval of k user data entries; Vicuna
Summ.: Baseline + summary of user data generated by vicuna; GPT-3.5 Summ.: Baseline + summary of user data generated by
GPT-3.5. Underline means summary improved the corresponding baselines with the same k, and bold means the best results
among compared experiments for each task. For all metrics, higher is better except in the case of MAE and RMSE used for
LaMP-3.
machine learning and data analytics for network monitoring and demonstrates a strong emphasis on efficient and scalable solutions
diagnosis. Their recent works also focus on addressing the latency for a wide range of machine learning and data analysis tasks.
and energy efficiency challenges in wireless multimedia networks GPT-3.5 Summ. Prediction: [1]
and leveraging software-defined networking for next-generation
cellular networks. In summary, this researcher’s work revolves
around developing novel algorithms, protocols, and systems that
improve the performance, resilience, and security of networked
systems.
GPT-3.5 Summ. Prediction: [1]
A.2 Example 2
Task: LaMP-1
Input: For an author who has written the paper with the title
"Visual-audio integration for user authentication system of partner
robots", which reference is related? Just answer with [1] or [2]
without explanation. [1]: "Pronunciation Modeling for Improved
Spelling Correction" [2]: "Path following algorithm for highly re-
dundant manipulators"
Output: [1]
Retrieved user data: Advancing Matrix Completion by Model-
ing Extra Structures beyond Low-Rankness
Improved Asymmetric Locality Sensitive Hashing (ALSH) for
Maximum Inner Product Search (MIPS)
Real-Time Implementation of Improved State-Space MPC for Air
Supply in a Coke Furnace
On Practical Algorithms for Entropy Estimation and the Im-
proved Sample Complexity of Compressed Counting
Baseline Prediction: [2]
User Summary by Vicuna: 1. Nystrom Method for Approxi-
mating the GMM Kernel.
2. Very sparse random projections.
3. Very sparse stable random projections for dimension reduction
in lalpha (0 lt;alphalt;=2) norm.
4. A Comparison Study of Nonlinear Kernels.
5. One sketch for all: Theory and Application of Conditional
Random Sampling.
6. Collaborative Multi-objective Ranking.
7. Accurate Estimators for Improving Minwise Hashing and b-Bit
Minwise Hashing.
8. A new space for comparing graphs.
9. Stability and Risk Bounds of Iterative Hard Thresholding.
10. Hashing Algorithms for Large-Scale Learning.
11. Variational Flow Graphical Model.
Vicuna Summ. Prediction: [2]
User Summary by GPT-3.5: The researcher’s primary research
interests revolve around the development of practical algorithms
for entropy estimation, compressed counting, and matrix comple-
tion. They explore various techniques for improving the efficiency
and accuracy of these methods. Additionally, the researcher focuses
on advancing hashing algorithms, particularly in the context of
maximum inner product search (MIPS) and locality-sensitive hash-
ing (LSH). They also investigate sketch-based sampling techniques,
especially for sparse data, and propose novel approaches for esti-
mating entropy of data streams. Furthermore, the researcher has
contributions in optimization methods, classification, and dimen-
sionality reduction using stable random projections. Their work