LogoLogo
Continuum WebsiteContinuum ApplicationsContinuum KnowledgeAxolotl Platform
Continuum Knowledge
Continuum Knowledge
  • Continuum
  • Data
    • Datasets
      • Pre Training Data
      • Types of Fine Tuning
      • Self Instruct Paper
      • Self-Alignment with Instruction Backtranslation
      • Systematic Evaluation of Instruction-Tuned Large Language Models on Open Datasets
      • Instruction Tuning
      • Instruction Fine Tuning - Alpagasus
      • Less is More For Alignment
      • Enhanced Supervised Fine Tuning
      • Visualising Data using t-SNE
      • UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction
      • Training and Evaluation Datasets
      • What is perplexity?
  • MODELS
    • Foundation Models
      • The leaderboard
      • Foundation Models
      • LLama 2 - Analysis
      • Analysis of Llama 3
      • Llama 3.1 series
      • Google Gemini 1.5
      • Platypus: Quick, Cheap, and Powerful Refinement of LLMs
      • Mixtral of Experts
      • Mixture-of-Agents (MoA)
      • Phi 1.5
        • Refining the Art of AI Training: A Deep Dive into Phi 1.5's Innovative Approach
      • Phi 2.0
      • Phi-3 Technical Report
  • Training
    • The Fine Tuning Process
      • Why fine tune?
        • Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
        • Explanations in Fine Tuning
      • Tokenization
        • Tokenization Is More Than Compression
        • Tokenization - SentencePiece
        • Tokenization explore
        • Tokenizer Choice For LLM Training: Negligible or Crucial?
        • Getting the most out of your tokenizer for pre-training and domain adaptation
        • TokenMonster
      • Parameter Efficient Fine Tuning
        • P-Tuning
          • The Power of Scale for Parameter-Efficient Prompt Tuning
        • Prefix-Tuning: Optimizing Continuous Prompts for Generation
        • Harnessing the Power of PEFT: A Smarter Approach to Fine-tuning Pre-trained Models
        • What is Low-Rank Adaptation (LoRA) - explained by the inventor
        • Low Rank Adaptation (Lora)
        • Practical Tips for Fine-tuning LMs Using LoRA (Low-Rank Adaptation)
        • QLORA: Efficient Finetuning of Quantized LLMs
        • Bits and Bytes
        • The Magic behind Qlora
        • Practical Guide to LoRA: Tips and Tricks for Effective Model Adaptation
        • The quantization constant
        • QLORA: Efficient Finetuning of Quantized Language Models
        • QLORA and Fine-Tuning of Quantized Language Models (LMs)
        • ReLoRA: High-Rank Training Through Low-Rank Updates
        • SLoRA: Federated Parameter Efficient Fine-Tuning of Language Models
        • GaLora: Memory-Efficient LLM Training by Gradient Low-Rank Projection
      • Hyperparameters
        • Batch Size
        • Padding Tokens
        • Mixed precision training
        • FP8 Formats for Deep Learning
        • Floating Point Numbers
        • Batch Size and Model loss
        • Batch Normalisation
        • Rethinking Learning Rate Tuning in the Era of Language Models
        • Sample Packing
        • Gradient accumulation
        • A process for choosing the learning rate
        • Learning Rate Scheduler
        • Checkpoints
        • A Survey on Efficient Training of Transformers
        • Sequence Length Warmup
        • Understanding Training vs. Evaluation Data Splits
        • Cross-entropy loss
        • Weight Decay
        • Optimiser
        • Caching
      • Training Processes
        • Extending the context window
        • PyTorch Fully Sharded Data Parallel (FSDP)
        • Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
        • YaRN: Efficient Context Window Extension of Large Language Models
        • Sliding Window Attention
        • LongRoPE
        • Reinforcement Learning
        • An introduction to reinforcement learning
        • Reinforcement Learning from Human Feedback (RLHF)
        • Direct Preference Optimization: Your Language Model is Secretly a Reward Model
  • INFERENCE
    • Why is inference important?
      • Grouped Query Attention
      • Key Value Cache
      • Flash Attention
      • Flash Attention 2
      • StreamingLLM
      • Paged Attention and vLLM
      • TensorRT-LLM
      • Torchscript
      • NVIDIA L40S GPU
      • Triton Inference Server - Introduction
      • Triton Inference Server
      • FiDO: Fusion-in-Decoder optimised for stronger performance and faster inference
      • Is PUE a useful measure of data centre performance?
      • SLORA
  • KNOWLEDGE
    • Vector Databases
      • A Comprehensive Survey on Vector Databases
      • Vector database management systems: Fundamental concepts, use-cases, and current challenges
      • Using the Output Embedding to Improve Language Models
      • Decoding Sentence-BERT
      • ColBERT: Efficient and Effective Passage Search via Contextualized Late Interaction over BERT
      • SimCSE: Simple Contrastive Learning of Sentence Embeddings
      • Questions Are All You Need to Train a Dense Passage Retriever
      • Improving Text Embeddings with Large Language Models
      • Massive Text Embedding Benchmark
      • RocketQAv2: A Joint Training Method for Dense Passage Retrieval and Passage Re-ranking
      • LLM2Vec: Large Language Models Are Secretly Powerful Text Encoders
      • Embedding and Fine-Tuning in Neural Language Models
      • Embedding Model Construction
      • Demystifying Embedding Spaces using Large Language Models
      • Fine-Tuning Llama for Multi-Stage Text Retrieval
      • Large Language Model Based Text Augmentation Enhanced Personality Detection Model
      • One Embedder, Any Task: Instruction-Finetuned Text Embeddings
      • Vector Databases are not the only solution
      • Knowledge Graphs
        • Harnessing Knowledge Graphs to Elevate AI: A Technical Exploration
        • Unifying Large Language Models and Knowledge Graphs: A Roadmap
      • Approximate Nearest Neighbor (ANN)
      • High Dimensional Data
      • Principal Component Analysis (PCA)
      • Vector Similarity Search - HNSW
      • FAISS (Facebook AI Similarity Search)
      • Unsupervised Dense Retrievers
    • Retrieval Augmented Generation
      • Retrieval-Augmented Generation for Large Language Models: A Survey
      • Fine-Tuning or Retrieval?
      • Revolutionising Information Retrieval: The Power of RAG in Language Models
      • A Survey on Retrieval-Augmented Text Generation
      • REALM: Retrieval-Augmented Language Model Pre-Training
      • Retrieve Anything To Augment Large Language Models
      • Generate Rather Than Retrieve: Large Language Models Are Strong Context Generators
      • Active Retrieval Augmented Generation
      • DSPy: LM Assertions: Enhancing Language Model Pipelines with Computational Constraints
      • DSPy: Compiling Declarative Language Model Calls
      • DSPy: In-Context Learning for Extreme Multi-Label Classification
      • Optimizing Instructions and Demonstrations for Multi-Stage Language Model Programs
      • HYDE: Revolutionising Search with Hypothetical Document Embeddings
      • Enhancing Recommender Systems with Large Language Model Reasoning Graphs
      • Retrieval Augmented Generation (RAG) versus fine tuning
      • RAFT: Adapting Language Model to Domain Specific RAG
      • Summarisation Methods and RAG
      • Lessons Learned on LLM RAG Solutions
      • Stanford: Retrieval Augmented Language Models
      • Overview of RAG Approaches with Vector Databases
      • Mastering Chunking in Retrieval-Augmented Generation (RAG) Systems
    • Semantic Routing
    • Resource Description Framework (RDF)
  • AGENTS
    • What is agency?
      • Rephrase and Respond: Let Large Language Models Ask Better Questions for Themselves
      • Types of Agents
      • The risk of AI agency
      • Understanding Personality in Large Language Models: A New Frontier in AI Psychology
      • AI Agents - Reasoning, Planning, and Tool Calling
      • Personality and Brand
      • Agent Interaction via APIs
      • Bridging Minds and Machines: The Legacy of Newell, Shaw, and Simon
      • A Survey on Language Model based Autonomous Agents
      • Large Language Models as Agents
      • AI Reasoning: A Deep Dive into Chain-of-Thought Prompting
      • Enhancing AI Reasoning with Self-Taught Reasoner (STaR)
      • Exploring the Frontier of AI: The "Tree of Thoughts" Framework
      • Toolformer: Revolutionising Language Models with API Integration - An Analysis
      • TaskMatrix.AI: Bridging Foundational AI Models with Specialised Systems for Enhanced Task Completion
      • Unleashing the Power of LLMs in API Integration: The Rise of Gorilla
      • Andrew Ng's presentation on AI agents
      • Making AI accessible with Andrej Karpathy and Stephanie Zhan
  • Regulation and Ethics
    • Regulation and Ethics
      • Privacy
      • Detecting AI Generated content
      • Navigating the IP Maze in AI: The Convergence of Blockchain, Web 3.0, and LLMs
      • Adverse Reactions to generative AI
      • Navigating the Ethical Minefield: The Challenge of Security in Large Language Models
      • Navigating the Uncharted Waters: The Risks of Autonomous AI in Military Decision-Making
  • DISRUPTION
    • Data Architecture
      • What is a data pipeline?
      • What is Reverse ETL?
      • Unstructured Data and Generatve AI
      • Resource Description Framework (RDF)
      • Integrating generative AI with the Semantic Web
    • Search
      • BM25 - Search Engine Ranking Function
      • BERT as a reranking engine
      • BERT and Google
      • Generative Engine Optimisation (GEO)
      • Billion-scale similarity search with GPUs
      • FOLLOWIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions
      • Neural Collaborative Filtering
      • Federated Neural Collaborative Filtering
      • Latent Space versus Embedding Space
      • Improving Text Embeddings with Large Language Models
    • Recommendation Engines
      • On Interpretation and Measurement of Soft Attributes for Recommendation
      • A Survey on Large Language Models for Recommendation
      • Model driven recommendation systems
      • Recommender AI Agent: Integrating Large Language Models for Interactive Recommendations
      • Foundation Models for Recommender Systems
      • Exploring the Impact of Large Language Models on Recommender Systems: An Extensive Review
      • AI driven recommendations - harming autonomy?
    • Logging
      • A Taxonomy of Anomalies in Log Data
      • Deeplog
      • LogBERT: Log Anomaly Detection via BERT
      • Experience Report: Deep Learning-based System Log Analysis for Anomaly Detection
      • Log-based Anomaly Detection with Deep Learning: How Far Are We?
      • Deep Learning for Anomaly Detection in Log Data: A Survey
      • LogGPT
      • Adaptive Semantic Gate Networks (ASGNet) for log-based anomaly diagnosis
  • Infrastructure
    • The modern data centre
      • Enhancing Data Centre Efficiency: Strategies to Improve PUE
      • TCO of NVIDIA GPUs and falling barriers to entry
      • Maximising GPU Utilisation with Kubernetes and NVIDIA GPU Operator
      • Data Centres
      • Liquid Cooling
    • Servers and Chips
      • The NVIDIA H100 GPU
      • NVIDIA H100 NVL
      • Lambda Hyperplane 8-H100
      • NVIDIA DGX Servers
      • NVIDIA DGX-2
      • NVIDIA DGX H-100 System
      • NVLink Switch
      • Tensor Cores
      • NVIDIA Grace Hopper Superchip
      • NVIDIA Grace CPU Superchip
      • NVIDIA GB200 NVL72
      • Hopper versus Blackwell
      • HGX: High-Performance GPU Platforms
      • ARM Chips
      • ARM versus x86
      • RISC versus CISC
      • Introduction to RISC-V
    • Networking and Connectivity
      • Infiniband versus Ethernet
      • NVIDIA Quantum InfiniBand
      • PCIe (Peripheral Component Interconnect Express)
      • NVIDIA ConnectX InfiniBand adapters
      • NVMe (Non-Volatile Memory Express)
      • NVMe over Fabrics (NVMe-oF)
      • NVIDIA Spectrum-X
      • NVIDIA GPUDirect
      • Evaluating Modern GPU Interconnect
      • Scalable Hierarchical Aggregation and Reduction Protocol (SHARP)
      • Next-generation networking in AI environments
      • NVIDIA Collective Communications Library (NCCL)
    • Data and Memory
      • NVIDIA BlueField Data Processing Units (DPUs)
      • Remote Direct Memory Access (RDMA)
      • High Bandwidth Memory (HBM3)
      • Flash Memory
      • Model Requirements
      • Calculating GPU memory for serving LLMs
      • Transformer training costs
      • GPU Performance Optimisation
    • Libraries and Complements
      • NVIDIA Base Command
      • NVIDIA AI Enterprise
      • CUDA - NVIDIA GTC 2024 presentation
      • RAPIDs
      • RAFT
    • Vast Data Platform
      • Vast Datastore
      • Vast Database
      • Vast Data Engine
      • DASE (Disaggregated and Shared Everything)
      • Dremio and VAST Data
    • Storage
      • WEKA: A High-Performance Storage Solution for AI Workloads
      • Introduction to NVIDIA GPUDirect Storage (GDS)
        • GDS cuFile API
      • NVIDIA Magnum IO GPUDirect Storage (GDS)
      • Vectors in Memory
Powered by GitBook
LogoLogo

Continuum - Accelerated Artificial Intelligence

  • Continuum Website
  • Axolotl Platform

Copyright Continuum Labs - 2023

On this page
  • Key points from the survey
  • Taxonomy of FM4RecSys
  • Key points from the survey
  • Personalised recommendations through fine-tuning
  • Improved cold-start recommendations
  • Explainable recommendations through prompting
  • Context-aware recommendations
  • Interactive and conversational recommendations
  • Fairness-aware recommendations
  • Cross-domain recommendations
  • Examples of models
  • Language Foundation Models for RecSys
  • Personalised Agent for RecSys with Foundation Models
  • Open problems and opportunities
  • Conclusion
  • References

Was this helpful?

  1. DISRUPTION
  2. Recommendation Engines

Foundation Models for Recommender Systems

PreviousRecommender AI Agent: Integrating Large Language Models for Interactive RecommendationsNextExploring the Impact of Large Language Models on Recommender Systems: An Extensive Review

Last updated 11 months ago

Was this helpful?

This February 2024 paper provides a comprehensive survey of the emerging field of Foundation Model-based Recommender Systems (FM4RecSys).

While I am not a fan of this acronym they have created, it is an important and growing field.

The authors review the research background, provide a taxonomy of existing FM4RecSys works, and discuss open problems and opportunities in this area.

The taxonomy is divided into four parts: data characteristics, representation learning, model type, and downstream tasks.

Key developments in each area are reviewed, with a focus on representative models and their characteristics.

The authors also classify FM4RecSys frameworks by model type, including language foundation models, personalised agents, and multi-modal foundation models.

They then discuss applications of FM4RecSys in tasks such as top-k recommendation, context-aware recommendation, interactive recommendation, cross-domain recommendation, and interpretability and fairness.

Key points from the survey

Motivation for FM4RecSys

Enhanced generalisation capabilities: FMs can learn from large-scale data and generalise better to new, unseen data, improving recommendations for new users or items with limited information.

Elevated recommendation experience: FMs enable more dynamic and unstructured conversational interactions, offering enhanced interactivity and flexibility in user engagements.

Improved explanation and reasoning capabilities: FMs can generate more coherent and logically sound explanations by leveraging a comprehensive grasp of commonsense and user-specific context.

Distinguishing features from recent LLM-based RecSys surveys

The authors provide a unique viewpoint for examining FM4RecSys, systematically outlining the framework from data characteristics to downstream tasks.

The survey categorises FM4RecSys based on both the types of models used and the recommendation tasks themselves.

The survey covers a wider array of foundation models, not just large language models (LLMs).

Taxonomy of FM4RecSys

Data characteristics: User/Item ID, user profile, item side information, and external knowledge bases.

Representation learning: ID-based representation, multi-modal representation, and hybrid representation.

Model type: Language FM, personalised agent FM, and multi-modal FM.

Downstream tasks: Top-K recommendation, interpretability in recommendation, context-aware recommendation, fairness in recommendation, and interactive recommendation.

Key points from the survey

Large Language Models (LLMs) have shown potential in enhancing recommendation systems (RSs) by leveraging their extensive knowledge, reasoning capabilities, and generative power.

Here are some key ways LLMs can be used to create better recommendation systems, along with implementation details:

Personalised recommendations through fine-tuning

  • LLMs can be fine-tuned on user-specific data, such as user profiles, item interactions, and reviews, to generate personalised recommendations.

  • Models like InstructRec, TallRec, and BIGRec demonstrate the effectiveness of fine-tuning LLMs for recommendation tasks.

  • Implementation: Collect user data, preprocess it into a suitable format (e.g., user-item interaction sequences), and fine-tune the LLM using techniques like instruction tuning or parameter-efficient methods such as LoRA.

Improved cold-start recommendations

  • LLMs can leverage their pre-trained knowledge to provide accurate recommendations even for new users or items with limited data.

  • By using item titles, descriptions, or other side information, LLMs can infer user preferences and generate relevant recommendations.

  • Implementation: Represent new users or items using available side information (e.g., item titles) and feed this into the fine-tuned LLM to generate recommendations without relying on extensive interaction data.

Explainable recommendations through prompting

  • LLMs can generate natural language explanations for recommended items, improving transparency and user trust.

  • Prompting techniques, such as asking the LLM to explain why an item is recommended to a specific user, can be used to generate coherent and informative explanations.

  • Implementation: Design prompts that include user and item information, and use the LLM to generate explanations. Optionally, incorporate item features or continuous prompt vectors to guide the explanation process.

Context-aware recommendations

  • LLMs can integrate various contextual information, such as user preferences, item attributes, and external knowledge, to provide more relevant recommendations.

  • Techniques like knowledge prompt tuning and hybrid item representation can be used to incorporate structured knowledge and bridge the semantic gap between traditional RSs and LLMs.

  • Implementation: Construct prompts that include contextual information (e.g., user preferences, item attributes) and use the LLM to generate context-aware recommendations. Optionally, integrate external knowledge bases or adapt the LLM's input space to facilitate knowledge transfer.

Interactive and conversational recommendations

  • LLMs can engage in multi-turn conversations with users, refining preferences and providing personalized recommendations based on the conversation history.

  • Techniques like role-playing prompts and augmentations such as Retrieval-Augmented Generation (RAG) and Chain-of-Thought (CoT) can enhance the LLM's conversational abilities.

  • Implementation: Design conversational prompts and fine-tune the LLM on conversation datasets. Use the fine-tuned LLM to engage in multi-turn conversations with users, updating recommendations based on the evolving user preferences.

Fairness-aware recommendations

  • LLMs can be guided to generate recommendations that are fair and unbiased towards different user and item groups.

  • Techniques like Counterfactually-Fair-Prompting (CFP) and conditional ranking can be used to promote fairness in LLM-based RSs.

  • Implementation: Craft prompts that include fairness constraints or sensitive attributes, and train the LLM to generate recommendations that satisfy these constraints. Evaluate the fairness of the generated recommendations using appropriate metrics.

Cross-domain recommendations

  • LLMs can leverage their pre-trained knowledge across various domains to provide recommendations in data-scarce target domains.

  • Techniques like domain-specific adapters, behavior mixing, and knowledge adaptation can be used to facilitate cross-domain knowledge transfer.

  • Implementation: Fine-tune the LLM on data from multiple domains, using techniques like adapter-based learning or prompting strategies that capture domain-specific characteristics. Use the fine-tuned LLM to generate recommendations in the target domain.

To implement these ideas effectively, it is important to have access to high-quality user and item data, design appropriate prompts and fine-tuning strategies, and carefully evaluate the generated recommendations using relevant metrics and user feedback.

Collaboration between domain experts, data scientists, and ML engineers is essential to successfully integrate LLMs into recommendation systems and create a more personalised, explainable, and fair user experience.

Examples of models

To better understand the practical applications of Foundation Model-based Recommender Systems (FM4RecSys), it is essential to examine specific examples of models and their characteristics.

This section will explore various types of LLM-based recommendation models.

By reviewing their workings, architecture, and best applications of these models, we can gain insights into how FM4RecSys can be effectively implemented to enhance the performance and user experience of recommendation systems.

Language Foundation Models for RecSys

Pre-trained models for Foundation Model-based Recommender Systems (FM4RecSys)

These models are pre-trained on massive recommendation datasets using transformer-based architectures. They employ language modeling tasks such as masked language modeling and permutation language modeling.

They require a large amount of domain data and have high training costs. They are best suited for scenarios where extensive domain-specific data is available.

Directly fine-tuned models for Foundation Model-based Recommender Systems (FM4RecSys)

These models are fine-tune pre-trained LLMs using techniques like instruction tuning and parameter-efficient methods such as LoRA.

These models can understand and follow different instructions for recommendations and are effective for cold-start recommendations. They are best applied in situations where specific instructions or user history is available for fine-tuning.

Prompting for Foundation Model-based Recommender Systems (FM4RecSys)

This approach focuses on extracting knowledge from LLMs using prompting strategies without fine-tuning the model parameters.

These models design appropriate prompts to stimulate the recommendation abilities of LLMs. They are suitable for scenarios where fine-tuning is not feasible or when quick adaptation to new recommendation tasks is required.

Personalised Agent for RecSys with Foundation Models

Agent as User Simulator

These models treat users as FM-based autonomous agents within a virtual simulator.

These agents can simulate complex user behaviours and interactions within the recommendation system.

They are best applied in scenarios where gathering real user behaviour data is expensive or ethically complex, and when simulating realistic user interactions is crucial for training and evaluation.

Agent as Recommender Systems (RecSys)

These models leverage the reasoning, reflection, and tool usage capabilities of FMs for recommendation.

They treat FMs as the brain and recommendation models as tools that supply domain-specific knowledge. These models are suitable for scenarios that require advanced reasoning and tool integration for personalised recommendations.

While the examples of Foundation Model-based Recommender models demonstrate the potential for improved recommendation systems, it is important to acknowledge and address the open problems and opportunities in this field.

By identifying and tackling these challenges, researchers and practitioners can further advance the development and deployment of FM4RecSys, ensuring that these systems continue to evolve and meet the growing demands of businesses and users alike.

In the following section, we will explore some of the key open problems and opportunities in FM4RecSys.

Open problems and opportunities

Long-sequences in FM4RecSys

  • Issue: FMs have fixed context window limitations, which can impact their effectiveness in tasks requiring extensive context, such as context-aware recommendations.

  • Resolution: Adapt techniques from NLP, including segmenting and summarising inputs to fit within the context window, and employ strategies like attention mechanisms and memory augmentation to focus on relevant input segments. The RoPE technique with rotary position embedding shows promise in managing long inputs.

Explainability and Trustworthiness

  • Issue: The complexity and size of FMs introduce new hurdles in explaining FM4RecSys, making it challenging to enhance explainability and trustworthiness.

  • Resolution: Align FMs with explicit knowledge bases like knowledge graphs to make the decision-making process traceable. Techniques like Chain/Tree of Thoughts can further enhance the explanations. Generate natural language explanations for recommendations to improve transparency.

Temporal Extrapolation

  • Issue: Time series modeling in recommendation systems has not fully benefited from large-scale pretraining due to its unique challenges, such as variable scales, sampling rates, and data gaps.

  • Resolution: Leverage approaches like LLMTIME2, which encodes time series as numerical strings and treats forecasting as a next-token prediction task. Transform token distributions into continuous densities to facilitate the application of LLMs to time series forecasting without specialised knowledge or high computational costs.

Multi-modal Agent AI for RecSys

  • Issue: Integrating multi-modal understanding and interaction in recommendation systems to achieve more intelligent and personalised recommendations.

  • Resolution: Develop agent AI systems that can perceive and act in various domains, process visual and contextual data, understand user actions and behaviours, and produce meaningful responses. Use these agents as simulators for both RSs and users to collect data and train in an offline environment, reducing real-world A/B testing costs.

Retrieval-Augmented Generation (RAG) for RecSys

  • Issue: FMs may have outdated knowledge, generate incorrect information, or lack domain expertise, which can negatively impact the quality of recommendations.

  • Resolution: Integrate RAG techniques into FM4RecSys to enhance the generative capability by incorporating external data retrieval into the generative process. Use RAG to selectively extract relevant portions of a user's interaction history and associated external knowledge, ensuring up-to-date and reliable recommendations.

System Performance Analysis

  • Issue: The development of FM-based recommendation systems involves costs related to training, inference, and API usage, which can escalate with more complex or extensive usage.

  • Resolution: Employ targeted solutions such as data selection for pre-training and fine-tuning, parameter-efficient fine-tuning methods (e.g., LoRA), embedding caching, lightweight FMs (e.g., distillation, pruning, quantization), prompt selection, and adaptive RAG to reduce training costs, inference latency, and API costs.

Benchmarks and Evaluation Metrics

  • Issue: The current benchmarks for FM4RecSys are limited in terms of datasets, recommendation tasks, and evaluation metrics, hindering comprehensive evaluation and comparison of different models.

  • Resolution: Create a holistic and diversified benchmark that includes a variety of datasets, diverse recommendation tasks, and adaptable metrics for different models. Devise new benchmarks and evaluation metrics specifically for multi-modal and personalized agent FMs in recommendation scenarios.

Emerging Trends

  • Issue: FMs in RSs face challenges related to safety (e.g., red teaming attacks, undesirable content generation) and privacy (e.g., potential access to sensitive user data through prompt injection).

  • Resolution: Align FMs with human values by gathering relevant negative data and employing supervised fine-tuning techniques like online and offline human preference training. Incorporate approaches such as federated learning and machine unlearning into FM4RecSys to enhance user privacy protection.

By addressing these key issues and implementing the proposed resolutions, researchers and practitioners can further advance the field of FM4RecSys, improving the effectiveness, explainability, trustworthiness, and efficiency of foundation model-based recommender systems.]

Conclusion

The field of Foundation Model-based Recommender Systems (FM4RecSys) has shown significant promise in enhancing the capabilities and performance of recommendation systems.

By leveraging the power of large language models and other foundation models, FM4RecSys can provide more personalised, explainable, and context-aware recommendations, even in scenarios with limited data.

However, the development and deployment of FM4RecSys come with their own set of challenges, such as handling long sequences, ensuring explainability and trustworthiness, and addressing privacy and safety concerns.

By focusing on these open problems and opportunities, researchers and practitioners can further advance the field of FM4RecSys and create more effective, efficient, and user-centric recommendation systems that cater to the diverse needs of businesses and consumers alike.

The most promising commercial applications of FM4RecSys likely lie in domains where personalisation, explainability, and cross-domain knowledge integration are crucial, such as e-commerce, content streaming, and digital marketing.

References

Here is a summary of the mentioned papers including their names, topics, authors, and publication years:

Abi-step Grounding Paradigm for Large Language Models in Recommendation Systems

  • Topic: The paper proposes a new paradigm to integrate large language models effectively in recommendation systems.

  • Authors: Keqin Bao, Jizhi Zhang, Wenjie Wang, Yang Zhang, Zhengyi Yang, Yancheng Luo, Fuli Feng, Xiangnan He, Qi Tian

  • Year: 2023

TallRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation

  • Topic: This study introduces an efficient tuning framework called TallRec, which aligns large language models to enhance recommendation systems.

  • Authors: Keqin Bao, Jizhi Zhang, Yang Zhang, Wenjie Wang, Fuli Feng, Xiangnan He

  • Year: 2023

Large Language Models for Recommendation: Progresses and Future Directions

  • Topic: A review of the current state and future directions of using large language models in recommendation systems.

  • Authors: Keqin Bao, Jizhi Zhang, Yang Zhang, Wang Wenjie, Fuli Feng, Xiangnan He

  • Year: 2023

Longformer: The Long-Document Transformer

  • Topic: Introduction of Longformer, a transformer model designed to efficiently process longer documents.

  • Authors: Iz Beltagy, Matthew E. Peters, Arman Cohan

  • Year: 2020

On the Opportunities and Risks of Foundation Models

  • Topic: An extensive discussion on the potential benefits and risks associated with foundation models in AI.

  • Authors: Rishi Bommasani and colleagues

  • Year: 2021

Language Models are Few-Shot Learners

  • Topic: An exploration of the capabilities of language models to perform tasks with minimal task-specific data.

  • Authors: Tom Brown and colleagues

  • Year: 2020

Instruction Mining: High-Quality Instruction Data Selection for Large Language Models

  • Topic: The paper focuses on selecting high-quality instructional data to improve the performance of large language models.

  • Authors: Yihan Cao, Yanbin Kang, Lichao Sun

  • Year: 2023

The Lottery Ticket Hypothesis for Pretrained BERT Networks

  • Topic: Investigates the "lottery ticket hypothesis" in the context of pretrained BERT networks, suggesting that certain subnetworks are especially effective.

  • Authors: Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu, Yang Zhang, Zhangyang Wang, Michael Carbin

  • Year: 2020

Recommendation Unlearning

  • Topic: Discusses methods for 'unlearning' in recommendation systems to potentially improve model adaptability and data privacy.

  • Authors: Chong Chen, Fei Sun, Min Zhang, Bolin Ding

  • Year: 2022

Maybe Only 0.5% Data is Needed: A Preliminary Exploration of Low Training Data Instruction Tuning

  • Topic: Explores the efficiency of using significantly reduced datasets for tuning large language models.

  • Authors: Hao Chen, Yiming Zhang, Qi Zhang, Hantao Yang, Xiaomeng Hu, Xuetao Ma, Yifan Yanggong, Junbo Zhao

  • Year: 2023

M6-Rec: Generative Pretrained Language Models are Open-Ended Recommender Systems

  • Topic: Introduces M6-Rec, a framework that utilizes generative pretrained language models as flexible, open-ended recommender systems.

  • Authors: Zeyu Cui, Jianxin Ma, Chang Zhou, Jingren Zhou, Hongxia Yang

  • Year: 2022

Uncovering ChatGPT’s Capabilities in Recommender Systems

  • Topic: An investigation into how ChatGPT can be utilized to enhance the functionality of recommender systems.

  • Authors: Sunhao Dai and colleagues

  • Year: 2023

An Unified Search and Recommendation Foundation Model for Cold-Start Scenario

  • Topic: This paper presents a foundation model designed to address the cold-start problem in search and recommendation systems.

  • Authors: Yuqi Gong, Xichen Ding, Yehui Su, Kaiming Shen, Zhongyi Liu, Guannan Zhang

  • Year: 2023

Large Language Models are Zero-Shot Time Series Forecasters

  • Topic: Exploration of the capability of large language models to perform zero-shot forecasting for time series data.

  • Authors: Nate Gruver, Marc Anton Finzi, Shikai Qiu, Andrew Gordon Wilson

  • Year: 2023

Leveraging Large Language Models for Sequential Recommendation

  • Topic: Investigates how large language models can be utilized for making sequential recommendations.

  • Authors: Jesse Harte, Wouter Zorgdrager, Panos Louridas, Asterios Katsifodimos, Dietmar Jannach, Marios Fragkoulis

  • Year: 2023

Large Language Models as Zero-Shot Conversational Recommenders

  • Topic: Discusses the use of large language models for conversational recommendation without specific training on the task.

  • Authors: Zhankui He, Zhouhang Xie, Rahul Jha, Harald Steck, Dawen Liang, Yesu Feng, Bodhisattwa Prasad Majumder, Nathan Kallus, Julian J. McAuley

  • Year: 2023

Learning Vector-Quantized Item Representation for Transferable Sequential Recommenders

  • Topic: The study focuses on developing transferable sequential recommendation systems using vector-quantized item representation.

  • Authors: Yupeng Hou, Zhankui He, Julian McAuley, Wayne Xin Zhao

  • Year: 2023

Large Language Models are Zero-Shot Rankers for Recommender Systems

  • Topic: Examines the potential of large language models to function as zero-shot rankers in recommender systems.

  • Authors: Yupeng Hou, Junjie Zhang, Zihan Lin, Hongyu Lu, Ruobing Xie, Julian McAuley, Wayne Xin Zhao

  • Year: 2023

LoRA: Low-Rank Adaptation of Large Language Models

  • Topic: Introduces a method for adapting large language models using low-rank matrices to improve efficiency and performance.

  • Authors: Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, Weizhu Chen

  • Year: 2022

In-Context Analogical Reasoning with Pre-Trained Language Models

  • Topic: This paper explores how pre-trained language models can be used for analogical reasoning within a given context.

  • Authors: Xiaoyang Hu, Shane Storks, Richard L. Lewis, Joyce Chai

  • Year: 2023

UP5: Unbiased Foundation Model for Fairness-Aware Recommendation

  • Topic: Proposes a new foundation model aimed at ensuring fairness in recommendation systems by addressing bias.

  • Authors: Wenyue Hua, Yingqiang Ge, Shuyuan Xu, Jianchao Ji, Yongfeng Zhang

  • Year: 2023

Tutorial on Large Language Models for Recommendation

  • Topic: A comprehensive tutorial on employing large language models for various recommendation tasks.

  • Authors: Wenyue Hua, Lei Li, Shuyuan Xu, Li Chen, Yongfeng Zhang

  • Year: 2023

How to Index Item IDs for Recommendation Foundation Models

  • Topic: Discusses techniques for effectively indexing item IDs to optimize foundation models for recommendation.

  • Authors: Wenyue Hua, Shuyuan Xu, Yingqiang Ge, Yongfeng Zhang

  • Year: 2023

Recommender AI Agent: Integrating Large Language Models for Interactive Recommendations

  • Topic: Introduces an AI agent that integrates large language models to provide interactive and dynamic recommendations.

  • Authors: Xu Huang, Jianxun Lian, Yuxuan Lei, Jing Yao, Defu Lian, Xing Xie

  • Year: 2023

Here is the continuation of the summary for the listed papers:

LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models

  • Topic: Presents LoftQ, a quantization approach that considers fine-tuning aspects of LoRA to enhance the performance of large language models.

  • Authors: Yixiao Li, Yifan Yu, Chen Liang, Pengcheng He, Nikos Karampatziakis, Weizhu Chen, Tuo Zhao

  • Year: 2023

LLARA: Aligning Large Language Models with Sequential Recommenders

  • Topic: Introduces LLARA, a methodology for integrating large language models with sequential recommendation systems to improve alignment.

  • Authors: Jiayi Liao, Sihang Li, Zhengyi Yang, Jiancan Wu, Yancheng Yuan, Xiang Wang

  • Year: 2023

Sparks of Artificial General Recommender (AGR): Early Experiments with ChatGPT

  • Topic: Discusses early experimental results using ChatGPT in developing an Artificial General Recommender system.

  • Authors: Guo Lin, Yongfeng Zhang

  • Year: 2023

AWQ: Activation-Aware Weight Quantization for LLM Compression and Acceleration

  • Topic: Explores AWQ, a technique for quantizing the weights of large language models based on their activation characteristics to optimize performance.

  • Authors: Ji Lin, Jiaming Tang, Haotian Tang, Shang Yang, Xingyu Dang, Song Han

  • Year: 2023

How Can Recommender Systems Benefit from Large Language Models: A Survey

  • Topic: A comprehensive survey on the integration of large language models into recommender systems, highlighting benefits and methodologies.

  • Authors: Jianghao Lin, Xinyi Dai, Yunjia Xi, Weiwen Liu, Bo Chen, Xiangyang Li, Chenxu Zhu, Huifeng Guo, Yong Yu, Ruiming Tang, Weinan Zhang

  • Year: 2023

Rella: Retrieval-Enhanced Large Language Models for Lifelong Sequential Behavior Comprehension in Recommendation

  • Topic: Proposes Rella, a retrieval-enhanced approach for large language models aimed at comprehending sequential behaviors across a user's lifetime for recommendations.

  • Authors: Jianghao Lin, Rong Shan, Chenxu Zhu, Kounianhua Du, Bo Chen, Shigang Quan, Ruiming Tang, Yong Yu, Weinan Zhang

  • Year: 2023

A Multi-Facet Paradigm to Bridge Large Language Model and Recommendation

  • Topic: Introduces a multi-facet paradigm that bridges the gap between large language models and recommendation systems, focusing on various integration aspects.

  • Authors: Xinyu Lin, Wenjie Wang, Yongqi Li, Fuli Feng, See-Kiong Ng, Tat-Seng Chua

  • Year: 2023

Is ChatGPT a Good Recommender? A Preliminary Study

  • Topic: This preliminary study investigates the effectiveness of ChatGPT as a recommender system.

  • Authors: Junling Liu, Chao Liu, Renjie Lv, Kang Zhou, Yan Zhang

  • Year: 2023

LLMRec: Benchmarking Large Language Models on Recommendation Task

  • Topic: LLMRec benchmarks the performance of large language models on recommendation tasks, providing insights into their capabilities and limitations.

  • Authors: Junling Liu, Chao Liu, Peilin Zhou, Qichen Ye, Dading Chong, Kang Zhou, Yueqi Xie, Yuwei Cao, Shoujin Wang, Chenyu You, Philip S. Yu

  • Year: 2023

Pre-Train, Prompt, and Recommendation: A Comprehensive Survey of Language Modelling Paradigm Adaptations in Recommender Systems

  • Topic: This survey reviews how language modeling paradigms, particularly pre-training and prompting, are adapted for use in recommender systems.

  • Authors: Peng Liu, Lemei Zhang, Jon Atle Gulla

  • Year: 2023

MissRec: Pre-training and Transferring Multi-modal Interest-Aware Sequence Representation for Recommendation

  • Topic: This paper presents MissRec, a model that pre-trains and transfers multi-modal, interest-aware sequence representations for improving recommendation systems.

  • Authors: Jinpeng Wang, Ziyun Zeng, Yunxiao Wang, Yuting Wang, Xingyu Lu, Tianxiang Li, Jun Yuan, Rui Zhang, Hai-Tao Zheng, Shu-Tao Xia

  • Year: 2023

RecAgent: A Novel Simulation Paradigm for Recommender Systems

  • Topic: Introduces RecAgent, a novel simulation-based framework to evaluate and enhance the performance of recommender systems.

  • Authors: Lei Wang, Jingsen Zhang, Xu Chen, Yankai Lin, Ruihua Song, Wayne Xin Zhao, Ji-Rong Wen

  • Year: 2023

Rethinking the Evaluation for Conversational Recommendation in the Era of Large Language Models

  • Topic: Explores new methodologies for evaluating conversational recommender systems in light of advancements in large language models.

  • Authors: Xiaolei Wang, Xinyu Tang, Wayne Xin Zhao, Jingyuan Wang, Ji-Rong Wen

  • Year: 2023

Enhancing Recommender Systems with Large Language Model Reasoning Graphs

  • Topic: Discusses the integration of reasoning graphs derived from large language models into recommender systems to enhance performance.

  • Authors: Yan Wang, Zhixuan Chu, Xin Ouyang, Simeng Wang, Hongyan Hao, Yue Shen, Jinjie Gu, Siqiao Xue, James Y. Zhang, Qing Cui, Longfei Li, Jun Zhou, Sheng Li

  • Year: 2023

RecMind: Large Language Model Powered Agent for Recommendation

  • Topic: Describes RecMind, an agent powered by large language models to provide dynamic and effective recommendation services.

  • Authors: Yancheng Wang, Ziyan Jiang, Zheng Chen, Fan Yang, Yingxue Zhou, Eunah Cho, Xing Fan, Xiaojiang Huang, Yanbin Lu, Yingzhen Yang

  • Year: 2023

DRDT: Dynamic Reflection with Divergent Thinking for LLM-Based Sequential Recommendation

  • Topic: Introduces DRDT, a novel approach combining dynamic reflection and divergent thinking strategies for enhancing sequential recommendations using large language models.

  • Authors: Yu Wang, Zhiwei Liu, Jianguo Zhang, Weiran Yao, Shelby Heinecke, Philip S. Yu

  • Year: 2023

Aligning Large Language Models with Human: A Survey

  • Topic: This survey reviews the strategies and methodologies for aligning large language models with human-like reasoning and decision-making processes.

  • Authors: Yufei Wang, Wanjun Zhong, Liangyou Li, Fei Mi, Xingshan Zeng, Wenyong Huang, Lifeng Shang, Xin Jiang, Qun Liu

  • Year: 2023

PTUM: Pretraining User Model from Unlabeled User Behaviors via Self-Supervision

  • Topic: Discusses PTUM, a method for pretraining user models using self-supervised learning from unlabeled user behavior data.

  • Authors: Chuhan Wu, Fangzhao Wu, Tao Qi, Jianxun Lian, Yongfeng Huang, Xing Xie

  • Year: 2020

Personalized Prompts for Sequential Recommendation

  • Topic: Explores the use of personalized prompts to improve the efficacy of sequential recommendation systems.

  • Authors: Yiqing Wu, Ruobing Xie, Yongchun Zhu, Fuzhen Zhuang, Xu Zhang, Leyu Lin, Qing He

  • Year: 2022

A Survey on Large Language Models for Recommendation

  • Topic: Provides a comprehensive survey on the application of large language models in recommendation systems, highlighting key advancements and future directions.

  • Authors: Likang Wu, Zhi Zheng, Zhaopeng Qiu, Hao Wang, Hongchao Gu, Tingjia Shen, Chuan Qin, Chen Zhu, Hengshu Zhu, Qi Liu, Hui Xiong, Enhong Chen

  • Year: 2023

LogoFoundation Models for Recommender Systems: A Survey and New PerspectivesarXiv.org
The taxonomy of FM4RecSys from data characteristics to open problems and opportunities
Page cover image