Kubishi Research Group

Backlinks

  • Welcome
Home

❯

Publications

❯

Crowd SFT: Crowdsourcing for LLM Alignment

Crowd-SFT: Crowdsourcing for LLM Alignment

  • decentralized-computing
  • distributed-ledger-technology
  • artificial-intelligence
  • large-language-models
Alex Sotiropoulos, Linus Lei, Jared Coleman, Bhaskar Krishnamachari, Sulyab Thottungal Valapu
DAPPS 2025 - The 7th IEEE International Conference on Decentralized Applications and Infrastructures
July 24, 2025
10.48550/arXiv.2506.04063
PDF

Abstract

Large Language Models (LLMs) increasingly rely on Supervised Fine-Tuning (SFT) and Reinforcement Learning from Human Feedback (RLHF) to align model responses with human preferences. While RLHF employs a reinforcement learning approach with a separate reward model, SFT uses human-curated datasets for supervised learning. Both approaches traditionally depend on small, vetted groups of annotators, making them costly, prone to bias, and limited in scalability. We propose an open, crowd-sourced fine-tuning framework that addresses these limitations by enabling broader feedback collection for SFT without extensive annotator training. Our framework promotes incentive fairness via a point-based reward system correlated with Shapley values and guides model convergence through iterative model updates. Our multi-model selection framework demonstrates up to a 55% reduction in target distance over single-model selection, enabling subsequent experiments that validate our point-based reward mechanism's close alignment with Shapley values (a well-established method for attributing individual contributions) thereby supporting fair and scalable participation.


Created with Quartz v4.5.2 © 2025

  • GitHub
  • Discord Community