Large Language Models (LLMs) are often deployed as advisors to consumers, e.g., recommending purchases, and to managers, e.g., suggesting new hires. In fact, LLMs provide advice based on cost or convenience, overlooking broader societal impacts (e.g., carbon footprint when recommending products to a potential customer, or fairness when recommending a new hire to a company). To align LLMs' advice with societal goals like environmental sustainability and gender parity, tuning strategies must integrate the notion of common good. We discuss why Direct Alignment tuning could be preferable to classic Reinforcement Learning from Human Feedback to achieve this integration. Then, we describe and compare two approaches to Direct Preference Optimization: (1) exposing the model tuning examples taken from recommendations and regulations, and (2) mythopoiesis, i.e., model tuning based on synthetic ``legends'', fictional success stories of regulatory compliance (also generated by LLMs). We present a pipeline to evaluate legends' effectiveness in reducing bias and fostering compliance. Our preliminary results suggest that legend-based tuning may enhance engagement and generalization, while direct exposure ensures factual accuracy but risks rigidity.

Tuning LLM-Based Advisors for the Common Good: The Case for Direct Preference Optimization / L. Mauri, G. Sargsyan, E. Damiani - In: FLLM[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2026. - ISBN 979-8-3315-9410-7. - pp. 910-915 (( 3. International Conference on Foundation and Large Language Models : November, 25 - 28 Vienna 2025 [10.1109/fllm67465.2025.11391029].

Tuning LLM-Based Advisors for the Common Good: The Case for Direct Preference Optimization

L. Mauri
Primo
;
E. Damiani
Ultimo
2026

Abstract

Large Language Models (LLMs) are often deployed as advisors to consumers, e.g., recommending purchases, and to managers, e.g., suggesting new hires. In fact, LLMs provide advice based on cost or convenience, overlooking broader societal impacts (e.g., carbon footprint when recommending products to a potential customer, or fairness when recommending a new hire to a company). To align LLMs' advice with societal goals like environmental sustainability and gender parity, tuning strategies must integrate the notion of common good. We discuss why Direct Alignment tuning could be preferable to classic Reinforcement Learning from Human Feedback to achieve this integration. Then, we describe and compare two approaches to Direct Preference Optimization: (1) exposing the model tuning examples taken from recommendations and regulations, and (2) mythopoiesis, i.e., model tuning based on synthetic ``legends'', fictional success stories of regulatory compliance (also generated by LLMs). We present a pipeline to evaluate legends' effectiveness in reducing bias and fostering compliance. Our preliminary results suggest that legend-based tuning may enhance engagement and generalization, while direct exposure ensures factual accuracy but risks rigidity.
large language models; legend-based tuning; AI alignment; regulatory compliance; benchmarking LLMs;
Settore INFO-01/A - Informatica
   MUSA - Multilayered Urban Sustainability Actiona
   MUSA
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA

   Sovereign Edge-Hub: un’architettura cloud-edge per la sovranità digitale nelle scienze della vita (SOV-EDGE-HUB)Linea Strategica 4 - Sicurezza informatica/Cloud
   SOV-EDGE-HUB
   UNIVERSITA' DEGLI STUDI DI MILANO
2026
Institute of Electrical and Electronics Engineers (IEEE)
Book Part (author)
File in questo prodotto:
File Dimensione Formato  
paper_FLLM2025.pdf

accesso aperto

Tipologia: Post-print, accepted manuscript ecc. (versione accettata dall'editore)
Licenza: Creative commons
Dimensione 1.09 MB
Formato Adobe PDF
1.09 MB Adobe PDF Visualizza/Apri
Tuning_LLM-Based_Advisors_for_the_Common_Good_The_Case_for_Direct_Preference_Optimization(2).pdf

accesso riservato

Tipologia: Publisher's version/PDF
Licenza: Nessuna licenza
Dimensione 1.16 MB
Formato Adobe PDF
1.16 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2434/1233599
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact