Large Language Models (LLMs) are often deployed as advisors to consumers, e.g., recommending purchases, and to managers, e.g., suggesting new hires. In fact, LLMs provide advice based on cost or convenience, overlooking broader societal impacts (e.g., carbon footprint when recommending products to a potential customer, or fairness when recommending a new hire to a company). To align LLMs' advice with societal goals like environmental sustainability and gender parity, tuning strategies must integrate the notion of common good. We discuss why Direct Alignment tuning could be preferable to classic Reinforcement Learning from Human Feedback to achieve this integration. Then, we describe and compare two approaches to Direct Preference Optimization: (1) exposing the model tuning examples taken from recommendations and regulations, and (2) mythopoiesis, i.e., model tuning based on synthetic ``legends'', fictional success stories of regulatory compliance (also generated by LLMs). We present a pipeline to evaluate legends' effectiveness in reducing bias and fostering compliance. Our preliminary results suggest that legend-based tuning may enhance engagement and generalization, while direct exposure ensures factual accuracy but risks rigidity.
Tuning LLM-Based Advisors for the Common Good: The Case for Direct Preference Optimization / L. Mauri, G. Sargsyan, E. Damiani - In: FLLM[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2026. - ISBN 979-8-3315-9410-7. - pp. 910-915 (( 3. International Conference on Foundation and Large Language Models : November, 25 - 28 Vienna 2025 [10.1109/fllm67465.2025.11391029].
Tuning LLM-Based Advisors for the Common Good: The Case for Direct Preference Optimization
L. Mauri
Primo
;E. DamianiUltimo
2026
Abstract
Large Language Models (LLMs) are often deployed as advisors to consumers, e.g., recommending purchases, and to managers, e.g., suggesting new hires. In fact, LLMs provide advice based on cost or convenience, overlooking broader societal impacts (e.g., carbon footprint when recommending products to a potential customer, or fairness when recommending a new hire to a company). To align LLMs' advice with societal goals like environmental sustainability and gender parity, tuning strategies must integrate the notion of common good. We discuss why Direct Alignment tuning could be preferable to classic Reinforcement Learning from Human Feedback to achieve this integration. Then, we describe and compare two approaches to Direct Preference Optimization: (1) exposing the model tuning examples taken from recommendations and regulations, and (2) mythopoiesis, i.e., model tuning based on synthetic ``legends'', fictional success stories of regulatory compliance (also generated by LLMs). We present a pipeline to evaluate legends' effectiveness in reducing bias and fostering compliance. Our preliminary results suggest that legend-based tuning may enhance engagement and generalization, while direct exposure ensures factual accuracy but risks rigidity.| File | Dimensione | Formato | |
|---|---|---|---|
|
paper_FLLM2025.pdf
accesso aperto
Tipologia:
Post-print, accepted manuscript ecc. (versione accettata dall'editore)
Licenza:
Creative commons
Dimensione
1.09 MB
Formato
Adobe PDF
|
1.09 MB | Adobe PDF | Visualizza/Apri |
|
Tuning_LLM-Based_Advisors_for_the_Common_Good_The_Case_for_Direct_Preference_Optimization(2).pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Licenza:
Nessuna licenza
Dimensione
1.16 MB
Formato
Adobe PDF
|
1.16 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




