Comparative Analysis of Multi-Agent Reinforcement Learning Policies for Crop Planning Decision Support
- URL: http://arxiv.org/abs/2412.02057v1
- Date: Tue, 03 Dec 2024 00:30:19 GMT
- Title: Comparative Analysis of Multi-Agent Reinforcement Learning Policies for Crop Planning Decision Support
- Authors: Anubha Mahajan, Shreya Hegde, Ethan Shay, Daniel Wu, Aviva Prins,
- Abstract summary: In India, majority of farmers are classified as small or marginal, making their livelihoods particularly vulnerable to economic losses due to market saturation and climate risks.
Existing decision support systems (DSS) often provide generic recommendations that fail to account for real-time market dynamics and the interactions among multiple farmers.
In this paper, we evaluate the viability of three multi-agent reinforcement learning (MARL) approaches for optimizing total farmer income and promoting fairness in crop planning.
- Score: 0.873811641236639
- License:
- Abstract: In India, the majority of farmers are classified as small or marginal, making their livelihoods particularly vulnerable to economic losses due to market saturation and climate risks. Effective crop planning can significantly impact their expected income, yet existing decision support systems (DSS) often provide generic recommendations that fail to account for real-time market dynamics and the interactions among multiple farmers. In this paper, we evaluate the viability of three multi-agent reinforcement learning (MARL) approaches for optimizing total farmer income and promoting fairness in crop planning: Independent Q-Learning (IQL), where each farmer acts independently without coordination, Agent-by-Agent (ABA), which sequentially optimizes each farmer's policy in relation to the others, and the Multi-agent Rollout Policy, which jointly optimizes all farmers' actions for global reward maximization. Our results demonstrate that while IQL offers computational efficiency with linear runtime, it struggles with coordination among agents, leading to lower total rewards and an unequal distribution of income. Conversely, the Multi-agent Rollout policy achieves the highest total rewards and promotes equitable income distribution among farmers but requires significantly more computational resources, making it less practical for large numbers of agents. ABA strikes a balance between runtime efficiency and reward optimization, offering reasonable total rewards with acceptable fairness and scalability. These findings highlight the importance of selecting appropriate MARL approaches in DSS to provide personalized and equitable crop planning recommendations, advancing the development of more adaptive and farmer-centric agricultural decision-making systems.
Related papers
- Fairness Aware Reinforcement Learning via Proximal Policy Optimization [7.061167083587786]
This paper introduces fairness in Proximal Policy Optimization (PPO) with a penalty term derived from demographic parity, counterfactual fairness, and conditional statistical parity.
We evaluate our approach in the Allelopathic Harvest game, a cooperative and competitive MAS focused on resource collection.
arXiv Detail & Related papers (2025-02-06T10:45:55Z) - From Novice to Expert: LLM Agent Policy Optimization via Step-wise Reinforcement Learning [62.54484062185869]
We introduce StepAgent, which utilizes step-wise reward to optimize the agent's reinforcement learning process.
We propose implicit-reward and inverse reinforcement learning techniques to facilitate agent reflection and policy adjustment.
arXiv Detail & Related papers (2024-11-06T10:35:11Z) - Cooperation and Fairness in Multi-Agent Reinforcement Learning [6.164771707307928]
In resource-constrained environments of mobility and transportation systems, efficiency may be achieved at the expense of fairness.
We consider the problem of fair multi-agent navigation for a group of decentralized agents using multi-agent reinforcement learning (MARL)
We find that our model yields a 14% improvement in efficiency and a 5% improvement in fairness over a baseline trained using random assignments.
arXiv Detail & Related papers (2024-10-19T00:10:52Z) - Fair Allocation in Dynamic Mechanism Design [57.66441610380448]
We consider a problem where an auctioneer sells an indivisible good to groups of buyers in every round, for a total of $T$ rounds.
The auctioneer aims to maximize their discounted overall revenue while adhering to a fairness constraint that guarantees a minimum average allocation for each group.
arXiv Detail & Related papers (2024-05-31T19:26:05Z) - Principal-Agent Reward Shaping in MDPs [50.914110302917756]
Principal-agent problems arise when one party acts on behalf of another, leading to conflicts of interest.
We study a two-player Stack game where the principal and the agent have different reward functions, and the agent chooses an MDP policy for both players.
Our results establish trees and deterministic decision processes with a finite horizon.
arXiv Detail & Related papers (2023-12-30T18:30:44Z) - An Online Optimization-Based Decision Support Tool for Small Farmers in
India: Learning in Non-stationary Environments [1.3597551064547502]
Small farmers in India, who could greatly benefit from these tools, do not have access to them.
In this paper, we model an individual greenhouse as a Markov Decision Process (MDP) and adapt Li and Li's Follow the Leader (FWL) online learning algorithm to offer crop planning advice.
arXiv Detail & Related papers (2023-11-28T23:33:16Z) - Quantifying Agent Interaction in Multi-agent Reinforcement Learning for
Cost-efficient Generalization [63.554226552130054]
Generalization poses a significant challenge in Multi-agent Reinforcement Learning (MARL)
The extent to which an agent is influenced by unseen co-players depends on the agent's policy and the specific scenario.
We present the Level of Influence (LoI), a metric quantifying the interaction intensity among agents within a given scenario and environment.
arXiv Detail & Related papers (2023-10-11T06:09:26Z) - MESOB: Balancing Equilibria & Social Optimality [12.702156510015628]
Motivated by bid recommendation in online ad auctions, this paper considers a class of multi-level and multi-agent games.
We propose a novel and tractable bi-objective optimization formulation with mean-field approximation.
MESOB-OMO enables obtaining approximately efficient solutions in terms of the dual objectives of competition and cooperation.
arXiv Detail & Related papers (2023-07-16T00:43:54Z) - Finding Regularized Competitive Equilibria of Heterogeneous Agent
Macroeconomic Models with Reinforcement Learning [151.03738099494765]
We study a heterogeneous agent macroeconomic model with an infinite number of households and firms competing in a labor market.
We propose a data-driven reinforcement learning framework that finds the regularized competitive equilibrium of the model.
arXiv Detail & Related papers (2023-02-24T17:16:27Z) - Distributional Reward Estimation for Effective Multi-Agent Deep
Reinforcement Learning [19.788336796981685]
We propose a novel Distributional Reward Estimation framework for effective Multi-Agent Reinforcement Learning (DRE-MARL)
Our main idea is to design the multi-action-branch reward estimation and policy-weighted reward aggregation for stabilized training.
The superiority of the DRE-MARL is demonstrated using benchmark multi-agent scenarios, compared with the SOTA baselines in terms of both effectiveness and robustness.
arXiv Detail & Related papers (2022-10-14T08:31:45Z) - Monotonic Improvement Guarantees under Non-stationarity for
Decentralized PPO [66.5384483339413]
We present a new monotonic improvement guarantee for optimizing decentralized policies in cooperative Multi-Agent Reinforcement Learning (MARL)
We show that a trust region constraint can be effectively enforced in a principled way by bounding independent ratios based on the number of agents in training.
arXiv Detail & Related papers (2022-01-31T20:39:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.