MIMIR: A Streamlined Platform for Personalized Agent Tuning in Domain Expertise
- URL: http://arxiv.org/abs/2404.04285v1
- Date: Wed, 3 Apr 2024 23:42:38 GMT
- Title: MIMIR: A Streamlined Platform for Personalized Agent Tuning in Domain Expertise
- Authors: Chunyuan Deng, Xiangru Tang, Yilun Zhao, Hanming Wang, Haoran Wang, Wangchunshu Zhou, Arman Cohan, Mark Gerstein,
- Abstract summary: textscMimir is a streamlined platform offering a customizable pipeline for personalized agent tuning.
textscMimir supports the generation of general instruction-tuning datasets from the same input.
textscMimir integrates these features into a cohesive end-to-end platform, facilitating everything from the uploading of personalized files to one-click agent fine-tuning.
- Score: 49.83486066403154
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, large language models (LLMs) have evolved into interactive agents, proficient in planning, tool use, and task execution across a wide variety of tasks. However, without specific agent tuning, open-source models like LLaMA currently struggle to match the efficiency of GPT- 4, particularly given the scarcity of agent-tuning datasets for fine-tuning. In response, we introduce \textsc{Mimir}: a streamlined platform offering a customizable pipeline that enables users to leverage both private knowledge and publicly available, legally compliant datasets at scale for \textbf{personalized agent tuning}. Additionally, \textsc{Mimir} supports the generation of general instruction-tuning datasets from the same input. This dual capability ensures that language agents developed through the platform possess both specific agent abilities and general competencies. \textsc{Mimir} integrates these features into a cohesive end-to-end platform, facilitating everything from the uploading of personalized files to one-click agent fine-tuning.
Related papers
- CRAB: Cross-environment Agent Benchmark for Multimodal Language Model Agents [52.83132876539399]
Crab is the first benchmark framework designed to support cross-environment tasks.
Our framework supports multiple devices and can be easily extended to any environment with a Python interface.
The experimental results demonstrate that the single agent with GPT-4o achieves the best completion ratio of 35.
arXiv Detail & Related papers (2024-07-01T17:55:04Z) - LEGENT: Open Platform for Embodied Agents [60.71847900126832]
We introduce LEGENT, an open, scalable platform for developing embodied agents using Large Language Models (LLMs) and Large Multimodal Models (LMMs)
LEGENT offers a rich, interactive 3D environment with communicable and actionable agents, paired with a user-friendly interface.
In experiments, an embryonic vision-language-action model trained on LEGENT-generated data surpasses GPT-4V in embodied tasks.
arXiv Detail & Related papers (2024-04-28T16:50:12Z) - Personalized Federated Instruction Tuning via Neural Architecture Search [25.19100691411643]
Federated Instruction Tuning (FIT) has shown the ability to achieve collaborative model instruction tuning among massive data owners without sharing private data.
Due to the varying data distribution and preferences among data owners, FIT cannot adapt to the personalized data of individual owners.
We propose a novel Personalized Federated Instruction Tuning (PerFIT) framework based on architecture search.
arXiv Detail & Related papers (2024-02-26T06:29:05Z) - AgentScope: A Flexible yet Robust Multi-Agent Platform [66.64116117163755]
AgentScope is a developer-centric multi-agent platform with message exchange as its core communication mechanism.
The abundant syntactic tools, built-in agents and service functions, user-friendly interfaces for application demonstration and utility monitor, zero-code programming workstation, and automatic prompt tuning mechanism significantly lower the barriers to both development and deployment.
arXiv Detail & Related papers (2024-02-21T04:11:28Z) - Communication-Efficient Personalized Federated Learning for
Speech-to-Text Tasks [66.78640306687227]
To protect privacy and meet legal regulations, federated learning (FL) has gained significant attention for training speech-to-text (S2T) systems.
The commonly used FL approach (i.e., textscFedAvg) in S2T tasks typically suffers from extensive communication overhead.
We propose a personalized federated S2T framework that introduces textscFedLoRA, a lightweight LoRA module for client-side tuning and interaction with the server, and textscFedMem, a global model equipped with a $k$-near
arXiv Detail & Related papers (2024-01-18T15:39:38Z) - PRODIGy: a PROfile-based DIalogue Generation dataset [14.123548564209068]
We propose a new resource where each dialogue is aligned with all possible speaker representations such as communication style, biographies, and personality.
This framework allows to test several baselines built using generative language models with several profile configurations.
arXiv Detail & Related papers (2023-11-09T08:19:34Z) - Recommender AI Agent: Integrating Large Language Models for Interactive
Recommendations [53.76682562935373]
We introduce an efficient framework called textbfInteRecAgent, which employs LLMs as the brain and recommender models as tools.
InteRecAgent achieves satisfying performance as a conversational recommender system, outperforming general-purpose LLMs.
arXiv Detail & Related papers (2023-08-31T07:36:44Z) - Gentopia: A Collaborative Platform for Tool-Augmented LLMs [21.09079715807735]
We present gentopia, an Augmented Language Models (ALMs) framework enabling flexible customization of agents through simple configurations.
We also establish gentpool, a public platform enabling the registration and sharing of user-customized agents.
gentbench, an integral component of gentpool, is designed to thoroughly evaluate user-customized agents across diverse aspects such as safety, robustness, efficiency, etc.
arXiv Detail & Related papers (2023-08-08T04:12:29Z) - ELIT: Emory Language and Information Toolkit [15.340540198612826]
ELIT is a comprehensive framework providing transformer-based end-to-end models for core tasks.
ELIT features an efficient Multi-Task Learning (MTL) model with many downstream tasks that include lemmatization, part-of-speech tagging, named entity recognition, dependency parsing, constituency parsing, semantic role labeling, and AMR parsing.
arXiv Detail & Related papers (2021-09-08T19:50:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.