Recommendations to overcome language barriers in the Vera C. Rubin Observatory Research Ecosystem
- URL: http://arxiv.org/abs/2507.18682v1
- Date: Thu, 24 Jul 2025 16:34:26 GMT
- Title: Recommendations to overcome language barriers in the Vera C. Rubin Observatory Research Ecosystem
- Authors: José Antonio Alonso Pavón, Andrés Alejandro Plazas Malagón,
- Abstract summary: Report presents recommendations to reduce language barriers within the Vera C. Rubin Observatory Research Ecosystem.<n>English linguistic hegemony in science limits participation and productivity.<n> multilingual presentation formats, academic writing training, a Virtual Writing Center, language support programs, and writing retreats.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The report presents a comprehensive set of five recommendations to reduce language barriers within the Vera C. Rubin Observatory Research Ecosystem, promoting greater inclusion of researchers who are speakers of English as an additional language. Recognizing that English linguistic hegemony in science limits participation and productivity, the document proposes multilingual presentation formats, academic writing training, a Virtual Writing Center, language support programs, and writing retreats. Each recommendation is grounded in both pedagogical theory and empirical evidence, with an emphasis on collaborative, socially embedded approaches to scientific writing. The proposed academic writing training integrates constructivist and socio-cultural perspectives, emphasizing genre awareness, rhetorical competence, and reflective practices. The Virtual Writing Center would serve as a permanent infrastructure offering personalized tutoring and peer review support, while the language support programs address ongoing needs through workshops, consultations, and access to language tools. Writing retreats provide immersive environments for focused work and mentorship. The recommendations also encourage ethical use of AI tools for translation and writing assistance, fostering digital literacy alongside linguistic proficiency. Collectively, these initiatives aim to transform language from a barrier into a resource, recognizing multilingualism as an asset in global research collaboration. Rather than offering a one-size-fits-all solution, the document advocates for adaptable, community-driven strategies that can evolve within the diverse institutional and disciplinary contexts of the Rubin Research Ecosystem. By implementing these practices, the Ecosystem could lead efforts to democratize scientific communication and foster a more equitable, multilingual research culture.
Related papers
- From Word to World: Evaluate and Mitigate Culture Bias via Word Association Test [48.623761108859085]
We extend the human-centered word association test (WAT) to assess the alignment of large language models with cross-cultural cognition.<n>To mitigate the culture preference, we propose CultureSteer, an innovative approach that integrates a culture-aware steering mechanism.
arXiv Detail & Related papers (2025-05-24T07:05:10Z) - Preserving Cultural Identity with Context-Aware Translation Through Multi-Agent AI Systems [0.4218593777811082]
Language is a cornerstone of cultural identity, yet globalization and the dominance of major languages have placed nearly 3,000 languages at risk of extinction.<n>Existing AI-driven translation models prioritize efficiency but often fail to capture cultural nuances, idiomatic expressions, and historical significance.<n>We propose a multi-agent AI framework designed for culturally adaptive translation in underserved language communities.
arXiv Detail & Related papers (2025-03-05T06:43:59Z) - LIMBA: An Open-Source Framework for the Preservation and Valorization of Low-Resource Languages using Generative Models [62.47865866398233]
This white paper proposes a framework to generate linguistic tools for low-resource languages.
By addressing the data scarcity that hinders intelligent applications for such languages, we contribute to promoting linguistic diversity.
arXiv Detail & Related papers (2024-11-20T16:59:41Z) - Responsible Multilingual Large Language Models: A Survey of Development, Applications, and Societal Impact [5.803667039914564]
This work bridges the gap by providing an end-to-end framework for developing and deploying MLLMs in production environments.
Our findings reveal critical challenges in supporting linguistic diversity, with 88.38% of world languages categorized as low-resource.
This survey provides essential guidance for practitioners and researchers working to develop more inclusive and effective multilingual AI systems.
arXiv Detail & Related papers (2024-10-23T03:19:15Z) - A global AI community requires language-diverse publishing [1.4579344926652844]
We argue that the requirement for English language publishing upholds and reinforces broader regimes of extraction in AI.
We propose alternative futures for a healthier publishing culture, organized around three themes.
arXiv Detail & Related papers (2024-08-27T04:20:10Z) - A Systematic Survey of Natural Language Processing for the Greek Language [2.3499129784547663]
This study introduces a generalizable framework for systematic monolingual NLP surveys.<n>Our approach integrates a structured search protocol to minimize bias, an NLP task taxonomy for classification, and language resource to identify potential benchmarks.<n>We apply this framework to Greek NLP (2012-2023), providing an in-depth analysis of its current state, task-specific progress, and resource gaps.
arXiv Detail & Related papers (2024-07-13T12:01:52Z) - A Survey on Large Language Models with Multilingualism: Recent Advances and New Frontiers [51.8203871494146]
The rapid development of Large Language Models (LLMs) demonstrates remarkable multilingual capabilities in natural language processing.<n>Despite the breakthroughs of LLMs, the investigation into the multilingual scenario remains insufficient.<n>This survey aims to help the research community address multilingual problems and provide a comprehensive understanding of the core concepts, key techniques, and latest developments in multilingual natural language processing based on LLMs.
arXiv Detail & Related papers (2024-05-17T17:47:39Z) - FREDSum: A Dialogue Summarization Corpus for French Political Debates [26.76383031532945]
We present a dataset of French political debates for the purpose of enhancing resources for multi-lingual dialogue summarization.
Our dataset consists of manually transcribed and annotated political debates, covering a range of topics and perspectives.
arXiv Detail & Related papers (2023-12-08T05:42:04Z) - Computer Science Framework to Teach Community-Based Environmental
Literacy and Data Literacy to Diverse Students [4.734693086769023]
This study introduces an integrated curriculum designed to empower underrepresented students by combining environmental literacy, data literacy, and computer science.
The framework promotes environmental awareness, data literacy, and civic engagement using a culturally sustaining approach.
arXiv Detail & Related papers (2023-09-25T12:49:39Z) - BabySLM: language-acquisition-friendly benchmark of self-supervised
spoken language models [56.93604813379634]
Self-supervised techniques for learning speech representations have been shown to develop linguistic competence from exposure to speech without the need for human labels.
We propose a language-acquisition-friendly benchmark to probe spoken language models at the lexical and syntactic levels.
We highlight two exciting challenges that need to be addressed for further progress: bridging the gap between text and speech and between clean speech and in-the-wild speech.
arXiv Detail & Related papers (2023-06-02T12:54:38Z) - Interactive Natural Language Processing [67.87925315773924]
Interactive Natural Language Processing (iNLP) has emerged as a novel paradigm within the field of NLP.
This paper offers a comprehensive survey of iNLP, starting by proposing a unified definition and framework of the concept.
arXiv Detail & Related papers (2023-05-22T17:18:29Z) - Experience Grounds Language [185.73483760454454]
Language understanding research is held back by a failure to relate language to the physical world it describes and to the social interactions it facilitates.
Despite the incredible effectiveness of language processing models to tackle tasks after being trained on text alone, successful linguistic communication relies on a shared experience of the world.
arXiv Detail & Related papers (2020-04-21T16:56:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.