How much can ChatGPT really help Computational Biologists in
Programming?
- URL: http://arxiv.org/abs/2309.09126v2
- Date: Mon, 4 Dec 2023 06:53:49 GMT
- Title: How much can ChatGPT really help Computational Biologists in
Programming?
- Authors: Chowdhury Rafeed Rahman, Limsoon Wong
- Abstract summary: This paper focuses on the potential influence (both positive and negative) of ChatGPT in the mentioned aspects with illustrative examples from different perspectives.
Compared to other fields of computer science, computational biology has - (1) less coding resources, (2) more sensitivity and bias issues (deals with medical data) and (3) more necessity of coding assistance.
- Score: 0.40792653193642503
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: ChatGPT, a recently developed product by openAI, is successfully leaving its
mark as a multi-purpose natural language based chatbot. In this paper, we are
more interested in analyzing its potential in the field of computational
biology. A major share of work done by computational biologists these days
involve coding up bioinformatics algorithms, analyzing data, creating
pipelining scripts and even machine learning modeling and feature extraction.
This paper focuses on the potential influence (both positive and negative) of
ChatGPT in the mentioned aspects with illustrative examples from different
perspectives. Compared to other fields of computer science, computational
biology has - (1) less coding resources, (2) more sensitivity and bias issues
(deals with medical data) and (3) more necessity of coding assistance (people
from diverse background come to this field). Keeping such issues in mind, we
cover use cases such as code writing, reviewing, debugging, converting,
refactoring and pipelining using ChatGPT from the perspective of computational
biologists in this paper.
Related papers
- An Evaluation of Large Language Models in Bioinformatics Research [52.100233156012756]
We study the performance of large language models (LLMs) on a wide spectrum of crucial bioinformatics tasks.
These tasks include the identification of potential coding regions, extraction of named entities for genes and proteins, detection of antimicrobial and anti-cancer peptides, molecular optimization, and resolution of educational bioinformatics problems.
Our findings indicate that, given appropriate prompts, LLMs like GPT variants can successfully handle most of these tasks.
arXiv Detail & Related papers (2024-02-21T11:27:31Z) - Diversifying Knowledge Enhancement of Biomedical Language Models using
Adapter Modules and Knowledge Graphs [54.223394825528665]
We develop an approach that uses lightweight adapter modules to inject structured biomedical knowledge into pre-trained language models.
We use two large KGs, the biomedical knowledge system UMLS and the novel biochemical OntoChem, with two prominent biomedical PLMs, PubMedBERT and BioLinkBERT.
We show that our methodology leads to performance improvements in several instances while keeping requirements in computing power low.
arXiv Detail & Related papers (2023-12-21T14:26:57Z) - Unmasking the giant: A comprehensive evaluation of ChatGPT's proficiency in coding algorithms and data structures [0.6990493129893112]
We evaluate ChatGPT's ability to generate correct solutions to the problems fed to it, its code quality, and nature of run-time errors thrown by its code.
We look into patterns in the test cases passed in order to gain some insights into how wrong ChatGPT code is in these kinds of situations.
arXiv Detail & Related papers (2023-07-10T08:20:34Z) - Evaluation of ChatGPT on Biomedical Tasks: A Zero-Shot Comparison with
Fine-Tuned Generative Transformers [2.5027382653219155]
ChatGPT is a large language model developed by OpenAI.
This paper aims to evaluate the performance of ChatGPT on various benchmark biomedical tasks.
arXiv Detail & Related papers (2023-06-07T15:11:26Z) - ChatGPT Beyond English: Towards a Comprehensive Evaluation of Large
Language Models in Multilingual Learning [70.57126720079971]
Large language models (LLMs) have emerged as the most important breakthroughs in natural language processing (NLP)
This paper evaluates ChatGPT on 7 different tasks, covering 37 diverse languages with high, medium, low, and extremely low resources.
Compared to the performance of previous models, our extensive experimental results demonstrate a worse performance of ChatGPT for different NLP tasks and languages.
arXiv Detail & Related papers (2023-04-12T05:08:52Z) - ChatGPT for Programming Numerical Methods [2.741266294612776]
ChatGPT is a large language model recently released by the OpenAI company.
We explore for the first time the capability of ChatGPT for programming numerical algorithms.
arXiv Detail & Related papers (2023-03-21T12:18:17Z) - Many bioinformatics programming tasks can be automated with ChatGPT [3.2698789104455677]
Recent advances in artificial intelligence have made it possible to translate human-language prompts to functional code.
We evaluated the extent to which one such model -- OpenAI's ChatGPT -- can successfully complete basic- to moderate-level programming tasks.
arXiv Detail & Related papers (2023-03-07T23:32:17Z) - Natural Language to Code Generation in Interactive Data Science
Notebooks [35.621936471322385]
We build ARCADE, a benchmark of 1082 code generation problems using the pandas data analysis framework in data science notebooks.
We develop PaChiNCo, a 62B code language model (LM) for Python computational notebooks, which significantly outperforms public code LMs.
arXiv Detail & Related papers (2022-12-19T05:06:00Z) - Neural Fields in Visual Computing and Beyond [54.950885364735804]
Recent advances in machine learning have created increasing interest in solving visual computing problems using coordinate-based neural networks.
neural fields have seen successful application in the synthesis of 3D shapes and image, animation of human bodies, 3D reconstruction, and pose estimation.
This report provides context, mathematical grounding, and an extensive review of literature on neural fields.
arXiv Detail & Related papers (2021-11-22T18:57:51Z) - Ten Quick Tips for Deep Learning in Biology [116.78436313026478]
Machine learning is concerned with the development and applications of algorithms that can recognize patterns in data and use them for predictive modeling.
Deep learning has become its own subfield of machine learning.
In the context of biological research, deep learning has been increasingly used to derive novel insights from high-dimensional biological data.
arXiv Detail & Related papers (2021-05-29T21:02:44Z) - Information-Theoretic Probing for Linguistic Structure [74.04862204427944]
We propose an information-theoretic operationalization of probing as estimating mutual information.
We evaluate on a set of ten typologically diverse languages often underrepresented in NLP research.
arXiv Detail & Related papers (2020-04-07T01:06:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.