Calorie Aware Automatic Meal Kit Generation from an Image
- URL: http://arxiv.org/abs/2112.09839v1
- Date: Sat, 18 Dec 2021 04:16:12 GMT
- Title: Calorie Aware Automatic Meal Kit Generation from an Image
- Authors: Ahmad Babaeian Jelodar and Yu Sun
- Abstract summary: Given a single cooking image, a pipeline for calorie estimation and meal re-production is proposed.
Portion estimation introduced in the model helps improve calorie estimation and is also beneficial for meal re-production in different serving sizes.
- Score: 7.170180366236038
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Calorie and nutrition research has attained increased interest in recent
years. But, due to the complexity of the problem, literature in this area
focuses on a limited subset of ingredients or dish types and simple
convolutional neural networks or traditional machine learning. Simultaneously,
estimation of ingredient portions can help improve calorie estimation and meal
re-production from a given image. In this paper, given a single cooking image,
a pipeline for calorie estimation and meal re-production for different servings
of the meal is proposed. The pipeline contains two stages. In the first stage,
a set of ingredients associated with the meal in the given image are predicted.
In the second stage, given image features and ingredients, portions of the
ingredients and finally the total meal calorie are simultaneously estimated
using a deep transformer-based model. Portion estimation introduced in the
model helps improve calorie estimation and is also beneficial for meal
re-production in different serving sizes. To demonstrate the benefits of the
pipeline, the model can be used for meal kits generation. To evaluate the
pipeline, the large scale dataset Recipe1M is used. Prior to experiments, the
Recipe1M dataset is parsed and explicitly annotated with portions of
ingredients. Experiments show that using ingredients and their portions
significantly improves calorie estimation. Also, a visual interface is created
in which a user can interact with the pipeline to reach accurate calorie
estimations and generate a meal kit for cooking purposes.
Related papers
- NutritionVerse-Direct: Exploring Deep Neural Networks for Multitask Nutrition Prediction from Food Images [63.314702537010355]
Self-reporting methods are often inaccurate and suffer from substantial bias.
Recent work has explored using computer vision prediction systems to predict nutritional information from food images.
This paper aims to enhance the efficacy of dietary intake estimation by leveraging various neural network architectures.
arXiv Detail & Related papers (2024-05-13T14:56:55Z) - How Much You Ate? Food Portion Estimation on Spoons [63.611551981684244]
Current image-based food portion estimation algorithms assume that users take images of their meals one or two times.
We introduce an innovative solution that utilizes stationary user-facing cameras to track food items on utensils.
The system is reliable for estimation of nutritional content of liquid-solid heterogeneous mixtures such as soups and stews.
arXiv Detail & Related papers (2024-05-12T00:16:02Z) - NutritionVerse-Real: An Open Access Manually Collected 2D Food Scene
Dataset for Dietary Intake Estimation [68.49526750115429]
We introduce NutritionVerse-Real, an open access manually collected 2D food scene dataset for dietary intake estimation.
The NutritionVerse-Real dataset was created by manually collecting images of food scenes in real life, measuring the weight of every ingredient and computing the associated dietary content of each dish.
arXiv Detail & Related papers (2023-11-20T11:05:20Z) - NutritionVerse: Empirical Study of Various Dietary Intake Estimation Approaches [59.38343165508926]
Accurate dietary intake estimation is critical for informing policies and programs to support healthy eating.
Recent work has focused on using computer vision and machine learning to automatically estimate dietary intake from food images.
We introduce NutritionVerse- Synth, the first large-scale dataset of 84,984 synthetic 2D food images with associated dietary information.
We also collect a real image dataset, NutritionVerse-Real, containing 889 images of 251 dishes to evaluate realism.
arXiv Detail & Related papers (2023-09-14T13:29:41Z) - A Large-Scale Benchmark for Food Image Segmentation [62.28029856051079]
We build a new food image dataset FoodSeg103 (and its extension FoodSeg154) containing 9,490 images.
We annotate these images with 154 ingredient classes and each image has an average of 6 ingredient labels and pixel-wise masks.
We propose a multi-modality pre-training approach called ReLeM that explicitly equips a segmentation model with rich and semantic food knowledge.
arXiv Detail & Related papers (2021-05-12T03:00:07Z) - Revamping Cross-Modal Recipe Retrieval with Hierarchical Transformers
and Self-supervised Learning [17.42688184238741]
Cross-modal recipe retrieval has recently gained substantial attention due to the importance of food in people's lives.
We propose a simplified end-to-end model based on well established and high performing encoders for text and images.
Our proposed method achieves state-of-the-art performance in the cross-modal recipe retrieval task on the Recipe1M dataset.
arXiv Detail & Related papers (2021-03-24T10:17:09Z) - An End-to-End Food Image Analysis System [8.622335099019214]
We propose an image-based food analysis framework that integrates food localization, classification and portion size estimation.
Our proposed framework is end-to-end, i.e., the input can be an arbitrary food image containing multiple food items.
Our framework is evaluated on a real life food image dataset collected from a nutrition feeding study.
arXiv Detail & Related papers (2021-02-01T05:36:20Z) - Multi-Task Learning for Calorie Prediction on a Novel Large-Scale Recipe
Dataset Enriched with Nutritional Information [25.646488178514186]
In this work, we aim to estimate the calorie amount of a meal directly from an image by learning from recipes people have published on the Internet.
We propose the pic2kcal benchmark comprising 308,000 images from over 70,000 recipes including photographs, ingredients and instructions.
Our experiments demonstrate clear benefits of multi-task learning for calorie estimation, surpassing the single-task calorie regression by 9.9%.
arXiv Detail & Related papers (2020-11-02T16:11:51Z) - Picture-to-Amount (PITA): Predicting Relative Ingredient Amounts from
Food Images [24.26111169033236]
We study the novel and challenging problem of predicting the relative amount of each ingredient from a food image.
We propose PITA, the Picture-to-Amount deep learning architecture to solve the problem.
Experiments on a dataset of recipes collected from the Internet show the model generates promising results.
arXiv Detail & Related papers (2020-10-17T06:43:18Z) - Multi-Task Image-Based Dietary Assessment for Food Recognition and
Portion Size Estimation [6.603050343996914]
We propose an end-to-end multi-task framework that can achieve both food classification and food portion size estimation.
Our results outperforms the baseline methods for both classification accuracy and mean absolute error for portion estimation.
arXiv Detail & Related papers (2020-04-27T21:35:07Z) - Cross-Modal Food Retrieval: Learning a Joint Embedding of Food Images
and Recipes with Semantic Consistency and Attention Mechanism [70.85894675131624]
We learn an embedding of images and recipes in a common feature space, such that the corresponding image-recipe embeddings lie close to one another.
We propose Semantic-Consistent and Attention-based Networks (SCAN), which regularize the embeddings of the two modalities through aligning output semantic probabilities.
We show that we can outperform several state-of-the-art cross-modal retrieval strategies for food images and cooking recipes by a significant margin.
arXiv Detail & Related papers (2020-03-09T07:41:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.