Project Introduction
As part of my AI coursework
(CS224n) at Stanford, my team attempted
to develop a transformer model to generate
recipe instructions from only a recipe
title. Existing recipee generation pipelines requir
significant compute to achieve. We set out to identify
less complex architectures that could provide sufficient
recipe generation outputs to expand the possible use of
recipe geneeation.
Abstract
Recipe generation from recipe titles only is currently unsolved,
as state of the art models require both recipe titles and ingredients lists for
instruction generation (Lee et al., 2020). This project investigates if a number
of different architectures such as Long Short-Term Memory (LSTM) encoder-decoders,
LSTM decoders, or Transformer-based decoders, can produce meaningful ingredient lists
when given recipe titles only. The recipe titles and generated ingredients are then passed
into an existing recipe instruction generation framework to produce cooking instructions
(Liu et al., 2022). Our best ingredient generation model yielded qualitatively coherent
ingredients lists with BLEU score 11.2 and F1 score 8.9, however, the BLEU and ROUGE-L
scores for the final recipe instructions with ingredients from our selected transformer
decode were 3.4 and 22.7. The baseline plug-and-play recipe instruction generation framework,
relying on RecipeGPT and ground truth recipe title and ingredients demonstrates BLEU and
ROUGE-L scores of 13.73 and 39.1 respectively for instruction generation. Since BLEU and
ROUGE-L performance are influenced by n-gram matching and order, further evaluation would
be required with metrics such as Semantic Textual Similarity (STS) to evaluate the meaning
of the produced ingredients in the context of each recipe.