Phil 1.12.2022

Spanish!

GPT Agents

  • I think I want to put the results into three sections: 1) Memorization, or the learning of the meta-wrapper, 2) Interpolation, or how the model re-creates correct reviews 3) Extrapolation, how the model creates new (zero shot) reviews
  • Add a section to the beginning of the methods section stating that all finetuning was done on the Huggingface GPT-2 117M parameter model.
    • For speed (easier to produce a model for comparison)
    • For the environment
    • To show that state-of-the art insight into TLMs does not require building large models