Book
- Finished the Google Doodle. Next is balloon challenge.
GOES
- https://datascience.stackexchange.com/questions/51065/what-is-the-positional-encoding-in-the-transformer-model – Positional Encoding, the terms were a bit vague in some other articles
- kazemnejad.com/blog/transformer_architecture_positional_encoding
- From Tensorflow: tutorials/text/transformer
- http://jalammar.github.io/illustrated-transformer/ – For Attention Heads
SBIR
- Meeting on NN architecture
GPT Agents
- Language Models are few-shot learners (video). Lots of good stuff about how to build probes.
- Was able to get the ecco library to work with my local model!
from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer
from ecco.lm import LM
activations=False
attention=False
hidden_states=True
activations_layer_nums=None
model_str = '../models/chess_model'
tokenizer = AutoTokenizer.from_pretrained(model_str)
model = AutoModelForCausalLM.from_pretrained(model_str, output_hidden_states=hidden_states, output_attentions=attention)
lm_kwargs = {
'collect_activations_flag': activations,
'collect_activations_layer_nums': activations_layer_nums}
lm = LM(model, tokenizer, **lm_kwargs)
# Input text
text = "Check."
# Generate 100 tokens to complete the input text.
output = lm.generate(text, generate=100, do_sample=True)
print(output)
Had a nice chat with Antonio about an introduction for the special issue
