It snowed again! I think that’s more snow in one week than the past two years
GPT Agents
Need to compare the performance of each model for each probe and compare to ground truth. One thing to point out is how little data there is to sample:
SBIRs
Fixing the “find matching”
Make node size log-based
Book
Put together some more data. Need to change the maps a bit
We evaluate OTS and custom methods on the following datasets. While some of these datasets have common targets, for example, Trump is present in four of them, they are all collected in different periods of time, with different keywords (c.f Appendix B). All datasets have stance labels of ‘favor’, ‘against’, and ‘none’ towards the targets. (EMNLP)
Finished with generating the new data, now we get to see if it works!
It’s pretty good. Here’s the two GPT models, one trained on the first 50k reviews of the American dataset (iso) and the other trained on the first 50k of the American dataset that do not contain the string “vegetarian options”. The probes are:
no vegetarian options
some vegetarian options
several vegetarian options
many vegetarian options
Basically identical
Now I need to compare the response vs the ground truth for each of the probes
Jamie Raskin just released a book that apparently has some overlap with my work? Trying to track it down. Here’s something from CBS
GPT Agents
Creating unistar models from the corpora that have ‘vegetarian options’ removed. As they are trained, I’m also generating responses to the vegetarian prompts that I’ll do the star and unigram compares with. Then put that in a table and write the paper around it. Also, add the Floober part or something fanciful.
Models are all created. Finished running the first two and am now adding sentiment to them
SBIRs
Continue code cleanup and documenting. I managed to remove a good deal of code that had to do with handing raw text selection of topics, since that seems to be broken in tk
Finished commenting QueryFrame. Now I need to fix that listing problem in on_link_existing_clicked()
It got really cold last night and I had forgotten to turn the water off to the outside and lost the faucet on the deck. Could have been worse. At least the pipes didn’t burst
Thinking about submitting a writeup on Sanhedrin 17a (Section 10.4 of the dissertation. Mostly) for the We Robot conference
Abstracts due: March 7
Decisions: May 9
Final papers due: August 8
Book
Playing around with negative scalars to see how that works. This resulted in some code cleanup and a better color gradient. Not sure if it looks better though:
Still like this better:
SBIRs
Sprint planning
Working on code cleanup for MabBuilder. First, adding comments!
Fixed the exit condition that happened when clicking the ‘X’ close icon in the text compare popup
Next, check through all the button behavior in QueryFrame
Set Group
Add Topic/Seed
Add Topic
Add Seed
Find Closest (and dialog)
Add Group
Next Seed
Rerun Seed
Get Topic Details
Direct Prompt
Wikipedia
Link Existing (make this work with descending length topics)
GPT Agents
3:30 Meeting. Going to make some models that explicitly are missing the phrase ‘vegetarian options’ from the training corpora. I’ll then run those as to compare to ‘vegetarian options’ in the ground truth by star and the other GPT models
Get the number of POSITIVE and NEGATIVE sentiment for each isolated model and compare to ground truth. Make a chart and add to the draft. This is the part that shows that creating models for a population captures that population’s patterns, and that this method is more accurate and reliable than assuming that one general model has all the information needed in an accessible way. Done
Happy New Year everyone! It’s been warm here in the Baltimore region. Working on terrain visualization.
Got lighting working. You attach the lighting node to the node you want it to move with and then set it to the node you want to light. Here’s the code:
You must be logged in to post a comment.