Phil 10.5.2023

Take a look at the IUI abstracts and maybe put together a sortable spreadsheet?

SBIRs

  • 9:00 standup
  • See if the (relative?) ship position data Loren used to create his FOM curves can be incorporated as input data in our app
  • First read of Language Models Represent Space and Time – done. Boy is there a backlash on Xitter
    • Found this in the citations: Mapping Language Models to Grounded Conceptual Spaces
      • A fundamental criticism of text-only language models (LMs) is their lack of grounding—that is, the ability to tie a word for which they have learned a representation, to its actual use in the world. However, despite this limitation, large pre-trained LMs have been shown to have a remarkable grasp of the conceptual structure of language, as demonstrated by their ability to answer questions, generate fluent text, or make inferences about entities, objects, and properties that they have never physically observed. In this work we investigate the extent to which the rich conceptual structure that LMs learn indeed reflects the conceptual structure of the non-linguistic world—which is something that LMs have never observed. We do this by testing whether the LMs can learn to map an entire conceptual domain (e.g., direction or colour) onto a grounded world representation given only a small number of examples. For example, we show a model what the word “left” means using a textual depiction of a grid world, and assess how well it can generalise to related concepts, for example, the word “right”, in a similar grid world. We investigate a range of generative language models of varying sizes (including GPT-2 and GPT-3), and see that although the smaller models struggle to perform this mapping, the largest model can not only learn to ground the concepts that it is explicitly taught, but appears to generalise to several instances of unseen concepts as well. Our results suggest an alternative means of building grounded language models: rather than learning grounded representations “from scratch”, it is possible that large text-only models learn a sufficiently rich conceptual structure that could allow them to be grounded in a data-efficient way.
    • Understanding intermediate layers using linear classifier probes
      • Neural network models have a reputation for being black boxes. We propose to monitor the features at every layer of a model and measure how suitable they are for classification. We use linear classifiers, which we refer to as “probes”, trained entirely independently of the model itself. This helps us better understand the roles and dynamics of the intermediate layers. We demonstrate how this can be used to develop a better intuition about models and to diagnose potential problems. We apply this technique to the popular models Inception v3 and Resnet-50. Among other things, we observe experimentally that the linear separability of features increase monotonically along the depth of the model.
  • Slides for demos

GPT agents

  • 2:00 Meeting
  • Send story to CACM and see if they would like to pursue and what the lead times are – done
  • Worked a bit on Neema’s Senate testimony

Phil 10.4.2023

The bidding phase of IUI 2024 is now open. Now my present/future self has to live up to the commitments made by me in the past.

Just got back from the excellent Digital Platforms and Societal Harms IEEE event at American University. Some of the significant points that were discussed over the past two days:

  • Moderation is hard. Determining, for example, what is hate speech in the ten seconds or so allocated to moderators is mostly straightforward but often complicated and very dependent of locale and culture. I get the feeling that – based on examining content alone – machine learning could easily take care of 50% or so, particularly if you just decide to lump in satire and mockery. Add network analysis and you could probably be more sophisticated and get up to 70%? Handling the remaining 30% is a crushing job that would send most normal people running. Which means that the job of moderating for unacceptable content is its own form of exploitation.
  • Governments that were well positioned to detect and disrupt organizations like ISIS are no better prepared than a company like Meta when it comes to handling radical extremists from within the dominant culture that produced the company. In the US, that’s largely white and some variant of Christian. I’d assume that in China the same pattern exists for their dominant group.
  • There is a sense that all of our systems are reactive. That they only come into play when something has happened, not before something happens. Intervention for someone who is radicalizing requires human intervention. Which means it’s expensive and hard to scale. Moonshot is working to solve this problem, and has made surprisingly good progress, so there may be ways to make this work.
  • Militant accelerationism, or hastening societal collapse, is a thing. The exploitation of vulnerable people to become expendable munitions is being attempted by online actors. Generative AI will be a tool for these people, if it isn’t already.
  • There are quite a few good databases, but they are so toxic that they are largely kept in servers that are isolated from the internet to a greater or lesser degree. Public repositories are quite rare.
  • The transformation of Twitter to X is a new, very difficult problem. Twitter built up so much social utility as, for example, early warning, or reports from disaster areas that it can’t be removed from an App Store in the same way that an app that permits similar toxic behavior but only has 25 users can be. No one seems to have a good answer for this.
  • The Fediverse also appears to complicate harm tracking and prevention. Since there is no single source, how do you pull your Mastodon App if some people are accessing (possibly blacklisted) servers hosting hate speech? Most people are using the app for productive reasons. Now what?
  • Removing content doesn’t remove the person making the content. Even without any ability to post, or even with full bans from a platform, they can still search for targets and buy items that can enable them to cause harm in the real world. This is why moderation is only the lowest bar. Detection and treatment should be a goal.
  • Of course all these technologies are two edged swords. Detection and treatment in an authoritarian situation might mean finding reporters or human rights activist and imprisoning them.
  • The organizers are going to make this a full conference next year, with a call for papers and publication, so keep an eye on this space if you’re interested: https://tech-forum.computer.org/societal-harms-2023/

SBIRs

  • The War Elephants paper got a hard reject. Need to talk to Aaron to see How to proceed. Done
  • Add ASRC to letterhead – Done
  • Expense report! Done
  • Had a good chat with Rukan about using the SimAccel for interactive analysis of trajectories and FOM curves
  • Work on Senate story

GPT Agents

  • 3:00 Alden meeting Nope
  • Gotta get back to maps. Found this:
  • Language Models Represent Space and Time
  • The capabilities of large language models (LLMs) have sparked debate over whether such systems just learn an enormous collection of superficial statistics or a coherent model of the data generating process — a world model. We find evidence for the latter by analyzing the learned representations of three spatial datasets (world, US, NYC places) and three temporal datasets (historical figures, artworks, news headlines) in the Llama-2 family of models. We discover that LLMs learn linear representations of space and time across multiple scales. These representations are robust to prompting variations and unified across different entity types (e.g. cities and landmarks). In addition, we identify individual “space neurons” and “time neurons” that reliably encode spatial and temporal coordinates. Our analysis demonstrates that modern LLMs acquire structured knowledge about fundamental dimensions such as space and time, supporting the view that they learn not merely superficial statistics, but literal world models.
  • Because the world is mean, the paper cites two papers from 2022 on reconstructing the game board from knowledge in the model with Chess and Othello. My paper did this in 2020. Grumble

Phil 10.2.2023

Logically combines advanced AI with one of the world’s largest dedicated fact-checking teams. We help governments, businesses, and enterprise organizations uncover and address harmful misinformation and deliberate disinformation online.

Phil 9.29.2023

SBIRs

  • 4:00 Technical Fellows meeting. Try to remember this time. Done! Went… ok? I think they are looking for something very specific.
  • Made a cover

GPT Agents

  • Fold in the new Informed Consent and add a link to a pdf of the doc – done!

Phil 9.28.2023

Different ways to do learning that used to be RL. Need to look at this, and there is a repo.

SBIRs

  • 9:00 standup – done
  • 11:30 Touchpoint – nope
  • Finish poster and submit. Need a V2 – done!

GPT Agents

  • Finish training – done!
  • 2:00 Meeting – need to update the Informed consent – Rewrote it. Now I need to stuff all that text on the webpage. Ugh.

Phil 9.27.2023

Dinner with Greg at 6:00

SBIRs

  • Continue poster. Made a lot of synthetic art “in the style of Francis Bacon,” which captures the mood nicely. Really like this one, even though it’s not what I was after:
  • No meetings today?!
  • Fix headers and send WP to Lauren

GPT Agents

  • Finish CITI training
  • Start fixing IRB submission

Phil 9.26.2023

The radar patterns are still pretty confusing. Going to try to get out at 11:00

SBIRs

  • Ping Bob S. About contacting SEG – done
  • Added some text to the M30 doc and made a tabel of options
  • Start poster
  • 12:30 JSC discussion
  • 2:30 AI Ethics
  • 3:00 M30 Meeting

GPT Agents

  • Do more CITI training. One more down

Phil 9.25.2023

It’s Fall! Ping Nathan

SBIRs

  • Make a test_harness.tex file – done
  • Finish venues – done
  • Slides! – done
  • 9:00 Sprint demos – done
  • Stories! – done
  • 2:00 MDA
  • 3:00 Sprint planning

GPT Agents

  • Test out on people?

Phil 9.24.2023

Very rainy weekend. Worked a bit on the book, and on trimming long CVs

Connecting Large Language Models with Evolutionary Algorithms Yields Powerful Prompt Optimizers

  • Large Language Models (LLMs) excel in various tasks, but they rely on carefully crafted prompts that often demand substantial human effort. To automate this process, in this paper, we propose a novel framework for discrete prompt optimization, called EvoPrompt, which borrows the idea of evolutionary algorithms (EAs) as they exhibit good performance and fast convergence. To enable EAs to work on discrete prompts, which are natural language expressions that need to be coherent and human-readable, we connect LLMs with EAs. This approach allows us to simultaneously leverage the powerful language processing capabilities of LLMs and the efficient optimization performance of EAs. Specifically, abstaining from any gradients or parameters, EvoPrompt starts from a population of prompts and iteratively generates new prompts with LLMs based on the evolutionary operators, improving the population based on the development set. We optimize prompts for both closed- and open-source LLMs including GPT-3.5 and Alpaca, on 9 datasets spanning language understanding and generation tasks. EvoPrompt significantly outperforms human-engineered prompts and existing methods for automatic prompt generation by up to 25% and 14% respectively. Furthermore, EvoPrompt demonstrates that connecting LLMs with EAs creates synergies, which could inspire further research on the combination of LLMs and conventional algorithms.

Phil 9.22.2023

Mow lawn before it rains for days!

Inside Tiktok’s real-life frenzies – from riots to false murder accusations

  • the BBC has identified four episodes in recent months where disproportionate engagement on TikTok was connected to harmful behaviour:
    • An online obsession with a murder case in Idaho, USA, that led to innocent people being falsely accused
    • Interference in the police investigation of Nicola Bulley, who went missing in Lancashire, UK
    • School protests involving vandalism spreading across the UK
    • Fanning flames of riots in France, which spread at an unusual intensity and to unexpected locations
    • Ex-staffers at TikTok liken these frenzies to “wildfires” and describe them as “dangerous”, especially as the app’s audience can be young and impressionable.

SBIRs

  • 10:00 meeting with Aaron
  • 2:30 Tech Fellows interview
  • Finish Dahlgren WP draft 1

GPT Agents

  • Add pasting areas for Education history, Work history, and Publications. Cut publications when the entire text is > ~10,000 words

Phil 9.21.2023

Tweaked my The Great Chain of Being as a General Theory of Racism post

SBIRs

  • 9:00 standup – multiple issues with ENU reference frames
  • Working on the Dahlgren white paper – good progress
  • The Scale, well, I guess it’s on its way to being a book now is closing in on 40k words:
  • I found this helpful little site, which suggests that a “big idea” book should be between 60k and 80k words

GPT Agents

  • 2:00 Meeting
  • Need to fix a few more things

Phil 9.20.2023

SBIRs

  • Need to make a poster and submit by the 28th to the Digital Platforms and Societal Harms event. Probably show the 3 types of attacks (email examples) and mitigation. I could bring a laptop with ContextExplorer too.
  • Work on MAST whitepaper, then get together with Aaron at 1:00. Made good progress. The goal is to have a first draft by Friday COB
  • 10:00 JSC Data Review. There is a lot. Ron’s going to do some summary statistics.
  • Maybe more scale paper this evening? Yup, finished Arms Control

GPT Agents

  • 3:00 Meeting with Alden

Phil 8.19.2022

Caulk tub

SBIRs

  • MDA meeting from yesterday because Zac is back now. Done. Need to find out from Bob what the best target is.
  • More scale paper. Got started on the Arms Control section, which is coming along nicely. It seems that arms control is most effective when powers are not in open conflict (e.g. the cold war). Which is mostly the case now, though I wonder how much The Russian-Ukraine war would effect that. I think that there would be more focus on AI-enhanced weapons? Which for an agreement on Societal AI weapons might make things easier.
  • Need to get some work done on the MAST white paper

GPT Agents

  • Progress on getting lists of deans and chairs together to ask for participation.

Phil 9.18.2023

Centaurs and Cyborgs on the Jagged Frontier (from this paper)

  • …for 18 different tasks selected to be realistic samples of the kinds of work done at an elite consulting company, consultants using ChatGPT-4 outperformed those who did not, by a lot. On every dimension. Every way we measured performance.

El ingenioso hidalgo don Quijote de la Mancha

SBIRs

  • Meeting with Steve to talk about things
  • MDA weekly meeting

GPT Agents